Sample records for machine readable cataloging

  1. The critical evaluation of stellar data

    NASA Technical Reports Server (NTRS)

    Underhill, A. B.; Mead, J. M.; Nagy, T. A.

    1977-01-01

    The paper discusses the importance of evaluating a catalog of stellar data, whether it is an old catalog being made available in machine-readable form, or a new catalog written expressly in machine-readable form, and discusses some principles to be followed in the evaluation of such data. A procedure to be followed when checking out an astronomical catalog on magnetic tape is described. A cross index system which relates the different identification numbers of a star or other astronomical object as they appear in different catalogs in machine-readable form is described.

  2. Astronomical Data Center Bulletin, volume 1, number 2

    NASA Technical Reports Server (NTRS)

    Nagy, T. A.; Warren, W. H., Jr.; Mead, J. M.

    1981-01-01

    Work in progress on astronomical catalogs is presented in 16 papers. Topics cover astronomical data center operations; automatic astronomical data retrieval at GSFC; interactive computer reference search of astronomical literature 1950-1976; formatting, checking, and documenting machine-readable catalogs; interactive catalog of UV, optical, and HI data for 201 Virgo cluster galaxies; machine-readable version of the general catalog of variable stars, third edition; galactic latitude and magnitude distribution of two astronomical catalogs; the catalog of open star clusters; infrared astronomical data base and catalog of infrared observations; the Air Force geophysics laboratory; revised magnetic tape of the N30 catalog of 5,268 standard stars; positional correlation of the two-micron sky survey and Smithsonian Astrophysical Observatory catalog sources; search capabilities for the catalog of stellar identifications (CSI) 1979 version; CSI statistics: blue magnitude versus spectral type; catalogs available from the Astronomical Data Center; and status report on machine-readable astronomical catalogs.

  3. Documentation for the machine-readable version of the revised Catalogue of Stellar Rotational Velocities of Uesugi and Fukuda (1982)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    The machine-readable catalog provides mean data on the old Slettebak system for 6472 stars. The catalog results from the review, analysis and transformation of 11460 data from 102 sources. Star identification, (major catalog number, name if the star has one, or cluster identification, etc.), a man projected rotational velocity, and a list of source references re included. The references are given in a second file included with the catalog when it is distributed on magnetic tape. The contents and/formats of the the data and reference files of the machine-readable catalog are described to enable users to read and process the data.

  4. National Aspects of Creating and Using MARC/RECON Records.

    ERIC Educational Resources Information Center

    Rather, John C., Ed.; Avram, Henriette D., Ed.

    The Retrospective Conversion (RECON) Working Task Force investigated the problems of converting retrospective catalog records to machine readable form. The major conclusions and recommendations of the Task Force cover five areas: the level of machine-readable records, conversion of other machine-readable data bases, a machine-readable National…

  5. Documentation for the machine-readable version of The Revised AFGL Infrared Sky Survey Catalog (Price and Murdock 1983)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    A detailed description of the machine-readable catalog as it is currently being distributed from the Astronomical Data Center is given. The catalog contains a main data file of 2970 sources and a supplemental file of 3176 sources measured at wavelengths of 4.2, 11, 20 and 27 microns.

  6. A catalog of stellar spectrophotometry (Adelman, et al. 1989): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.; Adelman, Saul J.

    1990-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the astronomical data centers, is described. The catalog is a collection of spectrophotometric observations made using rotating grating scanners and calibrated with the fluxes of Vega. The observations cover various wavelength regions between about 330 and 1080 nm.

  7. Documentation for the machine-readable version of the Henry Draper Catalogue (edition 1985)

    NASA Technical Reports Server (NTRS)

    Roman, N. G.; Warren, W. H., Jr.

    1985-01-01

    An updated, corrected and extended machine-readable version of the catalog is described. Published and unpublished errors discovered in the previous version was corrected; letters indicating supplemental stars in the BD have been moved to a new byte to distinguish them from double-star components; and the machine readable portion of The Henry Draper Extension (HDE) (HA 100) was converted to the same format as the main catalog, with additional data added as necessary.

  8. A search for ultraviolet-excess objects (Kondo, Noguchi, and Maehara 1984): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    A list of 1186 ultraviolet-excess objects (designated KUV) was compiled as a result of a search conducted with the 105-cm Schmidt telescope of the Kiso station of the Tokyo Astronomical Observatory. This document describes the machine readable version of the KUV survey list and presents a sample listing showing the logical records as they are recorded in the machine readable catalog. The KUV data include equatorial coordinates, magnitudes, color indices, and identifications for previously cataloged objects.

  9. Documentation for the machine-readable version of the Revised S210 Catalog of Far-Ultraviolet Objects (Page, Carruthers and Heckathorn 1982)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    A detailed description of the machine-readable revised catalog as it is currently being distributed from the Astronomical Data Center is given. This catalog of star images was compiled from imagery obtained by the Naval Research Laboratory (NRL) Far-Ultraviolet Camera/Spectrograph (Experiments S201) operated from 21 to 23 April 1972 on the lunar surface during the Apollo 16 mission. The documentation includes a detailed data format description, a table of indigenous characteristics of the magnetic tape file, and a sample listing of data records exactly as they are presented in the machine-readable version.

  10. Documentation for the machine-readable version of the Cape Photographic Durchmusterung (CPD)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The complete catalog is contained in the magnetic tape file, and corrections published in all errata have been made to the data. The machine version contains 454877 records, but only 454875 stars (two stars were later deleted, but their logical records are retained in the file so that the zone counts are not diiferent from the published catalog).

  11. Documentation for the machine-readable version of the Cordoba Durchmusterung (CD)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is presented. The complete catalog is contained in the magnetic tape file, and corrections published in all corrigenda were made to the data. The machine version contains 613959 records, but only 613953 stars (six stars were later deleted, but their logical records are retained in the file so that the zone counts are not different from the published catalog).

  12. Documentation for the machine-readable version of the catalog of galactic O type stars

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The Catalog of Galactic O-Type Stars (Garmany, Conti and Chiosi 1982), a compilation from the literature of all O-type stars for which spectral types, luminosity classes and UBV photometry exist, contains 765 stars, for each of which designation (HD, DM, etc.), spectral type, V, B-V, cluster membership, Galactic coordinates, and source references are given. Derived values of absolute visual and bolometric magnitudes, and distances are included. The source reference should be consulted for additional details concerning the derived quantities. This description of the machine-readable version of the catalog seeks to enable users to read and process the data with a minimum of guesswork. A copy of this document should be distributed with any machine readable version of the catalog.

  13. A catalog of selected compact radio sources for the construction of an extragalactic radio/optical reference frame (Argue et al. 1984): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    1990-01-01

    This document describes the machine readable version of the Selected Compact Radio Source Catalog as it is currently being distributed from the international network of astronomical data centers. It is intended to enable users to read and process the computerized catalog. The catalog contains 233 strong, compact extragalactic radio sources having identified optical counterparts. The machine version contains the same data as the published catalog and includes source identifications, equatorial positions at J2000.0 and their mean errors, object classifications, visual magnitudes, redshift, 5-GHz flux densities, and comments.

  14. Documentation for the machine-readable version of the Lowell Proper Motion Survey northern hemisphere, the G numbered stars

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    Observed positions, proper motions, estimated photographic magnitudes and colors, and references to identifications in other catalogs are included. Photoelectric data on the UBV system are included for many stars, but no attempt was made to find all existing photometry. The machine-readable catalog is described.

  15. Documentation for the machine-readable version of the catalog of 5,268 standard stars, 1950.0 based on the normal system N30

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1981-01-01

    The machine-readable version of the N30 catalog available on magnetic tape from the Astronomical Data Center is described. Numerical representations of some data fields on the original catalog were changed to conform more closely to formats being used for star-catalog data, plus all records having asterisks indicating footnotes in the published catalog now have corresponding remarks entries in a second tape file; i.e. the footnotes in the published catalog were computerized and are contained in a second file of the tape.

  16. Lowell proper motion survey: Southern Hemisphere (Giclas, Burnham, and Thomas 1978). Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1989-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The catalog is a summary compilation of the Lowell Proper Motion Survey for the Southern Hemisphere, as completed to mid-1978 and published in the Lowell Observatory Bulletins. This summary catalog serves as a Southern Hemisphere companion to the Lowell Proper Motion Survey, Northern Hemisphere.

  17. Documentation for the machine-readable version of the Perth 75: A Catalogue of Positions of 2589 FK4 and FK4S Stars (Nikoloff and Hog 1982)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    Detailed descriptions of the data and format of the machine-readable astronomical catalog are given. The machine version is identical in data content to the published edition, but minor modifications in the data format were made in order to effect uniformity with machine versions of other astronomical catalogs. Stellar motions and positions at epoch and equinox 1950.0 are reported.

  18. Documentation for the machine-readable version of the Lowell Proper Motion Survey, Northern Hemisphere, the G numbered stars

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    This catalog contains a summary of many individual papers published in the Lowell Observatory Bulletins in the years 1958 to 1970. The data in the machine-readable version include observed positions, proper motions, estimated photographic magnitudes and colors, and references to identifications in other catalogs. Photoelectric data on the UBV system are included for many stars, but no attempt was made to find all existing photometry. The machine version contains all data of the published catalog, except the Lowell Bulletin numbers where finding charts can be found. A separate file contains the notes published in the original catalog.

  19. Documentation for the machine-readable version of A Finding List of Stars of Spectral Type F2 and Earlier in a North Galactic Pole Region

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable data set is the result of an objective-prism survey made with an 80 cm/120 cm Schmidt telescope. The F2 and earlier stars were isolated from later type objects by using the MK classification criteria. The catalog contains 601 stars and includes cross identifications to the BD and HD catalogs, coordinates, photographic magnitudes and spectral types. A separate file contains the remarks from the original data tables merged with those following the data. The machine-readable files are described.

  20. The HEAO A-1 X Ray Source Catalog (Wood Et Al. 1984): Documentation for the Machine-Readable Version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The catalog is a compilation of data for 842 sources detected with the U.S. Naval Research Laboratory Large Area Sky Survey Experiment flown aboard the HEAO 1 satellite. The data include source identifications, positions, error boxes, mean X-ray intensities, and cross identifications to other source designations.

  1. Documentation for the machine-readable version of the first Santiago-Pulkovo Fundamental Stars Catalogue (SPF1 catalogue)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of the first Santiago-Pulkovo Fundamental Stars catalog is described. It is intended to enable users to read and process the computerized catalog without the problems and guesswork often associated with such a task. The source reference should be consulted for additional details regarding the measurements, instrument characteristics, reductions, construction of the quasi-absolute system of right ascension, and star positions in the catalog.

  2. The US Naval Observatory Zodiacal Zone Catalog (Douglas and Harrington 1990): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The Zodiacal Zone Catalog is a catalog of positions and proper motions for stars in the magnitude range where m sub v is between 4 and 10, lying within 16 deg of the ecliptic and north of declination -30 deg. The catalog contains positions and proper motions, at epoch, for equator and equinox J2000.0, magnitudes and spectral types taken mostly from the Smithsonian Astrophysical Observatory Star Catalog, and reference positions and proper motions for equinox and epoch B1950.0.

  3. Proceedings of the Conference on Machine-Readable Catalog Copy (3rd, Library of Congress, February 25, 1966).

    ERIC Educational Resources Information Center

    Library of Congress, Washington, DC.

    A conference was held to permit a discussion between the libraries that will participate in the Library of Congress machine-readable cataloging (MARC) pilot project. The MARC pilot will provide an opportunity for the Library of Congress to assess the effect which data conversion places on the Library's normal processing procedures; the suitability…

  4. Documentation for the machine-readable version of the ANS Ultraviolet Photometry Catalogue of Point Sources (Wesselius et al 1982)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine-readable version of the Astronomical Netherlands Satellite ultraviolet photometry catalog is described in detail, with a byte-by-byte format description and characteristics of the data file given. The catalog is a compilation of ultraviolet photometry in five bands, within the wavelength range 155 nm to 330 nm, for 3573 mostly stellar objects. Additional cross reference data (object identification, UBV photometry and MK spectral types) are included in the catalog.

  5. Documentation for the machine readable version of the Yale Catalogue of the Positions and Proper Motions of Stars between Declinations -60 deg and -70 deg (Fallon 1983)

    NASA Technical Reports Server (NTRS)

    Roman, N. G.; Warren, W. H., Jr.

    1984-01-01

    The machine-readable, character-coded version of the catalog, as it is currently being distributed from the Astronomical Data Center(ADC), is described. The format and data provided in the magnetic tape version differ somewhat from those of the published catalog, which was also produced from a tape prepared at the ADC. The primary catalog data are positions and proper motions (equinox 1950.0) for 14597 stars.

  6. Far infrared supplement: Catalog of infrared observations

    NASA Technical Reports Server (NTRS)

    Gezari, D. Y.; Schmitz, M.; Mead, J. M.

    1982-01-01

    The development of a new generation of orbital, airborne and ground-based infrared astronomical observatory facilities, including the infrared astronomical satellite (IRAS), the cosmic background explorer (COBE), the NASA Kuiper airborne observatory, and the NASA infrared telescope facility, intensified the need for a comprehensive, machine-readable data base and catalog of current infrared astronomical observations. The Infrared Astronomical Data Base and its principal data product, this catalog, comprise a machine-readable library of infrared (1 micrometer to 1000 micrometers) astronomical observations published in the scientific literature since 1965.

  7. Southern Durchmusterung (Schoenfeld 1886): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.; Ochsenbein, Francois

    1989-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The Southern Durchmusterung (SD) was computerized at the Centre de Donnees Astronomiques de Strasbourg and at the Astronomical Data Center at the National Space Science Data Center, NASA/Goddard Space Flight Center. Corrigenda listed in the original SD volume and published by Kuenster and Sticker were incorporated into the machine file. In addition, one star indicated to be missing in a published list, and later verified, is flagged so that it can be omitted from computer plotted charts if desired. Stars deleted in the various errata lists were similarly flagged, while those with revised data are flagged and listed in a separate table. This catalog covers the zones -02 to -23 degrees; zones +89 to -01 degrees (the Bonner Durchmusterung) are included in a separate catalog available in machine-readable form.

  8. Documentation for the machine-readable version of A Catalogue of Extragalactic Radio Source Identifications (Veron-Cetty and Veron 1983)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    Detailed descriptions of the data and reference files of the updated and final version of the machine-readable catalog are given. The computerized catalog has greatly expanded since the original published version (1974), and additional information is given. A separate reference file contains bibliographical citations ordered simultaneously by numerical reference and alphabetically by author.

  9. Documentation for the machine-readable version of the lick Saturn-Voyager Reference Star Catalogue

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of the catalog is described. The catalog was prepared in order to determine accurate equatorial coordinates for reference stars in a band of sky against which cameras of the Voyager spacecraft were aligned for observations in the region of Saturn during the flyby. Tape contents and characteristics are described and a sample listing presented.

  10. Documentation for the machine-readable version of the general catalogue of 33342 stars for the epoch 1950 (Boss 1937)

    NASA Technical Reports Server (NTRS)

    Roman, N. G.; Warren, W. H., Jr.

    1983-01-01

    A revised and corrected version of the machine-readable catalog has been prepared. Cross identifications of the GC stars to the HD and DM catalogs have been replaced by data from the new SAO-HD-GC-DM Cross Index (Roman, Warren and Schofield 1983), including component identifications for multiple SAO entries having identical DM numbers in the SAO Catalog, supplemental Bonner Durchmusterung stars (lower case letter designations) and codes for multiple HD stars. Additional individual corrections have been incorporated based upon errors found during analyses of other catalogs.

  11. Documentation for the machine-readable version of the Fourth Cambridge Radio Survey Catalogue (4C) (Pilkington, Gower, Scott and Wills 1965, 1967)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    The machine readable catalogue contains survey data from the papers of Pilkington and Scott and Gower, Scott and Wills. These data result from a survey of radio sources between declinations -07 deg and +80 deg using the large Cambridge interferometer at 178 MHz. The computerized catalog contains for each source the 4C number, 1950 position, measured flux density, and accuracy class. For some sources miscellaneous brief comments such as cross identifications to the 3C catalog or remarks on contamination from nearby sources are given at the ends of the data records. A detailed description of the machine readable catalog as it is currently being distributed by the Astronomical Data Center is given to enable users to read and process the data.

  12. The machine-readable Durchmusterungen - Classical catalogs in contemporary form. [for positional astronomy and identification of stars

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.; Ochsenbein, Francois; Rappaport, Barry N.

    1990-01-01

    The entire series of Durchmusterung (DM) catalogs (Bonner, Southern, Cordoba, Cape Photographic) has been computerized through a collaborative effort among institutions and individuals in France and the United States of America. Complete verification of the data, both manually and by computer, the inclusion of all supplemental stars (represented by lower case letters), complete representation of all numerical data, and a consistent format for all catalogs, should make this collection of machine-readable data a valuable addition to digitized astronomical archives.

  13. Documentation for the machine-readable version of the Survey of the Astrographic Catalogue From 1 to 31 Degrees of Northern Declination (Fresneau 1983)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    A description of the machine readable catalog, including detailed format and tape file characteristics, is given. The machine file is a computation of mean values for position and magnitude at a mean epoch of observation for each unique star in the Oxford, Paris, Bordeaux, Toulouse and Northern Hemisphere Algiers zone. The format was changed to effect more efficient data searching by position and additional duplicate entries were removed. The final catalog contains data for 997311 stars.

  14. Documentation for the machine-readable version of a table of Redshifts for Abell clusters (Sarazin, Rood and Struble 1982)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    The machine readable catalog is described. The machine version contains the same data as the published table, which includes a second file with the notes. The computerized data files are prepared at the Astronomical Data Center. Detected discrepancies and cluster identifications based on photometric estimators are included.

  15. VizieR Online Data Catalog: de Houtman, Kepler and Halley star catalogs (Verbunt+ 2011)

    NASA Astrophysics Data System (ADS)

    Verbunt, F.; van Gent, R. H.

    2011-04-01

    We present Machine-readable versions of the star catalogues of de Houtman (1602), Kepler (1627: Secunda Classis and Tertia Classis) and Halley (1679). In addition to the data from the Historical catalogue, the machine-readable version contains the modern identification with a Hipparcos star and the latter's magnitude, and based on this identification the positional accuracy. For Kepler's catalogues we also give cross references to the catalogue of Ptolemaios (in the edition by Toomer 1998). (4 data files).

  16. Documentation for the machine-readable version of the AGK3 Star Catalogue of Positions and Proper Motions North of -2 deg .5 declination (Dieckvoss and Collaborators 1975)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    A detailed description of the machine-readable astronomical catalog as it is currently being distributed from the Astronomical Data Center is given. Stellar motions and positions are listed herein in tabular form.

  17. Fifth Fundamental Catalogue (FK5). Part 1: Basic fundamental stars (Fricke, Schwan, and Lederle 1988): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The Basic FK5 provides improved mean positions and proper motions for the 1535 classical fundamental stars that had been included in the FK3 and FK4 catalogs. The machine version of the catalog contains the positions and proper motions of the Basic FK5 stars for the epochs and equinoxes J2000.0 and B1950.0, the mean epochs of individual observed right ascensions and declinations used to determine the final positions, and the mean errors of the final positions and proper motions for the reported epochs. The cross identifications to other designations used for the FK5 stars that are given in the published catalog were not included in the original machine versions, but the Durchmusterung numbers have been added at the Astronomical Data Center.

  18. Design of Formats and Packs of Catalog Cards.

    ERIC Educational Resources Information Center

    OCLC Online Computer Library Center, Inc., Dublin, OH.

    The three major functions of the Ohio College Library Center's Shared Cataloging System are: 1) provision of union catalog location listing; 2) making available cataloging done by one library to all other users of the system; and 3) production of catalog cards. The system, based on a central machine readable data base, speeds cataloging and…

  19. Resource Sharing in Montana: A Study of Interlibrary Loan and Alternatives for a Montana Union Catalog.

    ERIC Educational Resources Information Center

    Matthews, Joseph R.

    This study recommends a variety of actions to create and maintain a Montana union catalog (MONCAT) for more effective usage of in-state resources and library funds. Specifically, it advocates (1) merger of existing COM, machine readable bibliographic records, and OCLC tapes into a single microform catalog; (2) acceptance of only machine readable…

  20. Documentation for the machine-readable version of the SAO-HD-GC-DM cross index version 1983

    NASA Technical Reports Server (NTRS)

    Roman, N. G.; Warren, W. H., Jr.; Schofield, N., Jr.

    1983-01-01

    An updated and extended machine readable version of the Smithsonian Astrophysical Observatory star catalog (SAO) is described. A correction of all errors which were found since preparation of the original catalog which resulted from misidentifications and omissions of components in multiple star systems and missing Durchmusterung numbers (the common identifier) in the SAO Catalog are included and component identifications from the Index of Visual Double Stars (IDS) are appended to all multiple SAO entries with the same DM numbers, and lower case letter identifiers for supplemental BD stars are added. A total of 11,398 individual corrections and data additions is incorporated into the present version of the cross index.

  1. Second catalog of interferometric measurements of binary stars (McAlister and Hartkopf 1988): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1989-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The catalog is a compilation of measurements of binary- and multiple-star systems obtained by speckle interferometric techniques; this version supersedes a previous edition of the catalog published in 1985. Stars that have been examined for multiplicity with negative results are included, in which case upper limits for the separation are given. The second version is expanded from the first in that a file of newly resolved systems and six cross-index files of alternate designations are included. The data file contains alternate identifications for the observed systems, epochs of observation, reported errors in position angles and separation, and bibliographical references.

  2. Convergence Toward Common Standards in Machine-Readable Cataloging *

    PubMed Central

    Gull, C. D.

    1969-01-01

    The adoption of the MARC II format for the communication of bibliographic information by the three National Libraries of the U.S.A. makes it possible for those libraries to converge on the remaining necessary common standards for machine-readable cataloging. Three levels of standards are identified: fundamental, the character set; intermediate, MARC II; and detailed, the codes for identifying data elements. The convergence on these standards implies that the National Libraries can create and operate a Joint Bibliographic Data Bank requiring standard book numbers and universal serial numbers for identifying monographs and serials and that the system will thoroughly process contributed catalog entries before adding them to the Data Bank. There is reason to hope that the use of the MARC II format will facilitate catalogers' decision processes. PMID:5782261

  3. Documentation for the machine-readable version of the catalog of supplemental stars to the Bonner Durchmusterung

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1980-01-01

    The magnetic tape version of the Bonn catalog is described. The catalog contains a listing of supplemental stars having lower case letter designations following the BD numbers after which they have been inserted. A sample catalog is also presented.

  4. Documentation for the machine-readable version of the third Santiago-Pulkovo Fundamental Stars Catalogue (SPF-3)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of a catalog of right ascensions of 671 bright stars is described. The observations were made in a series consisting of 70 stars observed along the meridian from +42 deg to the pole in upper culmination and from the pole to -70 deg in lower culmination.

  5. Smithsonian Astrophysical Observatory (SAO) star catalog (Sao staff 1966, edition ADC 1989): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Roman, Nancy G.; Warren, Wayne H., Jr.

    1989-01-01

    An updated, corrected, and extended machine readable version of the catalog is described. Published and unpublished errors discovered in the previous version were corrected, and multiple star and supplemental BD identifications were added to stars where more than one SAO entry has the same Durchmusterung number. Henry Draper Extension (HDE) numbers were added for stars found in both volumes of the extension. Data for duplicate SAO entries (those referring to the same star) were flagged. J2000 positions in usual units and in radians were added.

  6. An Investigation into the Economics of Retrospective Conversion Using a CD-ROM System.

    ERIC Educational Resources Information Center

    Co, Francisca K.

    This study compares the cost effectiveness of using a CD-ROM (compact disk read-only memory) system known as Bibliofile and the currently used OCLC (Online Computer Library Center)-based method to convert a university library's shelflist into a machine-readable database in the MARC (Machine-Readable Cataloging) format. The cost of each method of…

  7. A catalog of stellar spectrophotometry

    NASA Technical Reports Server (NTRS)

    Adelman, S. J.; Pyper, D. M.; Shore, S. N.; White, R. E.; Warren, W. H., Jr.

    1989-01-01

    A machine-readable catalog of stellar spectrophotometric measurements made with rotating grating scanner is introduced. Consideration is given to the processes by which the stellar data were collected and calibrated with the fluxes of Vega (Hayes and Latham, 1975). A sample page from the spectrophotometric catalog is presented.

  8. Documentation for the machine-readable version of the Uppsala general catalogue of galaxies

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of the catalog containing descriptions of galaxies, their surrounding areas, and position angles for flattened galaxies is described. In addition to the correction of several errors discovered in a previous computerized version, a few duplicate records were removed and the record structure was revised slightly to accommodate a large data value and to remove superfluous blanks.

  9. Documentation for the machine-readable version of the catalogue of 20457 Star positions obtained by photography in the declination zone -48 deg to -54 deg (1950)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1983-01-01

    The machine readable catalog, as it is distributed from the Astronomical Data Center, is described. Some minor reformatting of the magnetic tape version is received to decrease the record size and conserve space; the data content is identical to the sample shown in Table VI of the source reference.

  10. Documentation for the machine-readable version of the Catalogue of Nearby Stars, edition 1969

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The Catalogue of Nearby Stars, Edition 1969 (Gliese 1969) contains a number of modifications and additions to the 1957 catalog. It should be mentioned that the 1969 edition lists: (1) all 915 stars of the first edition, even though newer parallaxes place some of the stars below the catalog limit; (2) almost all known stars having trigonometric parallaxes or = 0.045 deg, although in some cases the mean values of trigonometric and spectral or photometric parallaxes are or = 0.045 deg. Pleiades stars and the carbon star X Cnc have been omitted; and (3) all stars with mean (resulting) parallaxes or = 0.045 deg. The resulting catalog contains 1529 single stars and systems with a total of 1890 components (not including spectroscopic and astrometric companions). The machine-readable version of the catalog is described. It is intended to enable users to read and process the data without problems or guesswork.

  11. Lick Saturn-Voyager reference star catalogue (Klemola, Taraji, and Ocampo 1979): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The catalog contains accurate equatorial coordinates for 4551 stars in a band of sky against which cameras of the Voyager spacecraft were pointed for observations in the region of Saturn during the flyby. All of the reference stars are in the range 12(exp h) 40(exp m) to 14(exp h) 12(exp m) in right ascension (1950) and +02 to -09 degs in declination. Mean errors of the positions are about 0.25 sec.

  12. IFLA General Conference, 1986. Bibliographic Control Division. Section: Cataloguing. Papers.

    ERIC Educational Resources Information Center

    International Federation of Library Associations and Institutions, The Hague (Netherlands).

    Papers on cataloging which were presented at the 1986 International Federation of Library Associations (IFLA) conference include: (1) "Cataloging of Government Documents in the Age of Automation" (Chong Y. Yoon, United States), which discusses the use of MARC (Machine-Readable Cataloging) formats to integrate government documents into…

  13. Planning for the Automation of School Library Media Centers.

    ERIC Educational Resources Information Center

    Caffarella, Edward P.

    1996-01-01

    Geared for school library media specialists whose centers are in the early stages of automation or conversion to a new system, this article focuses on major components of media center automation: circulation control; online public access catalogs; machine readable cataloging; retrospective conversion of print catalog cards; and computer networks…

  14. Procedures and Policies Manual

    ERIC Educational Resources Information Center

    Davis, Jane M.

    2006-01-01

    This document was developed by the Middle Tennessee State University James E. Walker Library Collection Management Department to provide policies and procedural guidelines for the cataloging and processing of bibliographic materials. This document includes policies for cataloging monographs, serials, government documents, machine-readable data…

  15. Demonstration of Cataloging Support Services and Marc II Conversion. Final Report.

    ERIC Educational Resources Information Center

    Buckland, Lawrence F.; And Others

    Beginning in December, 1967, the New England Library Information Network (NELINET) was demonstrated in actual operation using Machine-Readable Cataloging (MARC I) bibliographic data. Section 1 of this report is an introduction and summary of the project. Section 2 described the library processing function demonstrated which included catalog card…

  16. Documentation for the machine-readable version of the Smithsonian Astrophysical Observatory Star catalogue (SAO) version 1984

    NASA Technical Reports Server (NTRS)

    Roman, N. G.; Warren, W. H., Jr.

    1984-01-01

    An updated, corrected and extended machine readable version of the Smithsonian Astrophysical Observatory star catalog (SAO) is described. Published and unpublished errors discovered in the previous version have been corrected, and multiple star and supplemental BD identifications added to stars where more than one SAO entry has the same Durchmusterung number. Henry Draper Extension (HDE) numbers have been added for stars found in both volumes of the extension. Data for duplicate SAO entries (those referring to the same star) have been blanked out, but the records themselves have been retained and flagged so that sequencing and record count are identical to the published catalog.

  17. Lick Jupiter-Voyager reference star catalogue (Klemola, Morabito, and Taraji 1978): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The catalog contains accurate equatorial coordinates for 4983 stars in a band of sky against which cameras of the Voyager spacecraft were pointed for observations in the region of Jupiter during the flyby. All of the reference stars are in the range 6 hr 00 min to 8 hr 04 min in right ascension (1950), declination zones +16 to +23 degrees, and 8 hr to 31 min to 8 hr 57 min, zones +08 to +14 degrees. Mean errors of the positions are about 0.4 sec.

  18. A Mechanized Information Services Catalog.

    ERIC Educational Resources Information Center

    Marron, Beatrice; And Others

    The National Bureau of Standards is mechanizing a catalog of currently available information sources and services. Information from recent surveys of machine-readable, commercially-available bibliographic data bases, and the various current awareness, batch retrospective, and interactive retrospective services which can access them have been…

  19. Turning Archival Tapes into an Online “Cardless” Catalog

    PubMed Central

    Zuckerman, Alan E.; Ewens, Wilma A.; Cannard, Bonnie G.; Broering, Naomi C.

    1982-01-01

    Georgetown University has created an online card catalog based on machine readable cataloging records (MARC) loaded from archival tapes or online via the OCLC network. The system is programmed in MUMPS and uses the medical subject headings (MeSH) authority file created by the National Library of Medicine. The online catalog may be searched directly by library users and has eliminated the need for manual filing of catalog cards.

  20. Documentation for the machine-readable version of the general catalogue of trigonometric stellar parallaxes and supplement

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of the General Catalog of Trigonometric Stellar parallaxes as distributed by the Astronomical Data Center is described. It is intended to enable users to read and process the data without problems and guesswork. The source reference should be consulted for details concerning the compilation of the main catalogue and supplement, the probable errors, and the weighting system used to combine determinations from different observatories.

  1. A Feasibility Study on Data Distribution on Optical Media.

    ERIC Educational Resources Information Center

    Campbell (Bonnie) & Associates, Toronto (Ontario).

    This feasibility study assesses the potential of optical technology in the development of accessible bibliographic and location data networks both in Canada and within the international MARC (Machine-Readable Cataloging) network. The study is divided into four parts: (1) a market survey of cataloging and interlibrary loan librarians to determine…

  2. Documentation for the machine-readable version of a supplement to the Bright Star catalogue (Hoffleit, Saladyga and Wlasuk 1983)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    Detailed descriptions of the three files of the machine-readable catalog are given. The files of the original tape have been restructured and the data records reformatted to produce a uniform data file having a single logical record per star and homogeneous data fields. The characteristics of the tape version as it is presently being distributed from the Astronomical Data Center are given and the changes to the original tape supplied are described.

  3. A compilation of redshifts and velocity dispersions for Abell clusters (Struble and Rood 1987): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1989-01-01

    The machine readable version of the compilation, as it is currently being distributed from the Astronomical Data Center, is described. The catalog contains redshifts and velocity dispersions for all Abell clusters for which these data had been published up to 1986 July. Also included are 1950 equatorial coordinates for the centers of the listed clusters, numbers of observations used to determine the redshifts, and bibliographical references citing the data sources.

  4. Astronomical Data Center Bulletin, volume 1, number 3

    NASA Technical Reports Server (NTRS)

    Mead, J. M.; Warren, W. H., Jr.; Nagy, T. A.

    1983-01-01

    A catalog of galactic O-type stars, a machine-readable version of the bright star catalog, a two-micron sky survey, sky survey sources with problematical Durchmusterung identifications, data retrieval for visual binary stars, faint blue objects, the sixth catalog of galactic Wolf-Rayet stars, declination versus magnitude distribution, the SAO-HD-GC-DM cross index catalog, star cross-identification tables, astronomical sources, bibliographical star index search updates, DO-HD and HD-DO cross indices, and catalogs, are reviewed.

  5. Guidelines for Processing and Cataloging Computer Software for Schools and Area Education Agencies. Suggestions to Aid Schools and AEAs.

    ERIC Educational Resources Information Center

    Martin, Elizabeth; And Others

    Based on definitions of a machine-readable data file (MRDF) taken from the Anglo-American Cataloging Rules, second edition (AACR2) and Standards for Cataloging Nonprint Materials, the following recommendations for processing items of computer software are provided: (1) base main and added entry determination on AACR2; (2) place designation of form…

  6. Catalog of SAS-2 gamma-ray observations (Fichtel, et al. 1990)

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.

    1990-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The SAS-2 gamma ray catalog contains fluxes measured with the high energy gamma ray telescope flown aboard the second NASA Small Astronomy Satellite. The objects measured include various types of galaxies, quasi-stellar, and BL Lacertae objects, and pulsars. The catalog contains separate files for galaxies, pulsars, other objects, notes, and references.

  7. Collection Development Analysis Using OCLC Archival Tapes. Final Report.

    ERIC Educational Resources Information Center

    Evans, Glyn T.; And Others

    The purpose of this project is to develop a set of computer programs to perform a variety of collection development analyses on the machine-readable cataloging (MARC) records that are produced as a byproduct of use of the online cataloging subsystem of the Ohio College Library System (OCLC), and made available through the OCLC Distribution Tape…

  8. Documentation for the machine-readable version of a deep objective-prism survey for large Magellanic cloud members

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    This catalog contains 1273 proven or probable Large Magellanic Cloud (LMC) members, as found on deep objective-prism plates taken with the Curtis Schmidt telescope at Cerro Tololo Inter-American Observatory in Chile. The stars are generally brighter than about photographic magnitude 14. Approximate spectral types were determined by examination of the 580 A/mm objective-prism spectra; approximate 1975 positions were obtained by measuring relative to the 1975 coordinate grids on the Uppsala-Mount Stromlo Atlas of the LMC (Gascoigne and Westerlund 1961), and approximate photographic magnitudes were determined by averaging image density measures from the plates and image-diameter measures on the 'B' charts. The machine-readable version of the LMC survey catalog is described to enable users to read and process the tape file without problems or guesswork.

  9. Catalog of infrared observations

    NASA Technical Reports Server (NTRS)

    Gezari, D. Y.; Schmitz, M.; Mead, J. M.

    1982-01-01

    The infrared astronomical data base and its principal data product, the catalog of Infrared Observations (CIO), comprise a machine readable library of infrared (1 microns to 1000 microns astronomical observations. To date, over 1300 journal articles and 10 major survey catalogs are included in this data base, which contains about 55,000 individual observations of about 10,000 different infrared sources. Of these, some 8,000 sources are identifiable with visible objects, and about 2,000 do not have known visible counterparts.

  10. Investigation of the Searching Efficiency and Cost of Creating a Remote Access Catalog for the New York State Library. Final Report.

    ERIC Educational Resources Information Center

    Buckland, Lawrence F.; Madden, Mary

    From experimental work performed, and reported upon in this document, it is concluded that converting the New York State Library (NYSL) shelf list sample to machine readable form, and searching this shelf list using a remote access catalog are technically sound concepts though the capital costs of data conversion and system installation will be…

  11. NASA climate data catalog

    NASA Technical Reports Server (NTRS)

    Reph, M. G.

    1984-01-01

    This document provides a summary of information available in the NASA Climate Data Catalog. The catalog provides scientific users with technical information about selected climate parameter data sets and the associated sensor measurements from which they are derived. It is an integral part of the Pilot Climate Data System (PCDS), an interactive, scientific management system for locating, obtaining, manipulating, and displaying climate research data. The catalog is maintained in a machine readable representation which can easily be accessed via the PCDS. The purposes, format and content of the catalog are discussed. Summarized information is provided about each of the data sets currently described in the catalog. Sample detailed descriptions are included for individual data sets or families of related data sets.

  12. Astronomical catalog desk reference, 1994 edition

    NASA Technical Reports Server (NTRS)

    1994-01-01

    The Astronomical Catalog Desk Reference is designed to aid astronomers in locating machine readable catalogs in the Astronomical Data Center (ADC) archives. The key reference components of this document are as follows: A listing of shortened titles for all catalogs available from the ADC (includes the name of the lead author and year of publication), brief descriptions of over 300 astronomical catalogs, an index of ADC catalog numbers by subject keyword, and an index of ADC catalog numbers by author. The heart of this document is the set of brief descriptions generated by the ADC staff. The 1994 edition of the Astronomical Catalog Desk Reference contains descriptions for over one third of the catalogs in the ADC archives. Readers are encouraged to refer to this section for concise summaries of those catalogs and their contents.

  13. Beyond the online catalog: developing an academic information system in the sciences.

    PubMed Central

    Crawford, S; Halbrook, B; Kelly, E; Stucki, L

    1987-01-01

    The online public access catalog consists essentially of a machine-readable database with network capabilities. Like other computer-based information systems, it may be continuously enhanced by the addition of new capabilities and databases. It may also become a gateway to other information networks. This paper reports the evolution of the Bibliographic Access and Control System (BACS) of Washington University in end-user searching, current awareness services, information management, and administrative functions. Ongoing research and development and the future of the online catalog are also discussed. PMID:3315052

  14. Beyond the online catalog: developing an academic information system in the sciences.

    PubMed

    Crawford, S; Halbrook, B; Kelly, E; Stucki, L

    1987-07-01

    The online public access catalog consists essentially of a machine-readable database with network capabilities. Like other computer-based information systems, it may be continuously enhanced by the addition of new capabilities and databases. It may also become a gateway to other information networks. This paper reports the evolution of the Bibliographic Access and Control System (BACS) of Washington University in end-user searching, current awareness services, information management, and administrative functions. Ongoing research and development and the future of the online catalog are also discussed.

  15. Bonner Durchmusterung (Argelander 1859-1862): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.; Ochsenbein, Francois

    1989-01-01

    The machine-readable version of the catalog, as it is currently being distributed from the Astronomical Data Center, is described. The entire Bonner Durchmusterung (BD) was computerized through the collaborative efforts of the Centre de Donnees Astronomiques de Strasbourg, l'Observatoire de Nice, and the Astronomical Data Center at the NASA/Goddard Space Flight Center. All corrigenda published in the original BD volumes were incorporated into the machine file, along with changes published following the 1903 edition. In addition, stars indicated to be missing in published lists and verified by various techniques are flagged so that they can be omitted from computer plotted charts if desired. Stars deleted in the various errata lists were similarly flagged, while those with revised data are flagged and listed in a separate table.

  16. The Role of the National Agricultural Library.

    ERIC Educational Resources Information Center

    Howard, Joseph H.

    1989-01-01

    Describes the role, users, collections and services of the National Agricultural Library. Some of the services discussed include a machine readable bibliographic database, an international interlibrary loan system, programs to develop library networks and cooperative cataloging, and the development and use of information technologies such as laser…

  17. Comparing Characteristics of Highly Circulated Titles for Demand-Driven Collection Development.

    ERIC Educational Resources Information Center

    Britten, William A; Webster, Judith D.

    1992-01-01

    Describes methodology for analyzing MARC (machine-readable cataloging) records of highly circulating titles to document common characteristics for collection development purposes. Application of the methodology in a university library is discussed, and data are presented on commonality of subject heading, author, language, and imprint date for…

  18. IFLA General Conference, 1987. IFLA Core Programmes. Open Forum. Papers.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    The four papers in this compilation report on some of the recent core programs of the International Federation of Library Associations (IFLA): (1) "The IFLA Universal Bibliographic Control and International Machine Readable Cataloging Programme (UBCIM)" (Ross Bourne, IFLA UBCIM Programme Officer); (2) "The IFLA UAP (Universal…

  19. Issues in Retrospective Conversion for a Small Special Collection: A Case Study.

    ERIC Educational Resources Information Center

    Hieb, Fern

    1997-01-01

    Small special collections present unique problems for retrospective conversion of catalogs to machine-readable form. Examines retrospective conversion using the Moravian Music Foundation as a case study. Discusses advantages to automation, options for conversion process, quantifying conversion effort, costs, in-house conversion, national standards…

  20. NASA STI Program Coordinating Council Twelfth Meeting: Standards

    NASA Technical Reports Server (NTRS)

    1994-01-01

    The theme of this NASA Scientific and Technical Information Program Coordinating Council Meeting was standards and their formation and application. Topics covered included scientific and technical information architecture, the Open Systems Interconnection Transmission Control Protocol/Internet Protocol, Machine-Readable Cataloging (MARC) open system environment procurement, and the Government Information Locator Service.

  1. Retrieval techniques and graphics displays using a computerized stellar data base

    NASA Technical Reports Server (NTRS)

    Mead, J.; Nagy, T. A.

    1977-01-01

    The paper describes a stellar data retrieval system for which the data base consists of 28 machine-readable astronomical catalogs. Eleven of these catalogs have been combined into the Goddard Cross Index (GCI), which serves as the computer entry point to these catalogs. The full data entry from any of the GCI catalogs can be retrieved in a single computer run. With this system, it is possible to prepare candidates for observation by searching the data base for stars with given characteristics. Generation of plots of all catalog stars in or near the telescope's field of view to scale of Palomar, other atlases, or to the telescope itself for use as observing charts or to aid in identifying unknown sources, can be accomplished.

  2. Assessing Information on the Internet: Toward Providing Library Services for Computer-Mediated Communication. A Final Report.

    ERIC Educational Resources Information Center

    Dillon, Martin; And Others

    The Online Computer Library Center Internet Resource project focused on the nature of electronic textual information available through remote access using the Internet and the problems associated with creating machine-readable cataloging (MARC) records for these objects using current USMARC format for computer files and "Anglo-American…

  3. IFLA General Conference, 1989. Introduction to IFLA's Core Programmes; Contributed Papers; Plenary Session Papers. Booklet 00.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    This collection contains three papers providing an introduction to the International Federation of Library Associations (IFLA) Core Programs, four contributed papers, and two Plenary Session papers: (1) "The Universal Bibliographic Control and International MARC (Machine Readable Cataloging Formats) Program" (Winston D. Roberts); (2) "Le Programme…

  4. Retrospective Conversion at a Two-Year College.

    ERIC Educational Resources Information Center

    Krieger, Michael T.

    1982-01-01

    Findings of a project to convert a single LC class from cards to machine readable tapes at a two-year college suggest that an in-house retrospective conversion is feasible for academic libraries. A high conversion hit rate, implying minimal original cataloging, will keep project costs and duration low. There are five references. (RAA)

  5. Vatican Library Automates for the 21st Century.

    ERIC Educational Resources Information Center

    Carter, Thomas L.

    1994-01-01

    Because of space and staff constraints, the Vatican Library can issue only 2,000 reader cards a year. Describes IBM's Vatican Library Project: converting the library catalog records (prior to 1985) into machine readable form and digitally scanning 20,000 manuscript pages, print pages, and art works in gray scale and color, creating a database…

  6. VizieR Online Data Catalog: Parenago Catalog of Stars in Orion Nebula (Parenago 1954)

    NASA Astrophysics Data System (ADS)

    Parenago, P. P.

    1997-10-01

    The present catalogue is a machine-readable version of the catalogue of stars in the area of the Orion nebula, published by P.P. Parenago (1954). The sky area between 5h 24m and 5h 36m in right ascension (1900.0) and between -4 and -7 degrees in declination (1900.0), containing the Orion nebula, has been investigated in that work. Ten of variable stars in original Parenago (1954) catalogue had CSV numbers (Kukarkin et al., 1951) but since that time all of them were confirmed as variables and included in GCVS (Kholopov et al., 1985a&b, 1987). We superseded CSV-numbers by GCVS-names in the machine-readable version for the following stars: ------------------------------------------------ Number in CSV-number GCVS-name the catalogue ------------------------------------------------ 1605 606 V372 ORI 1613 607 V373 ORI 1635 608 V374 ORI 1713 609 V375 ORI 1748 610 V387 ORI 1762 100569 V376 ORI 1974 617 V377 ORI 2183 625 V388 ORI 2393 630 V380 ORI 2478 634 V381 ORI ------------------------------------------------ (1 data file).

  7. The astronomical data base and retrieval system at NASA

    NASA Technical Reports Server (NTRS)

    Mead, J. M.; Nagy, T. A.; Hill, R. S.; Warren, W. H., Jr.

    1982-01-01

    More than 250 machine-readable catalogs of stars and extended celestial objects are now available at the NASA/Goddard Space Flight Center (GSFC) as the result of over a decade of catalog acquisition, verification and documentation. Retrieval programs are described which permit the user to obtain from a remote terminal bibliographical listings for stars; to find all celestial objects from a given list that are within a defined angular separation from each object in another list; to plot celestial objects on overlays for sky survey plate areas; and to search selected catalogs for objects by criteria of position, identification number, magnitude or spectral type.

  8. California State Library: Processing Center Design and Specifications. Volume III, Coding Manual.

    ERIC Educational Resources Information Center

    Sherman, Don; Shoffner, Ralph M.

    As part of the report on the California State Library Processing Center design and specifications, this volume is a coding manual for the conversion of catalog card data to a machine-readable form. The form is compatible with the national MARC system, while at the same time it contains provisions for problems peculiar to the local situation. This…

  9. VizieR Online Data Catalog: GB6 catalog of radio sources (Gregory+ 1996)

    NASA Astrophysics Data System (ADS)

    Gregory, P. C.; Scott, W. K.; Douglas, K.; Condon, J. J.

    1997-01-01

    The final set of sky maps from the Green Bank 4.85 GHz survey (Condon J.J., Broderick J.J., Seielstad G.A., Douglas K., & Gregory P.C. in 1994AJ....107.1829C) was used to construct the GB6 Catalog of sources stronger than S ~ 18 mJy in the declination range 0deg < Dec. < +75deg (Gregory P.C., Scott W.K., Douglas K., & Condon J.J. in 1996ApJS..103..427G). There are two machine-readable versions of the GB6 catalog, with coordinates precessed to B1950 (file b1950.dat) and J2000 (file j2000.dat). Each catalog file contains one line per source (75,162 lines each), and the sources are sorted by increasing B1950 or J2000 right ascension. (3 data files).

  10. Washington Double Star Catalog Cross Index (1950 position sort)

    NASA Technical Reports Server (NTRS)

    1993-01-01

    A machine-readable version of the Washington Catalog of Visual Double Stars (WDS) was prepared in 1984 on the basis of a data file that was collected and maintained for more than a century by a succession of double-star observers. Although this catalog is being continually updated, a new copy for distribution is not expected to be available for a few years. The WDS contains DM numbers, but many of these are listed only in the notes, which makes it difficult to search for double-star information, except by position. Hence, a cross index that provides complete DM identifications is desirable, and it appears useful to add HD numbers for systems in that catalog. Aitken Double Star (ADS) numbers were retained from the WDS, but no attempt was made to correct these except for obvious errors.

  11. Design of an Automated Library Information Storage and Retrieval System for Library of Congress Division for the Blind and Physically Handicapped (DBPH). Final Report.

    ERIC Educational Resources Information Center

    Systems Architects, Inc., Randolph, MA.

    A practical system for producing a union catalog of titles in the collections of the Library of Congress Division for the Blind and Physically Handicapped (DBPH), its regional network, and related agencies from a machine-readable data base is presented. The DBPH organization and operations and the associated regional library network are analyzed.…

  12. Government documents and the online catalog.

    PubMed

    Lynch, F H; Lasater, M C

    1990-01-01

    Prior to planning for implementing the NOTIS system, the Vanderbilt Medical Center Library had not fully cataloged its government publications, and records for these materials were not in machine-readable format. A decision was made that patrons should need to look in only one place for all library materials, including the Health and Human Services Department publications received each year from the central library's Government Documents Unit. Beginning in 1985, these publications were added to the library's database, and the entire 7,200-piece collection is now in the online catalog. Working with these publications has taught the library much about the advantages and disadvantages of cataloging government documents in an online environment. It was found that OCLC cataloging copy is eventually available for most titles, although only about 10% of the records have MeSH headings. Staff time is the major expenditure; problems are caused by documents' irregular nature, frequent format changes, and difficult authority work. Since their addition to the online catalog, documents are used more and the library has better control.

  13. Government documents and the online catalog.

    PubMed Central

    Lynch, F H; Lasater, M C

    1990-01-01

    Prior to planning for implementing the NOTIS system, the Vanderbilt Medical Center Library had not fully cataloged its government publications, and records for these materials were not in machine-readable format. A decision was made that patrons should need to look in only one place for all library materials, including the Health and Human Services Department publications received each year from the central library's Government Documents Unit. Beginning in 1985, these publications were added to the library's database, and the entire 7,200-piece collection is now in the online catalog. Working with these publications has taught the library much about the advantages and disadvantages of cataloging government documents in an online environment. It was found that OCLC cataloging copy is eventually available for most titles, although only about 10% of the records have MeSH headings. Staff time is the major expenditure; problems are caused by documents' irregular nature, frequent format changes, and difficult authority work. Since their addition to the online catalog, documents are used more and the library has better control. PMID:2295010

  14. Documentation for the machine-readable version of photometric data for nearby stars

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    A computer list of all photometric systems (of those considered), in which each star was measured is provided. The file is a subset of a much larger and more comprehensive compilation, which lists all measured photoelectric photometric systems for any star that has been measured in at least one photoelectric system. In addition to the photometric system identifications, cross identifications to the Henry Draper and Durchmusterung catalogs and apparent visual magnitudes are included.

  15. Systems Design and Pilot Operation of a Regional Center for Technical Processing for the Libraries of the New England State Universities. NELINET, New England Library Information Network. Progress Report, July 1, 1967 - March 30, 1968, Volume II, Appendices.

    ERIC Educational Resources Information Center

    Agenbroad, James E.; And Others

    Included in this volume of appendices to LI 000 979 are acquisitions flow charts; a current operations questionnaire; an algorithm for splitting the Library of Congress call number; analysis of the Machine-Readable Cataloging (MARC II) format; production problems and decisions; operating procedures for information transmittal in the New England…

  16. 77 FR 7 - Revisions to Labeling Requirements for Blood and Blood Components, Including Source Plasma

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-01-03

    ... requirements will facilitate the use of a labeling system using machine-readable information that would be... components. Furthermore, we proposed the use of a labeling system using machine-readable information that...; Facilitates the use of a labeling system using machine- readable information that would be acceptable as a...

  17. 6 CFR 37.19 - Machine readable technology on the driver's license or identification card.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ..., States must use the ISO/IEC 15438:2006(E) Information Technology—Automatic identification and data... 6 Domestic Security 1 2011-01-01 2011-01-01 false Machine readable technology on the driver's..., Verification, and Card Issuance Requirements § 37.19 Machine readable technology on the driver's license or...

  18. 6 CFR 37.19 - Machine readable technology on the driver's license or identification card.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ..., States must use the ISO/IEC 15438:2006(E) Information Technology—Automatic identification and data... 6 Domestic Security 1 2010-01-01 2010-01-01 false Machine readable technology on the driver's..., Verification, and Card Issuance Requirements § 37.19 Machine readable technology on the driver's license or...

  19. Does Machine-Readable Documentation on Online Hosts and CD-ROMs Have a Role or Future?

    ERIC Educational Resources Information Center

    Harris, Stephen; Oppenheim, Charles

    1996-01-01

    Reports results of a United Kingdom-based mail survey of database users, CD-ROM producers, and hosts to assess trends and views concerning documentation in machine-readable form. Cost, convenience, and ease of use of print manuals are cited as reasons for the reluctance to switch to machine-readable documentation. Sample surveys are included.…

  20. VizieR Online Data Catalog: Zwicky Galaxy Catalog (Zwicky+ 1968)

    NASA Astrophysics Data System (ADS)

    Zwicky, F.; et al.

    1996-03-01

    This document describes a computer version of that part of the CGCG (Zwicky et al. 1961-68) containing all the alphanumeric information for galaxies. All known errors found by Zwicky and many others are corrected as well as erroneous quotations from other catalogs (Shapley & Ames 1932, Bigay 1951, Pettit 1954, Humason et al. 1956, Holmberg 1958). It is an illusion to consider all the errors are found. There are some misprints even in the most extended list of misprints (Paturel et al. 1991). We have compiled two files: zwigal.ori and zwigal.add. The first one contains the original information from CGCG for galaxies. The second one contains the data from above mentioned other catalogs given in CGCG. We have made no attempts to supply the catalog with any new information. A detailed comparison with the machine-readable version of Zwicky galaxies prepared by R.S. Hill (NSSDC ADC #7049 or CDS VII/49) was performed. Our version contains more data on individual galaxies - designation, description, magnitudes, velocity. All galaxies in the Coma center are included. However Hill's version contains data for Zwicky fields, Palomar Sky Survey plate number as well as Mead-Luyten-Palomar number. There are 27837 different galaxies and 29418 entries in CGCG. (2 data files).

  1. Department of Defense Logistics Roadmap 2008. Volume 1

    DTIC Science & Technology

    2008-07-01

    machine readable identification mark on the Department’s tangible qualifying assets, and establishes the data management protocols needed to...uniquely identify items with a Unique Item Identifier (UII) via machine - readable information (MRI) marking represented by a two-dimensional data...property items with a machine -readable Unique Item Identifier (UII), which is a set of globally unique data elements. The UII is used in functional

  2. A Catalog of Cool Dwarf Targets for the Transiting Exoplanet Survey Satellite

    NASA Astrophysics Data System (ADS)

    Muirhead, Philip S.; Dressing, Courtney D.; Mann, Andrew W.; Rojas-Ayala, Bárbara; Lépine, Sébastien; Paegert, Martin; De Lee, Nathan; Oelkers, Ryan

    2018-04-01

    We present a catalog of cool dwarf targets (V-J> 2.7, T eff ≲ 4000 K) and their stellar properties for the upcoming Transiting Exoplanet Survey Satellite (TESS), for the purpose of determining which cool dwarfs should be observed using two minute observations. TESS has the opportunity to search tens of thousands of nearby, cool, late K- and M-type dwarfs for transiting exoplanets, an order of magnitude more than current or previous transiting exoplanet surveys, such as Kepler, K2, and ground-based programs. This necessitates a new approach to choosing cool dwarf targets. Cool dwarfs are chosen by collating parallax and proper motion catalogs from the literature and subjecting them to a variety of selection criteria. We calculate stellar parameters and TESS magnitudes using the best possible relations from the literature while maintaining uniformity of methods for the sake of reproducibility. We estimate the expected planet yield from TESS observations using statistical results from the Kepler mission, and use these results to choose the best targets for two minute observations, optimizing for small planets for which masses can conceivably be measured using follow-up Doppler spectroscopy by current and future Doppler spectrometers. The catalog is available in machine readable format and is incorporated into the TESS Input Catalog and TESS Candidate Target List until a more complete and accurate cool dwarf catalog identified by ESA’s Gaia mission can be incorporated.

  3. Documentation for the machine-readable version of the Morphological Catalogue of Galaxies (MCG) of Vorontsov-Velyaminov et al, 1962-1968

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    Modifications, corrections, and the record format are provided for the machine-readable version of the "Morphological Catalogue of Galaxies.' In addition to hundreds of individual corrections, a detailed comparison of the machine-readable with the published catalogue resulted in the addition of 116 missing objects, the deletion of 10 duplicate records, and a format modification to increase storage efficiency.

  4. Survey of Commercially Available Computer-Readable Bibliographic Data Bases.

    ERIC Educational Resources Information Center

    Schneider, John H., Ed.; And Others

    This document contains the results of a survey of 94 U. S. organizations, and 36 organizations in other countries that were thought to prepare machine-readable data bases. Of those surveyed, 55 organizations (40 in U. S., 15 in other countries) provided completed camera-ready forms describing 81 commercially available, machine-readable data bases…

  5. An Open Catalog for Supernova Data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Guillochon, James; Parrent, Jerod; Kelley, Luke Zoltan

    We present the Open Supernova Catalog , an online collection of observations and metadata for presently 36,000+ supernovae and related candidates. The catalog is freely available on the web (https://sne.space), with its main interface having been designed to be a user-friendly, rapidly searchable table accessible on desktop and mobile devices. In addition to the primary catalog table containing supernova metadata, an individual page is generated for each supernova, which displays its available metadata, light curves, and spectra spanning X-ray to radio frequencies. The data presented in the catalog is automatically rebuilt on a daily basis and is constructed by parsingmore » several dozen sources, including the data presented in the supernova literature and from secondary sources such as other web-based catalogs. Individual supernova data is stored in the hierarchical, human- and machine-readable JSON format, with the entirety of each supernova’s data being contained within a single JSON file bearing its name. The setup we present here, which is based on open-source software maintained via git repositories hosted on github, enables anyone to download the entirety of the supernova data set to their home computer in minutes, and to make contributions of their own data back to the catalog via git. As the supernova data set continues to grow, especially in the upcoming era of all-sky synoptic telescopes, which will increase the total number of events by orders of magnitude, we hope that the catalog we have designed will be a valuable tool for the community to analyze both historical and contemporary supernovae.« less

  6. An Open Catalog for Supernova Data

    NASA Astrophysics Data System (ADS)

    Guillochon, James; Parrent, Jerod; Kelley, Luke Zoltan; Margutti, Raffaella

    2017-01-01

    We present the Open Supernova Catalog, an online collection of observations and metadata for presently 36,000+ supernovae and related candidates. The catalog is freely available on the web (https://sne.space), with its main interface having been designed to be a user-friendly, rapidly searchable table accessible on desktop and mobile devices. In addition to the primary catalog table containing supernova metadata, an individual page is generated for each supernova, which displays its available metadata, light curves, and spectra spanning X-ray to radio frequencies. The data presented in the catalog is automatically rebuilt on a daily basis and is constructed by parsing several dozen sources, including the data presented in the supernova literature and from secondary sources such as other web-based catalogs. Individual supernova data is stored in the hierarchical, human- and machine-readable JSON format, with the entirety of each supernova’s data being contained within a single JSON file bearing its name. The setup we present here, which is based on open-source software maintained via git repositories hosted on github, enables anyone to download the entirety of the supernova data set to their home computer in minutes, and to make contributions of their own data back to the catalog via git. As the supernova data set continues to grow, especially in the upcoming era of all-sky synoptic telescopes, which will increase the total number of events by orders of magnitude, we hope that the catalog we have designed will be a valuable tool for the community to analyze both historical and contemporary supernovae.

  7. A survey of machine readable data bases

    NASA Technical Reports Server (NTRS)

    Matlock, P.

    1981-01-01

    Forty-two of the machine readable data bases available to the technologist and researcher in the natural sciences and engineering are described and compared with the data bases and date base services offered by NASA.

  8. VizieR Online Data Catalog: Wisconsin soft X-ray diffuse background all-sky Survey (McCammon+ 1983)

    NASA Astrophysics Data System (ADS)

    McCammon, D.; Burrows, D. N.; Sanders, W. T.; Kraushaar, W. L.

    1997-10-01

    The catalog contains all-sky survey of the soft X-ray diffuse background and the count-rate data from which the maps were made for the ten flights included in the survey. It contains 40 files in the machine-readable version and includes documentation and utility subroutines. The data files contain different band maps (B, C, M, M1, M2, I, J, 2-6 keV) in a 0 degree-centered Aitoff projection, in a 180-degree-centered Aitoff projection, in a north polar projection, and in a south polar projection. Lookup tables in the form of FITS images are provided for conversion between pixel coordinates and Galactic coordinates for the various projections. The bands are: B = 130-188eV C = 160-284eV M1 = 440-930eV M2 = 600-1100eV I = 770-1500eV J = 1100-2200eV 2-6keV = 1800-6300eV (51 data files).

  9. The Exchange of Bibliographic Data in Non-Roman Scripts.

    ERIC Educational Resources Information Center

    Wellisch, Hans H.

    1980-01-01

    Advocates the use of machine readable codes to accomplish romanization and promote the exchange of bibliographic data. Proposals are presented for transliteration standards, design of machine readable conversion codes, and the establishment of databases. (RAA)

  10. Documentation for the machine-readable version of the revised new general catalogue of nonstellar astronomical objects

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The contents and format of the machine-readable version of the cataloque distributed by the Astronomical Data Center are described. Coding for the various scales and abbreviations used in the catalogue are tabulated and certain revisions to the machine version made to improve storage efficiency and notation are discussed.

  11. A Machine Learning Approach to Measurement of Text Readability for EFL Learners Using Various Linguistic Features

    ERIC Educational Resources Information Center

    Kotani, Katsunori; Yoshimi, Takehiko; Isahara, Hitoshi

    2011-01-01

    The present paper introduces and evaluates a readability measurement method designed for learners of EFL (English as a foreign language). The proposed readability measurement method (a regression model) estimates the text readability based on linguistic features, such as lexical, syntactic and discourse features. Text readability refers to the…

  12. Constructing and validating readability models: the method of integrating multilevel linguistic features with machine learning.

    PubMed

    Sung, Yao-Ting; Chen, Ju-Ling; Cha, Ji-Her; Tseng, Hou-Chiang; Chang, Tao-Hsing; Chang, Kuo-En

    2015-06-01

    Multilevel linguistic features have been proposed for discourse analysis, but there have been few applications of multilevel linguistic features to readability models and also few validations of such models. Most traditional readability formulae are based on generalized linear models (GLMs; e.g., discriminant analysis and multiple regression), but these models have to comply with certain statistical assumptions about data properties and include all of the data in formulae construction without pruning the outliers in advance. The use of such readability formulae tends to produce a low text classification accuracy, while using a support vector machine (SVM) in machine learning can enhance the classification outcome. The present study constructed readability models by integrating multilevel linguistic features with SVM, which is more appropriate for text classification. Taking the Chinese language as an example, this study developed 31 linguistic features as the predicting variables at the word, semantic, syntax, and cohesion levels, with grade levels of texts as the criterion variable. The study compared four types of readability models by integrating unilevel and multilevel linguistic features with GLMs and an SVM. The results indicate that adopting a multilevel approach in readability analysis provides a better representation of the complexities of both texts and the reading comprehension process.

  13. Ownership of Machine-Readable Bibliographic Data. Canadian Network Papers Number 5 = Propriete des Donnees Bibliographique Lisibles par Machine. Documents sur les Resaux Canadiens Numero 5.

    ERIC Educational Resources Information Center

    Duchesne, R. M.; And Others

    Because of data ownership questions raised by the interchange and sharing of machine readable bibliographic data, this paper was prepared for the Bibliographic and Communications Network Committee of the National Library Advisory Board. Background information and definitions are followed by a review of the legal aspects relating to property and…

  14. 48 CFR 252.211-7003 - Item unique identification and valuation.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... reader or interrogator, used to retrieve data encoded on machine-readable media. Concatenated unique item... identifier. Item means a single hardware article or a single unit formed by a grouping of subassemblies... manufactured under identical conditions. Machine-readable means an automatic identification technology media...

  15. Reference Manual for Machine-Readable Bibliographic Descriptions. Second Revised Edition.

    ERIC Educational Resources Information Center

    Dierickx, H., Ed.; Hopkinson, A., Ed.

    A product of the UNISIST International Centre for Bibliographic Descriptions (UNIBIB), this reference manual presents a standardized communication format for the exchange of machine-readable bibliographic information between bibliographic databases or other types of bibliographic information services, including libraries. The manual is produced in…

  16. BLS Machine-Readable Data and Tabulating Routines.

    ERIC Educational Resources Information Center

    DiFillipo, Tony

    This report describes the machine-readable data and tabulating routines that the Bureau of Labor Statistics (BLS) is prepared to distribute. An introduction discusses the LABSTAT (Labor Statistics) database and the BLS policy on release of unpublished data. Descriptions summarizing data stored in 25 files follow this format: overview, data…

  17. Banknotes and unattended cash transactions

    NASA Astrophysics Data System (ADS)

    Bernardini, Ronald R.

    2000-04-01

    There is a 64 billion dollar annual unattended cash transaction business in the US with 10 to 20 million daily transactions. Even small problems with the machine readability of banknotes can quickly become a major problem to the machine manufacturer and consumer. Traditional note designs incorporate overt security features for visual validation by the public. Many of these features such as fine line engraving, microprinting and watermarks are unsuitable as machine readable features in low cost note acceptors. Current machine readable features, mostly covert, were designed and implemented with the central banks in mind. These features are only usable by the banks large, high speed currency sorting and validation equipment. New note designs should consider and provide for low cost not acceptors, implementing features developed for inexpensive sensing technologies. Machine readable features are only as good as their consistency. Quality of security features as well as that of the overall printing process must be maintained to ensure reliable and secure operation of note readers. Variations in printing and of the components used to make the note are one of the major causes of poor performance in low cost note acceptors. The involvement of machine manufacturers in new currency designs will aid note producers in the design of a note that is machine friendly, helping to secure the acceptance of the note by the public as well as acting asa deterrent to fraud.

  18. A Study of Readability of Texts in Bangla through Machine Learning Approaches

    ERIC Educational Resources Information Center

    Sinha, Manjira; Basu, Anupam

    2016-01-01

    In this work, we have investigated text readability in Bangla language. Text readability is an indicator of the suitability of a given document with respect to a target reader group. Therefore, text readability has huge impact on educational content preparation. The advances in the field of natural language processing have enabled the automatic…

  19. Financial Statistics. Higher Education General Information Survey (HEGIS) [machine-readable data file].

    ERIC Educational Resources Information Center

    Center for Education Statistics (ED/OERI), Washington, DC.

    The Financial Statistics machine-readable data file (MRDF) is a subfile of the larger Higher Education General Information Survey (HEGIS). It contains basic financial statistics for over 3,000 institutions of higher education in the United States and its territories. The data are arranged sequentially by institution, with institutional…

  20. 45 CFR 205.57 - Maintenance of a machine readable file; requests for income and eligibility information.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... Public Welfare OFFICE OF FAMILY ASSISTANCE (ASSISTANCE PROGRAMS), ADMINISTRATION FOR CHILDREN AND FAMILIES, DEPARTMENT OF HEALTH AND HUMAN SERVICES GENERAL ADMINISTRATION-PUBLIC ASSISTANCE PROGRAMS § 205... 45 Public Welfare 2 2012-10-01 2012-10-01 false Maintenance of a machine readable file; requests...

  1. 45 CFR 205.57 - Maintenance of a machine readable file; requests for income and eligibility information.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... Public Welfare OFFICE OF FAMILY ASSISTANCE (ASSISTANCE PROGRAMS), ADMINISTRATION FOR CHILDREN AND FAMILIES, DEPARTMENT OF HEALTH AND HUMAN SERVICES GENERAL ADMINISTRATION-PUBLIC ASSISTANCE PROGRAMS § 205... 45 Public Welfare 2 2013-10-01 2012-10-01 true Maintenance of a machine readable file; requests...

  2. 45 CFR 205.57 - Maintenance of a machine readable file; requests for income and eligibility information.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... Public Welfare OFFICE OF FAMILY ASSISTANCE (ASSISTANCE PROGRAMS), ADMINISTRATION FOR CHILDREN AND FAMILIES, DEPARTMENT OF HEALTH AND HUMAN SERVICES GENERAL ADMINISTRATION-PUBLIC ASSISTANCE PROGRAMS § 205... 45 Public Welfare 2 2014-10-01 2012-10-01 true Maintenance of a machine readable file; requests...

  3. 45 CFR 205.57 - Maintenance of a machine readable file; requests for income and eligibility information.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... Public Welfare OFFICE OF FAMILY ASSISTANCE (ASSISTANCE PROGRAMS), ADMINISTRATION FOR CHILDREN AND FAMILIES, DEPARTMENT OF HEALTH AND HUMAN SERVICES GENERAL ADMINISTRATION-PUBLIC ASSISTANCE PROGRAMS § 205... 45 Public Welfare 2 2011-10-01 2011-10-01 false Maintenance of a machine readable file; requests...

  4. Elementary and Secondary School Civil Rights Survey, 1984 [machine-readable data file].

    ERIC Educational Resources Information Center

    DBS Corp., Arlington, VA.

    The "Elementary and Secondary School Civil Rights Survey" machine-readable data file (MRDF) contains data on the characteristics of student populations enrolled in public schools throughout the United States. The emphasis is on data by race/ethnicity and sex in the following areas: stereotyping in courses, special education, vocational education,…

  5. Choosing the Future: College Students' Projections of Their Personal Life Patterns [machine-readable data file].

    ERIC Educational Resources Information Center

    Thomas, Joan

    "Choosing the Future: College Students' Projections of Their Personal Life Patterns" is a machine-readable data file (MRDF) prepared by the principal investigator in connection with her doctoral program studies and her 1986 unpublished doctoral dissertation prepared in the Department of Psychology at the University of Cincinnati. The…

  6. COM: Decisions and Applications in a Small University Library.

    ERIC Educational Resources Information Center

    Schwarz, Philip J.

    Computer-output microfilm (COM) is used at the University of Wisconsin-Stout Library to generate reports from its major machine readable data bases. Conditions indicating the need to convert to COM include existence of a machine readable data base and high cost of report production. Advantages and disadvantages must also be considered before…

  7. Reference Manual for Machine-Readable Descriptions of Research Projects and Institutions.

    ERIC Educational Resources Information Center

    Dierickx, Harold; Hopkinson, Alan

    This reference manual presents a standardized communication format for the exchange between databases or other information services of machine-readable information on research in progress. The manual is produced in loose-leaf format to facilitate updating. Its first section defines in broad outline the format and content of applicable records. A…

  8. Student Achievement Study, 1970-1974. The IEA Six-Subject Data Bank [machine-readable data file].

    ERIC Educational Resources Information Center

    International Association for the Evaluation of Educational Achievement, Stockholm (Sweden).

    The "Student Achievement Study" machine-readable data files (MRDF) (also referred to as the "IEA Six-Subject Survey") are the result of an international data collection effort during 1970-1974 by 21 designated National Centers, which had agreed to cooperate. The countries involved were: Australia, Belgium, Chile, England-Wales,…

  9. Migrant Student Record Transfer System (MSRTS) [machine-readable data file].

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock. General Education Div.

    The Migrant Student Record Transfer System (MSRTS) machine-readable data file (MRDF) is a collection of education and health data on more than 750,000 migrant children in grades K-12 in the United States (except Hawaii), the District of Columbia, and the outlying territories of Puerto Rico and the Mariana and Marshall Islands. The active file…

  10. Machine-Readable Data Files in the Social Sciences: An Anthropologist and a Librarian Look at the Issues.

    ERIC Educational Resources Information Center

    Bernard, H. Russell; Jones, Ray

    1984-01-01

    Focuses on problems in making machine-readable data files (MRDFs) accessible and in using them: quality of data in MRDFs themselves (social scientists' concern) and accessibility--availability of bibliographic control, quality of documentation, level of user skills (librarians' concern). Skills needed by social scientists and librarians are…

  11. 49 CFR 573.4 - Definitions.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... than a tire) that was installed in or on a motor vehicle at the time of its delivery to the first purchaser if the item of equipment was installed on or in the motor vehicle at the time of its delivery to a... readable by machine. If readable by machine, the submitting party must obtain written confirmation from the...

  12. 49 CFR 573.4 - Definitions.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... than a tire) that was installed in or on a motor vehicle at the time of its delivery to the first purchaser if the item of equipment was installed on or in the motor vehicle at the time of its delivery to a... readable by machine. If readable by machine, the submitting party must obtain written confirmation from the...

  13. 49 CFR 573.4 - Definitions.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... than a tire) that was installed in or on a motor vehicle at the time of its delivery to the first purchaser if the item of equipment was installed on or in the motor vehicle at the time of its delivery to a... readable by machine. If readable by machine, the submitting party must obtain written confirmation from the...

  14. NOAA's Data Catalog and the Federal Open Data Policy

    NASA Astrophysics Data System (ADS)

    Wengren, M. J.; de la Beaujardiere, J.

    2014-12-01

    The 2013 Open Data Policy Presidential Directive requires Federal agencies to create and maintain a 'public data listing' that includes all agency data that is currently or will be made publicly-available in the future. The directive requires the use of machine-readable and open formats that make use of 'common core' and extensible metadata formats according to the best practices published in an online repository called 'Project Open Data', to use open licenses where possible, and to adhere to existing metadata and other technology standards to promote interoperability. In order to meet the requirements of the Open Data Policy, the National Oceanic and Atmospheric Administration (NOAA) has implemented an online data catalog that combines metadata from all subsidiary NOAA metadata catalogs into a single master inventory. The NOAA Data Catalog is available to the public for search and discovery, providing access to the NOAA master data inventory through multiple means, including web-based text search, OGC CS-W endpoint, as well as a native Application Programming Interface (API) for programmatic query. It generates on a daily basis the Project Open Data JavaScript Object Notation (JSON) file required for compliance with the Presidential directive. The Data Catalog is based on the open source Comprehensive Knowledge Archive Network (CKAN) software and runs on the Amazon Federal GeoCloud. This presentation will cover topics including mappings of existing metadata in standard formats (FGDC-CSDGM and ISO 19115 XML ) to the Project Open Data JSON metadata schema, representation of metadata elements within the catalog, and compatible metadata sources used to feed the catalog to include Web Accessible Folder (WAF), Catalog Services for the Web (CS-W), and Esri ArcGIS.com. It will also discuss related open source technologies that can be used together to build a spatial data infrastructure compliant with the Open Data Policy.

  15. 45 CFR 205.57 - Maintenance of a machine readable file; requests for income and eligibility information.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 45 Public Welfare 2 2010-10-01 2010-10-01 false Maintenance of a machine readable file; requests for income and eligibility information. 205.57 Section 205.57 Public Welfare Regulations Relating to Public Welfare OFFICE OF FAMILY ASSISTANCE (ASSISTANCE PROGRAMS), ADMINISTRATION FOR CHILDREN AND FAMILIES, DEPARTMENT OF HEALTH AND HUMAN SERVICES...

  16. CLOSSS: A Machine Readable Data Base of Social Science Serials, Progress Report, 1971-1972. Working Paper No. 8.

    ERIC Educational Resources Information Center

    Roberts, S. A.; Bradshaw, R. G.

    Deisgn of Information Systems in the Social Sciences (DISISS) is a research project conducted to describe the main characteristics of the literature of the social sciences using bibliometric techniques. A comprehensive machine readable file of social science serials was developed which is called CLOSSS (Check List of Social Science Serials). Data…

  17. Development of OCR system for portable passport and visa reader

    NASA Astrophysics Data System (ADS)

    Visilter, Yury V.; Zheltov, Sergey Y.; Lukin, Anton A.

    1999-01-01

    The modern passport and visa documents include special machine-readable zones satisfied the ICAO standards. This allows to develop the special passport and visa automatic readers. However, there are some special problems in such OCR systems: low resolution of character images captured by CCD-camera (down to 150 dpi), essential shifts and slopes (up to 10 degrees), rich paper texture under the character symbols, non-homogeneous illumination. This paper presents the structure and some special aspects of OCR system for portable passport and visa reader. In our approach the binarization procedure is performed after the segmentation step, and it is applied to the each character site separately. Character recognition procedure uses the structural information of machine-readable zone. Special algorithms are developed for machine-readable zone extraction and character segmentation.

  18. Salaries, Tenure, and Fringe Benefits of Full-Time Instructional Faculty. Higher Education General Information Survey (HEGIS) [machine-readable data file].

    ERIC Educational Resources Information Center

    VSE Corp., Alexandria, VA.

    The "Faculty Salary Survey" machine-readable data file (MRDF) is one component of the Higher Education General Information Survey (HEGIS). It contains data about salaries, tenure, and fringe benefits for full-time instructional faculty from over 3,000 institutions of higher education located in the United States and its outlying areas.…

  19. A multiplet table for Mn I (Adelman, Svatek, Van Winkler, Warren 1989): Documentation for the machine-readable version

    NASA Technical Reports Server (NTRS)

    Warren, Wayne H., Jr.; Adelman, Saul J.

    1989-01-01

    The machine-readable version of the multiplet table, as it is currently being distributed from the Astronomical Data Center, is described. The computerized version of the table contains data on excitation potentials, J values, multiplet terms, intensities of the transitions, and multiplet numbers. Files ordered by multiplet and by wavelength are included in the distributed version.

  20. State of the art of geoscience libraries and information services

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pruett, N.J.

    Geoscience libraries and geoscience information services are closely related. Both are trying to meet the needs of the geoscientists for information and data. Both are also being affected by many trends: increased availability of personal computers; decreased costs of machine readable storage; increased availability of maps in digital format (Pallatto, 1986); progress in graphic displays and in developing Geographic Information System, (GIS) (Kelly and Phillips, 1986); development in artificial intelligence; and the availability of new formats (e.g. CD-ROM). Some additional factors are at work at changing the role of libraries: libraries are coming to recognize the impossibility of collecting everythingmore » and the validity of Bradford's Law unobtrustive studies of library reference services have pointed out that only 50% of the questions are answered correctly it is clear that the number of databases is increasing although good figures for specifically geoscience databases are not available; lists of numeric database are beginning to appear; evaluative (as opposed to purely descriptive) reviews of available bibliographic databases are beginning to appear; more and more libraries are getting online catalogs and results of studies of users of online catalog are being used to improve catalog design; and research is raising consciousness about the value of; and research is raising consciousness about the value of information. All these trends are having or will have an effect on geoscience information.« less

  1. The National Longitudinal Study of the High School Class of 1972 (NLS-72), Fifth Follow-Up (1986) Data File [machine-readable data file].

    ERIC Educational Resources Information Center

    National Center for Education Statistics (ED), Washington, DC.

    This machine-readable data file (MDRF) contains information from the fifth follow-up survey of the National Longitudinal Survey of the High School Class of 1972. The survey was carried out along with the third survey of the High School and Beyond Study. The fifth follow-up data file consists of 12,841 records. The data tape contains information on…

  2. Documentation for the machine-readable version of A Finding List for the Multiplet Tables of NSRDS-NBS 3, Sections 1-10 (Adelman, Adelman, Fischel and Warren 1984)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine-readable finding list, as it is currently being distributed from the Astronomical Data Center, is described. This version of the list supersedes an earlier one (1977) containing only Sections 1 through 7 of the NSRDS-NBS 3 multiplet tables publications. Additional sections are to be incorporated into this list as they are published.

  3. The National Longitudinal Study of the High School Class of 1972 (NLS-72), Fifth Follow-Up (1986). Teaching Supplement Data File [machine-readable data file].

    ERIC Educational Resources Information Center

    National Center for Education Statistics (ED), Washington, DC.

    The National Longitudinal Survey of the High School Class of 1972 (NLS-72) Teaching Supplement Data File (TSDF) is presented. Data for the machine-readable data file (MDRF) were collected via a mail questionnaire that was sent to all respondents (N=1,517) to the fifth follow-up survey who indicated that they had a teaching background or training…

  4. Toolsets for Airborne Data (TAD): Improving Machine Readability for ICARTT Data Files

    NASA Astrophysics Data System (ADS)

    Northup, E. A.; Early, A. B.; Beach, A. L., III; Kusterer, J.; Quam, B.; Wang, D.; Chen, G.

    2015-12-01

    NASA has conducted airborne tropospheric chemistry studies for about three decades. These field campaigns have generated a great wealth of observations, including a wide range of the trace gases and aerosol properties. The ASDC Toolsets for Airborne Data (TAD) is designed to meet the user community needs for manipulating aircraft data for scientific research on climate change and air quality relevant issues. TAD makes use of aircraft data stored in the International Consortium for Atmospheric Research on Transport and Transformation (ICARTT) file format. ICARTT has been the NASA standard since 2010, and is widely used by NOAA, NSF, and international partners (DLR, FAAM). Its level of acceptance is due in part to it being generally self-describing for researchers, i.e., it provides necessary data descriptions for proper research use. Despite this, there are a number of issues with the current ICARTT format, especially concerning the machine readability. In order to overcome these issues, the TAD team has developed an "idealized" file format. This format is ASCII and is sufficiently machine readable to sustain the TAD system, however, it is not fully compatible with the current ICARTT format. The process of mapping ICARTT metadata to the idealized format, the format specifics, and the actual conversion process will be discussed. The goal of this presentation is to demonstrate an example of how to improve the machine readability of ASCII data format protocols.

  5. 48 CFR 252.211-7003 - Item identification and valuation.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ..., used to retrieve data encoded on machine-readable media. Concatenated unique item identifier means— (1... (or controlling) authority for the enterprise identifier. Item means a single hardware article or a...-readable means an automatic identification technology media, such as bar codes, contact memory buttons...

  6. 48 CFR 252.211-7003 - Item identification and valuation.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ..., used to retrieve data encoded on machine-readable media. Concatenated unique item identifier means— (1... (or controlling) authority for the enterprise identifier. Item means a single hardware article or a...-readable means an automatic identification technology media, such as bar codes, contact memory buttons...

  7. 48 CFR 252.211-7003 - Item identification and valuation.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ..., used to retrieve data encoded on machine-readable media. Concatenated unique item identifier means— (1... (or controlling) authority for the enterprise identifier. Item means a single hardware article or a...-readable means an automatic identification technology media, such as bar codes, contact memory buttons...

  8. Microcomputer Software Packages--Choose with Caution.

    ERIC Educational Resources Information Center

    Naumer, Janet Noll

    1983-01-01

    Briefly discusses types of software available for library and media center operations and library instruction, suggests three sources of software reviews, and describes almost 50 specific application programs available for bibliographic management, cataloging, circulation, inventory and purchasing, readability, and teaching library skills in…

  9. HD-SAO-DM cross index

    NASA Technical Reports Server (NTRS)

    Nagy, T. A.; Mead, J.

    1978-01-01

    A table of correspondence SAO-HD-DM-GC was prepared by Morin (1973). The machine-readable version of this cross identification was obtained from the Centre de Donnees Stellaires (Strasbourg, France). The table was sorted at the Goddard Space Flight Center by HD number and all blank HD number records were removed to produce the HD-SAO-DM table presented. There were 258997 entries in the original table; there are 180411 entries after removing the blank HD records. The Boss General Catalogue (GC) numbers were retained on the machine-readable version after the sort.

  10. Assessing the Readability of Medical Documents: A Ranking Approach.

    PubMed

    Zheng, Jiaping; Yu, Hong

    2018-03-23

    The use of electronic health record (EHR) systems with patient engagement capabilities, including viewing, downloading, and transmitting health information, has recently grown tremendously. However, using these resources to engage patients in managing their own health remains challenging due to the complex and technical nature of the EHR narratives. Our objective was to develop a machine learning-based system to assess readability levels of complex documents such as EHR notes. We collected difficulty ratings of EHR notes and Wikipedia articles using crowdsourcing from 90 readers. We built a supervised model to assess readability based on relative orders of text difficulty using both surface text features and word embeddings. We evaluated system performance using the Kendall coefficient of concordance against human ratings. Our system achieved significantly higher concordance (.734) with human annotators than did a baseline using the Flesch-Kincaid Grade Level, a widely adopted readability formula (.531). The improvement was also consistent across different disease topics. This method's concordance with an individual human user's ratings was also higher than the concordance between different human annotators (.658). We explored methods to automatically assess the readability levels of clinical narratives. Our ranking-based system using simple textual features and easy-to-learn word embeddings outperformed a widely used readability formula. Our ranking-based method can predict relative difficulties of medical documents. It is not constrained to a predefined set of readability levels, a common design in many machine learning-based systems. Furthermore, the feature set does not rely on complex processing of the documents. One potential application of our readability ranking is personalization, allowing patients to better accommodate their own background knowledge. ©Jiaping Zheng, Hong Yu. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 23.03.2018.

  11. Understanding and Writing G & M Code for CNC Machines

    ERIC Educational Resources Information Center

    Loveland, Thomas

    2012-01-01

    In modern CAD and CAM manufacturing companies, engineers design parts for machines and consumable goods. Many of these parts are cut on CNC machines. Whether using a CNC lathe, milling machine, or router, the ideas and designs of engineers must be translated into a machine-readable form called G & M Code that can be used to cut parts to precise…

  12. A Catalog of Performance Objectives, Performance Conditions, and Performance Guides for Machine Tool Operations.

    ERIC Educational Resources Information Center

    Stadt, Ronald; And Others

    This catalog provides performance objectives, tasks, standards, and performance guides associated with current occupational information relating to the job content of machinists, specifically tool grinder operators, production lathe operators, and production screw machine operators. The catalog is comprised of 262 performance objectives, tool and…

  13. Framework for Building Collaborative Research Environment

    DOE PAGES

    Devarakonda, Ranjeet; Palanisamy, Giriprakash; San Gil, Inigo

    2014-10-25

    Wide range of expertise and technologies are the key to solving some global problems. Semantic web technology can revolutionize the nature of how scientific knowledge is produced and shared. The semantic web is all about enabling machine-machine readability instead of a routine human-human interaction. Carefully structured data, as in machine readable data is the key to enabling these interactions. Drupal is an example of one such toolset that can render all the functionalities of Semantic Web technology right out of the box. Drupal’s content management system automatically stores the data in a structured format enabling it to be machine. Withinmore » this paper, we will discuss how Drupal promotes collaboration in a research setting such as Oak Ridge National Laboratory (ORNL) and Long Term Ecological Research Center (LTER) and how it is effectively using the Semantic Web in achieving this.« less

  14. Astronomical Catalogues - Definition Elements and Afterlife

    NASA Astrophysics Data System (ADS)

    Jaschek, C.

    1984-09-01

    Based on a look at the different meanings of the term catalogue (or catalog), a definition is proposed. In an analysis of the main elements, a number of requirements that catalogues should satisfy are pointed out. A section is devoted to problems connected with computer-readable versions of printed catalogues.

  15. U.S. announces one-year delay for visa waiver program change

    NASA Astrophysics Data System (ADS)

    The U.S. State Department has announced that it is delaying by one year a new rule affecting citizens from visa waiver program countries. The new rule, which was scheduled to go into effect on 1 October 2003, requires visitors from these countries to obtain non-immigrant visas to enter the United States if they do not have machine-readable passports. The announced delay means that this rule will now go into effect 26 October 2004 instead.The delay does not apply to five visa waiver countries—Andorra, Brunei, Liechtenstein, Luxembourg, and Slovenia—because most of the citizens of these nations already carry passports that are machine-readable.

  16. Machine-readable files developed for the High Plains Regional Aquifer-System analysis in parts of Colorado, Kansas, Nebraska, New Mexico, Oklahoma, South Dakota, Texas, and Wyoming

    USGS Publications Warehouse

    Ferrigno, C.F.

    1986-01-01

    Machine-readable files were developed for the High Plains Regional Aquifer-System Analysis project are stored on two magnetic tapes available from the U.S. Geological Survey. The first tape contains computer programs that were used to prepare, store, retrieve, organize, and preserve the areal interpretive data collected by the project staff. The second tape contains 134 data files that can be divided into five general classes: (1) Aquifer geometry data, (2) aquifer and water characteristics , (3) water levels, (4) climatological data, and (5) land use and water use data. (Author 's abstract)

  17. Document fraud deterrent strategies: four case studies

    NASA Astrophysics Data System (ADS)

    Mercer, John W.

    1998-04-01

    This paper discusses the approaches taken to deter fraud committed against four documents: the machine-readable passport; the machine-readable visa; the Consular Report of Birth Abroad; and the Border Crossing Card. General approaches are discussed first, with an emphasis on the reasons for the document, the conditions of its use and the information systems required for it to function. A cost model of counterfeit deterrence is introduced. Specific approaches to each of the four documents are then discussed, in light of the issuance circumstances and criteria, the intent of the issuing authority, the applicable international standards and the level of protection and fraud resistance appropriate for the document.

  18. Documentation for the machine-readable version of the catalogue of individual UBV and UVBY beta observations in the region of the Orion OB1 association

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable files of individual UBV observations of 106 stars in the vicinity of the Orion Nebula (the Sword region) and individual uvby beta observations of 508 stars in all regions of the Orion OB 1 association are described. For the UBV data the stars are identified by their Brun numbers, with cross identifications to the chart numbers used in Warren and Hesser; the uv by beta stars are identified by the aforementioned chart numbers and HD, BD or P ( = pi); numbers in that order of preference.

  19. Development of a Machine Form Union Catalog for the New England Library Information Network (NELINET). Final Report.

    ERIC Educational Resources Information Center

    Goldstein, Samuel; And Others

    Based on a literature survey of union cataloging, and New England libraries it was determined that: (1) New England's collective union catalog needs and problems had not been specified, especially regarding the possibilities of machine application; (2) crucial data and analysis needed for such specification was unavailable; and (3) the absence of…

  20. 3D Visualization of Machine Learning Algorithms with Astronomical Data

    NASA Astrophysics Data System (ADS)

    Kent, Brian R.

    2016-01-01

    We present innovative machine learning (ML) methods using unsupervised clustering with minimum spanning trees (MSTs) to study 3D astronomical catalogs. Utilizing Python code to build trees based on galaxy catalogs, we can render the results with the visualization suite Blender to produce interactive 360 degree panoramic videos. The catalogs and their ML results can be explored in a 3D space using mobile devices, tablets or desktop browsers. We compare the statistics of the MST results to a number of machine learning methods relating to optimization and efficiency.

  1. CDISC SHARE, a Global, Cloud-based Resource of Machine-Readable CDISC Standards for Clinical and Translational Research

    PubMed Central

    Hume, Samuel; Chow, Anthony; Evans, Julie; Malfait, Frederik; Chason, Julie; Wold, J. Darcy; Kubick, Wayne; Becnel, Lauren B.

    2018-01-01

    The Clinical Data Interchange Standards Consortium (CDISC) is a global non-profit standards development organization that creates consensus-based standards for clinical and translational research. Several of these standards are now required by regulators for electronic submissions of regulated clinical trials’ data and by government funding agencies. These standards are free and open, available for download on the CDISC Website as PDFs. While these documents are human readable, they are not amenable to ready use by electronic systems. CDISC launched the CDISC Shared Health And Research Electronic library (SHARE) to provide the standards metadata in machine-readable formats to facilitate the automated management and implementation of the standards. This paper describes how CDISC SHARE’S standards can facilitate collecting, aggregating and analyzing standardized data from early design to end analysis; and its role as a central resource providing information systems with metadata that drives process automation including study setup and data pipelining. PMID:29888049

  2. Documentation for the machine-readable version of the Bright Star Catalogue

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    The machine-readable version of The Bright Star Catalogue, 4th edition, is described. In addition to the large number of newly determined fundamental data, such as photoelectric magnitudes, MK spectral types, parallaxes, and radial velocities, the new version contains data and information not included in the third edition such as the identification of IR sources, U-B and R-I colors, radial velocity comments (indication and identification of spectroscopic and occultation binaries), and projected rotational velocities. The equatorial coordinates for equinoxes 1900 and 2000 are recorded to greater precision details concerning variability, spectral characteristics, duplicity, and group membership are included. Data compiled through 1979, some information and variable-star designations found through 1981 are considered.

  3. Multiple layer identification label using stacked identification symbols

    NASA Technical Reports Server (NTRS)

    Schramm, Harry F. (Inventor)

    2005-01-01

    An automatic identification system and method are provided which employ a machine readable multiple layer label. The label has a plurality of machine readable marking layers stacked one upon another. Each of the marking layers encodes an identification symbol detectable using one or more sensing technologies. The various marking layers may comprise the same marking material or each marking layer may comprise a different medium having characteristics detectable by a different sensing technology. These sensing technologies include x-ray, radar, capacitance, thermal, magnetic and ultrasonic. A complete symbol may be encoded within each marking layer or a symbol may be segmented into fragments which are then divided within a single marking layer or encoded across multiple marking layers.

  4. Documentation for the machine-readable version of A Library of Stellar Spectra (Jacoby, Hunter and Christian 1984)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine readable library as it is currently being distributed from the Astronomical Data Center is described. The library contains digital spectral for 161 stars of spectral classes O through M and luminosity classes 1, 3 and 5 in the wavelength range 3510 A to 7427 A. The resolution is approximately 4.5 A, while the typical photometric uncertainty of each resolution element is approximately 1 percent and broadband variations are 3 percent. The documentation includes a format description, a table of the indigenous characteristics of the magnetic tape file, and a sample listing of logical records exactly as they are recorded on the tape.

  5. Documentation for the machine-readable version of the Stellar Spectrophotometric Atlas, 3130 A lambda 10800 A of Gunn and Stryker (1983)

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1984-01-01

    The machine-readable version of the Atlas as it is currently being distributed from the Astronomical Data Center is described. The data were obtained with the Oke multichannel scanner on the 5-meter Hale reflector for purposes of synthesizing galaxy spectra, and the digitized Atlas contains normalized spectral energy distributions, computed colors, scan line and continuum indices for 175 selected stars covering the complete ranges of spectral type and luminosity class. The documentation includes a byte-by-byte format description, a table of the indigenous characteristics of the magnetic tape file, and a sample listing of logical records exactly as they are recorded on the tape.

  6. Biotea: RDFizing PubMed Central in support for the paper as an interface to the Web of Data

    PubMed Central

    2013-01-01

    Background The World Wide Web has become a dissemination platform for scientific and non-scientific publications. However, most of the information remains locked up in discrete documents that are not always interconnected or machine-readable. The connectivity tissue provided by RDF technology has not yet been widely used to support the generation of self-describing, machine-readable documents. Results In this paper, we present our approach to the generation of self-describing machine-readable scholarly documents. We understand the scientific document as an entry point and interface to the Web of Data. We have semantically processed the full-text, open-access subset of PubMed Central. Our RDF model and resulting dataset make extensive use of existing ontologies and semantic enrichment services. We expose our model, services, prototype, and datasets at http://biotea.idiginfo.org/ Conclusions The semantic processing of biomedical literature presented in this paper embeds documents within the Web of Data and facilitates the execution of concept-based queries against the entire digital library. Our approach delivers a flexible and adaptable set of tools for metadata enrichment and semantic processing of biomedical documents. Our model delivers a semantically rich and highly interconnected dataset with self-describing content so that software can make effective use of it. PMID:23734622

  7. Gloved Human-Machine Interface

    NASA Technical Reports Server (NTRS)

    Adams, Richard (Inventor); Hannaford, Blake (Inventor); Olowin, Aaron (Inventor)

    2015-01-01

    Certain exemplary embodiments can provide a system, machine, device, manufacture, circuit, composition of matter, and/or user interface adapted for and/or resulting from, and/or a method and/or machine-readable medium comprising machine-implementable instructions for, activities that can comprise and/or relate to: tracking movement of a gloved hand of a human; interpreting a gloved finger movement of the human; and/or in response to interpreting the gloved finger movement, providing feedback to the human.

  8. The feasibility of using natural language processing to extract clinical information from breast pathology reports.

    PubMed

    Buckley, Julliette M; Coopey, Suzanne B; Sharko, John; Polubriaginof, Fernanda; Drohan, Brian; Belli, Ahmet K; Kim, Elizabeth M H; Garber, Judy E; Smith, Barbara L; Gadd, Michele A; Specht, Michelle C; Roche, Constance A; Gudewicz, Thomas M; Hughes, Kevin S

    2012-01-01

    The opportunity to integrate clinical decision support systems into clinical practice is limited due to the lack of structured, machine readable data in the current format of the electronic health record. Natural language processing has been designed to convert free text into machine readable data. The aim of the current study was to ascertain the feasibility of using natural language processing to extract clinical information from >76,000 breast pathology reports. APPROACH AND PROCEDURE: Breast pathology reports from three institutions were analyzed using natural language processing software (Clearforest, Waltham, MA) to extract information on a variety of pathologic diagnoses of interest. Data tables were created from the extracted information according to date of surgery, side of surgery, and medical record number. The variety of ways in which each diagnosis could be represented was recorded, as a means of demonstrating the complexity of machine interpretation of free text. There was widespread variation in how pathologists reported common pathologic diagnoses. We report, for example, 124 ways of saying invasive ductal carcinoma and 95 ways of saying invasive lobular carcinoma. There were >4000 ways of saying invasive ductal carcinoma was not present. Natural language processor sensitivity and specificity were 99.1% and 96.5% when compared to expert human coders. We have demonstrated how a large body of free text medical information such as seen in breast pathology reports, can be converted to a machine readable format using natural language processing, and described the inherent complexities of the task.

  9. Application of XML to Journal Table Archiving

    NASA Astrophysics Data System (ADS)

    Shaya, E. J.; Blackwell, J. H.; Gass, J. E.; Kargatis, V. E.; Schneider, G. L.; Weiland, J. L.; Borne, K. D.; White, R. A.; Cheung, C. Y.

    1998-12-01

    The Astronomical Data Center (ADC) at the NASA Goddard Space Flight Center is a major archive for machine-readable astronomical data tables. Many ADC tables are derived from published journal articles. Article tables are reformatted to be machine-readable and documentation is crafted to facilitate proper reuse by researchers. The recent switch of journals to web based electronic format has resulted in the generation of large amounts of tabular data that could be captured into machine-readable archive format at fairly low cost. The large data flow of the tables from all major North American astronomical journals (a factor of 100 greater than the present rate at the ADC) necessitates the development of rigorous standards for the exchange of data between researchers, publishers, and the archives. We have selected a suitable markup language that can fully describe the large variety of astronomical information contained in ADC tables. The eXtensible Markup Language XML is a powerful internet-ready documentation format for data. It provides a precise and clear data description language that is both machine- and human-readable. It is rapidly becoming the standard format for business and information transactions on the internet and it is an ideal common metadata exchange format. By labelling, or "marking up", all elements of the information content, documents are created that computers can easily parse. An XML archive can easily and automatically be maintained, ingested into standard databases or custom software, and even totally restructured whenever necessary. Structuring astronomical data into XML format will enable efficient and focused search capabilities via off-the-shelf software. The ADC is investigating XML's expanded hyperlinking power to enhance connectivity within the ADC data/metadata and developing XSL display scripts to enhance display of astronomical data. The ADC XML Definition Type Document can be viewed at http://messier.gsfc.nasa.gov/dtdhtml/DTD-TREE.html

  10. E-Assessment Data Compatibility Resolution Methodology with Bidirectional Data Transformation

    ERIC Educational Resources Information Center

    Malik, Kaleem Razzaq; Ahmad, Tauqir

    2017-01-01

    Electronic Assessment (E-Assessment) also known as computer aided assessment for the purposes involving diagnostic, formative or summative examining using data analysis. Digital assessments come commonly from social, academic, and adaptive learning in machine readable forms to deliver the machine scoring function. To achieve real-time and smart…

  11. VizieR Online Data Catalog: Revised source list for the Rees 38-MHz survey (Hales+ 1995)

    NASA Astrophysics Data System (ADS)

    Hales, S. E. G.; Waldram, E. M.; Rees, N.; Warner, P. J.

    1994-11-01

    We present a revised machine-readable source list for the Rees 38-MHz (or '8C') survey with improved positions and no redundancy. The Rees 38-MHz survey covers an area of about 1 sr north of declination +60 degrees. The angular resolution is 4.5 x 4.5cosec(Dec) arcmin**2 and the limiting flux density over much of the survey area is about 1 Jy. Both of these figures are an improvement by nearly an order of magnitude on previous surveys at this frequency. Users of these data should consult and cite the original survey paper by Rees as primary reference (1990MNRAS.244..233R) with the present publication (1995MNRAS.274..447H) as a supplementary revision. The recommended style of reference is thus : "The revised Rees 38-MHz survey (Rees 1990, catalogue revised Hales et. al 1995)." Note that for interest the source list includes data on some sources at declinations lower than +60 degrees, but that the right ascension coverage is not complete below +60 degrees. (1 data file).

  12. Finding Atmospheric Composition (AC) Metadata

    NASA Technical Reports Server (NTRS)

    Strub, Richard F..; Falke, Stefan; Fiakowski, Ed; Kempler, Steve; Lynnes, Chris; Goussev, Oleg

    2015-01-01

    The Atmospheric Composition Portal (ACP) is an aggregator and curator of information related to remotely sensed atmospheric composition data and analysis. It uses existing tools and technologies and, where needed, enhances those capabilities to provide interoperable access, tools, and contextual guidance for scientists and value-adding organizations using remotely sensed atmospheric composition data. The initial focus is on Essential Climate Variables identified by the Global Climate Observing System CH4, CO, CO2, NO2, O3, SO2 and aerosols. This poster addresses our efforts in building the ACP Data Table, an interface to help discover and understand remotely sensed data that are related to atmospheric composition science and applications. We harvested GCMD, CWIC, GEOSS metadata catalogs using machine to machine technologies - OpenSearch, Web Services. We also manually investigated the plethora of CEOS data providers portals and other catalogs where that data might be aggregated. This poster is our experience of the excellence, variety, and challenges we encountered.Conclusions:1.The significant benefits that the major catalogs provide are their machine to machine tools like OpenSearch and Web Services rather than any GUI usability improvements due to the large amount of data in their catalog.2.There is a trend at the large catalogs towards simulating small data provider portals through advanced services. 3.Populating metadata catalogs using ISO19115 is too complex for users to do in a consistent way, difficult to parse visually or with XML libraries, and too complex for Java XML binders like CASTOR.4.The ability to search for Ids first and then for data (GCMD and ECHO) is better for machine to machine operations rather than the timeouts experienced when returning the entire metadata entry at once. 5.Metadata harvest and export activities between the major catalogs has led to a significant amount of duplication. (This is currently being addressed) 6.Most (if not all) Earth science atmospheric composition data providers store a reference to their data at GCMD.

  13. Automated Atmospheric Composition Dataset Level Metadata Discovery. Difficulties and Surprises

    NASA Astrophysics Data System (ADS)

    Strub, R. F.; Falke, S. R.; Kempler, S.; Fialkowski, E.; Goussev, O.; Lynnes, C.

    2015-12-01

    The Atmospheric Composition Portal (ACP) is an aggregator and curator of information related to remotely sensed atmospheric composition data and analysis. It uses existing tools and technologies and, where needed, enhances those capabilities to provide interoperable access, tools, and contextual guidance for scientists and value-adding organizations using remotely sensed atmospheric composition data. The initial focus is on Essential Climate Variables identified by the Global Climate Observing System - CH4, CO, CO2, NO2, O3, SO2 and aerosols. This poster addresses our efforts in building the ACP Data Table, an interface to help discover and understand remotely sensed data that are related to atmospheric composition science and applications. We harvested GCMD, CWIC, GEOSS metadata catalogs using machine to machine technologies - OpenSearch, Web Services. We also manually investigated the plethora of CEOS data providers portals and other catalogs where that data might be aggregated. This poster is our experience of the excellence, variety, and challenges we encountered.Conclusions:1.The significant benefits that the major catalogs provide are their machine to machine tools like OpenSearch and Web Services rather than any GUI usability improvements due to the large amount of data in their catalog.2.There is a trend at the large catalogs towards simulating small data provider portals through advanced services. 3.Populating metadata catalogs using ISO19115 is too complex for users to do in a consistent way, difficult to parse visually or with XML libraries, and too complex for Java XML binders like CASTOR.4.The ability to search for Ids first and then for data (GCMD and ECHO) is better for machine to machine operations rather than the timeouts experienced when returning the entire metadata entry at once. 5.Metadata harvest and export activities between the major catalogs has led to a significant amount of duplication. (This is currently being addressed) 6.Most (if not all) Earth science atmospheric composition data providers store a reference to their data at GCMD.

  14. U.S. Visa Waiver Program Changes

    NASA Astrophysics Data System (ADS)

    The U.S. State Department has just announced that a change to a new rule affecting citizens from visa waiver program countries. The rule, scheduled to go into effect on 1 October 2003, requires visitors from these countries to obtain non-immigrant visas to enter the United States if they do not have machine-readable passports. The change announced is that a visa waiver country can petition the U.S. government to delay the rule by one year. The State Department recommends that citizens of visa waiver program countries who are contemplating visiting the United States, and do not have machine-readable passports, contact the nearest U.S. embassy or consulate to find out if implementation of the rule has been temporarily waived for their countries.

  15. Documentation for the machine-readable version of OAO 2 filter photometry of 531 stars of diverse types

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1982-01-01

    A magnetic tape version of the ultraviolet photometry of 531 stars observed with the Wisconsin Experiment Package aboard the Orbiting Astronomical Observatory (OAO 2) is described. The data were obtained with medium band interference filters and were reduced to a uniform magnitude system. They represent a subset of partially reduced data currently on file at the National Space Science Data Center. The document is intended to enable users of the tape file to read and process data without problems or guesswork. For technical details concerning the observations, instrumentation limitations, and interpretation of the data the reference publication should be consulted. This document was designed for distribution with any machine-readable version of the OAO 2 photometric data.

  16. 18 CFR 125.2 - General instructions.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... stored on machine readable media. Internal control procedures must be documented by a responsible... associated companies. Public utilities and licensees must assure the availability of records of services...

  17. Toward Automated Benchmarking of Atomistic Force Fields: Neat Liquid Densities and Static Dielectric Constants from the ThermoML Data Archive.

    PubMed

    Beauchamp, Kyle A; Behr, Julie M; Rustenburg, Ariën S; Bayly, Christopher I; Kroenlein, Kenneth; Chodera, John D

    2015-10-08

    Atomistic molecular simulations are a powerful way to make quantitative predictions, but the accuracy of these predictions depends entirely on the quality of the force field employed. Although experimental measurements of fundamental physical properties offer a straightforward approach for evaluating force field quality, the bulk of this information has been tied up in formats that are not machine-readable. Compiling benchmark data sets of physical properties from non-machine-readable sources requires substantial human effort and is prone to the accumulation of human errors, hindering the development of reproducible benchmarks of force-field accuracy. Here, we examine the feasibility of benchmarking atomistic force fields against the NIST ThermoML data archive of physicochemical measurements, which aggregates thousands of experimental measurements in a portable, machine-readable, self-annotating IUPAC-standard format. As a proof of concept, we present a detailed benchmark of the generalized Amber small-molecule force field (GAFF) using the AM1-BCC charge model against experimental measurements (specifically, bulk liquid densities and static dielectric constants at ambient pressure) automatically extracted from the archive and discuss the extent of data available for use in larger scale (or continuously performed) benchmarks. The results of even this limited initial benchmark highlight a general problem with fixed-charge force fields in the representation low-dielectric environments, such as those seen in binding cavities or biological membranes.

  18. 5 CFR 841.1005 - State responsibilities.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ....1005 State responsibilities. The State will, in performance of this agreement: (a) Accept requests and...) Convert these requests on a monthly basis to a machine-readable magnetic tape using specifications...

  19. 5 CFR 841.1005 - State responsibilities.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ....1005 State responsibilities. The State will, in performance of this agreement: (a) Accept requests and...) Convert these requests on a monthly basis to a machine-readable magnetic tape using specifications...

  20. 32 CFR 701.41 - FOIA fee terms.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... can take the form of paper copy, microfiche, audiovisual, or machine readable documentation (e.g... duplication of computer tapes and audiovisual, the actual cost, including the operator's time, shall be...

  1. 18 CFR 356.2 - General instructions.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... procedures that assure the reliability of and ready access to data stored on machine readable media. Internal... of services performed by associated companies. Oil pipeline companies must assure the availability of...

  2. 5 CFR 831.1904 - State responsibilities.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... performance of this agreement: (a) Accept requests and revocations from annuitants who have designated that... machine-readable magnetic tape using specifications received from OPM, and forward that tape to OPM for...

  3. 5 CFR 831.1904 - State responsibilities.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... performance of this agreement: (a) Accept requests and revocations from annuitants who have designated that... machine-readable magnetic tape using specifications received from OPM, and forward that tape to OPM for...

  4. 41 CFR 51-8.3 - Definitions.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... COMMITTEE FOR PURCHASE FROM PEOPLE WHO ARE BLIND OR SEVERELY DISABLED 8-PUBLIC AVAILABILITY OF AGENCY... request. Such copies can take the form of paper copy, audio-visual materials, or machine readable...

  5. 41 CFR 51-8.3 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... COMMITTEE FOR PURCHASE FROM PEOPLE WHO ARE BLIND OR SEVERELY DISABLED 8-PUBLIC AVAILABILITY OF AGENCY... request. Such copies can take the form of paper copy, audio-visual materials, or machine readable...

  6. 32 CFR 701.41 - FOIA fee terms.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... OFFICIAL RECORDS AVAILABILITY OF DEPARTMENT OF THE NAVY RECORDS AND PUBLICATION OF DEPARTMENT OF THE NAVY... can take the form of paper copy, microfiche, audiovisual, or machine readable documentation (e.g...

  7. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Schlossberg, David J.; Bodner, Grant M.; Bongard, Michael W.

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in D.J. Schlossberg et al., 'Non-Inductively Driven Tokamak Plasmas at Near-Unity Toroidal Beta,' Phys. Rev. Lett. 119, 035001 (2017).

  8. 3/29/2018: Making Data Machine-Readable Webinar | National Agricultural

    Science.gov Websites

    Library Skip to main content Home National Agricultural Library United States Department of | USDA.gov | Agricultural Research Service | Plain Language | FOIA | Accessibility Statement | Information

  9. Making USGS Science Data more Open, Accessible, and Usable: Leveraging ScienceBase for Success

    NASA Astrophysics Data System (ADS)

    Chang, M.; Ignizio, D.; Langseth, M. L.; Norkin, T.

    2016-12-01

    In 2013, the White House released initiatives requiring federally funded research to be made publicly available and machine readable. In response, the U.S. Geological Survey (USGS) has been developing a unified approach to make USGS data available and open. This effort has involved the establishment of internal policies and the release of a Public Access Plan, which outlines a strategy for the USGS to move forward into the modern era in scientific data management. Originally designed as a catalog and collaborative data management platform, ScienceBase (www.sciencebase.gov) is being leveraged to serve as a robust data hosting solution for USGS researchers to make scientific data accessible. With the goal of maintaining persistent access to formal data products and developing a management approach to facilitate stable data citation, the ScienceBase Data Release Team was established to ensure the quality, consistency, and meaningful organization of USGS data through standardized workflows and best practices. These practices include the creation and maintenance of persistent identifiers for data, improving the use of open data formats, establishing permissions for read/write access, validating the quality of standards compliant metadata, verifying that data have been reviewed and approved prior to release, and connecting to external search catalogs such as the USGS Science Data Catalog (data.usgs.gov) and data.gov. The ScienceBase team is actively building features to support this effort by automating steps to streamline the process, building metrics to track site visits and downloads, and connecting published digital resources in line with USGS and Federal policy. By utilizing ScienceBase to achieve stewardship quality and employing a dedicated team to help USGS scientists improve the quality of their data, the USGS is helping to meet today's data quality management challenges and ensure that reliable USGS data are available to and reusable for the public.

  10. International Data Archive and Analysis Center. I. International Relations Archive. II. Voluntary International Coordination. III. Attachments.

    ERIC Educational Resources Information Center

    Miller, Warren; Tanter, Raymond

    The International Relations Archive undertakes as its primary goals the acquisition, management and dissemination of international affairs data. The first document enclosed is a copy of the final machine readable codebook prepared for the data from the Political Events Project, 1948-1965. Also included is a copy of the final machine-readable…

  11. 12 CFR 309.5 - Procedures for requesting records.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (4... processing. A requester may contact the FOIA/PA Group to learn whether a particular request has been assigned...

  12. 12 CFR 309.5 - Procedures for requesting records.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (4... processing. A requester may contact the FOIA/PA Group to learn whether a particular request has been assigned...

  13. 75 FR 54052 - Description of Office, Procedures, and Public Information

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-09-03

    ..., found at: http://www.ffiec.gov . Requests must reasonably describe the records sought. (ii) Contents of..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (D...

  14. 12 CFR 309.5 - Procedures for requesting records.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (4... processing. A requester may contact the FOIA/PA Group to learn whether a particular request has been assigned...

  15. 12 CFR 309.5 - Procedures for requesting records.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (4... processing. A requester may contact the FOIA/PA Group to learn whether a particular request has been assigned...

  16. 12 CFR 309.5 - Procedures for requesting records.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ..., microfilm, audiovisual records, or machine readable records (e.g., magnetic tape or computer disk). (4... processing. A requester may contact the FOIA/PA Group to learn whether a particular request has been assigned...

  17. Electronic Publishing.

    ERIC Educational Resources Information Center

    Lancaster, F. W.

    1989-01-01

    Describes various stages involved in the applications of electronic media to the publishing industry. Highlights include computer typesetting, or photocomposition; machine-readable databases; the distribution of publications in electronic form; computer conferencing and electronic mail; collaborative authorship; hypertext; hypermedia publications;…

  18. 32 CFR 701.18 - Agency record.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ..., photographs, machine readable materials, inclusive of those in electronic form or format, or other documentary...) Hard copy or electronic records, which are subject to FOIA requests under 5 U.S.C. 552(a)(3), and which...

  19. 19 CFR 163.1 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... following: Statements; declarations; documents; electronically generated or machine readable data; electronically stored or transmitted information or data; books; papers; correspondence; accounts; financial accounting data; technical data; computer programs necessary to retrieve information in a usable form; and...

  20. ClearTK 2.0: Design Patterns for Machine Learning in UIMA

    PubMed Central

    Bethard, Steven; Ogren, Philip; Becker, Lee

    2014-01-01

    ClearTK adds machine learning functionality to the UIMA framework, providing wrappers to popular machine learning libraries, a rich feature extraction library that works across different classifiers, and utilities for applying and evaluating machine learning models. Since its inception in 2008, ClearTK has evolved in response to feedback from developers and the community. This evolution has followed a number of important design principles including: conceptually simple annotator interfaces, readable pipeline descriptions, minimal collection readers, type system agnostic code, modules organized for ease of import, and assisting user comprehension of the complex UIMA framework. PMID:29104966

  1. ClearTK 2.0: Design Patterns for Machine Learning in UIMA.

    PubMed

    Bethard, Steven; Ogren, Philip; Becker, Lee

    2014-05-01

    ClearTK adds machine learning functionality to the UIMA framework, providing wrappers to popular machine learning libraries, a rich feature extraction library that works across different classifiers, and utilities for applying and evaluating machine learning models. Since its inception in 2008, ClearTK has evolved in response to feedback from developers and the community. This evolution has followed a number of important design principles including: conceptually simple annotator interfaces, readable pipeline descriptions, minimal collection readers, type system agnostic code, modules organized for ease of import, and assisting user comprehension of the complex UIMA framework.

  2. Library Information-Processing System

    NASA Technical Reports Server (NTRS)

    1985-01-01

    System works with Library of Congress MARC II format. System composed of subsystems that provide wide range of library informationprocessing capabilities. Format is American National Standards Institute (ANSI) format for machine-readable bibliographic data. Adaptable to any medium-to-large library.

  3. 76 FR 27048 - Information Collection Being Reviewed by the Federal Communications Commission

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-05-10

    ... Commission; (8) Ex parte notices must be submitted electronically in machine-readable format. PDF images created by scanning a paper document may not be submitted, except in cases in which a word-processing...

  4. 14 CFR 221.500 - Transmission of electronic tariffs to subscribers.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... to any subscriber to the on-line tariff database, including access to the justification required by... machine-readable data (raw tariff data) of all daily transactions made to its on-line tariff database. The...

  5. 14 CFR 221.500 - Transmission of electronic tariffs to subscribers.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... to any subscriber to the on-line tariff database, including access to the justification required by... machine-readable data (raw tariff data) of all daily transactions made to its on-line tariff database. The...

  6. 14 CFR 221.500 - Transmission of electronic tariffs to subscribers.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... to any subscriber to the on-line tariff database, including access to the justification required by... machine-readable data (raw tariff data) of all daily transactions made to its on-line tariff database. The...

  7. 14 CFR 221.500 - Transmission of electronic tariffs to subscribers.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... to any subscriber to the on-line tariff database, including access to the justification required by... machine-readable data (raw tariff data) of all daily transactions made to its on-line tariff database. The...

  8. Structuring supplemental materials in support of reproducibility.

    PubMed

    Greenbaum, Dov; Rozowsky, Joel; Stodden, Victoria; Gerstein, Mark

    2017-04-05

    Supplements are increasingly important to the scientific record, particularly in genomics. However, they are often underutilized. Optimally, supplements should make results findable, accessible, interoperable, and reusable (i.e., "FAIR"). Moreover, properly off-loading to them the data and detail in a paper could make the main text more readable. We propose a hierarchical organization for supplements, with some parts paralleling and "shadowing" the main text and other elements branching off from it, and we suggest a specific formatting to make this structure explicit. Furthermore, sections of the supplement could be presented in multiple scientific "dialects", including machine-readable and lay-friendly formats.

  9. Formulating qualitative features using interactive visualization for analysis of multivariate spatiotemporal data

    NASA Astrophysics Data System (ADS)

    Porter, M.; Hill, M. C.; Pierce, S. A.; Gil, Y.; Pennington, D. D.

    2017-12-01

    DiscoverWater is a web-based visualization tool developed to enable the visual representation of data, and thus, aid scientific and societal understanding of hydrologic systems. Open data sources are coalesced to, for example, illustrate the impacts on streamflow of irrigation withdrawals. Scientists and stakeholders are informed through synchronized time-series data plots that correlate multiple spatiotemporal datasets and an interactive time-evolving map that provides a spatial analytical context. Together, these components elucidate trends so that the user can try to envision the relations between groundwater-surface water interactions, the impacts of pumping on these interactions, and the interplay of climate. Aligning data in this manner has the capacity for interdisciplinary knowledge discovery and motivates dialogue about system processes that we seek to enhance through qualitative features informed through quantitative models. DiscoverWater and its connection is demonstrated using two field cases. First, it is used to visualize data sets from the High Plains aquifer, where reservoir- and groundwater-supported irrigation has affected the Arkansas River in western Kansas. Second, data and model results from Barton Springs segment of the Edwards aquifer in Texas reveal the effects of regional pumping on this important urbanizing aquifer system. Identifying what is interesting about the data and the modeled system in the two different case studies is a step towards moving typically static visualization capabilities to an adaptive framework. Additionally, the dashboard interface incorporates both quantitative and qualitative information about distinctive case studies in a machine-readable form, such that a catalog of qualitative models can capture subject matter expertise alongside associated datasets. As the catalog is expanded to include other case studies, the collection has potential to establish a standard framework able to inform intelligent system reasoning.

  10. InfoQUEST: An Online Catalog for Small Libraries.

    ERIC Educational Resources Information Center

    Campbell, Bonnie

    1984-01-01

    InfoQUEST is a microcomputer-based online public access catalog, designed for the small library handling file sizes up to 25,000 records. Based on the IBM-PC, or compatible machines, the system will accept downloading, in batch mode, of records from the library's file on the UTLAS Catalog Support System. (Author/EJS)

  11. 26 CFR 31.3406(d)-4 - Special rules for readily tradable instruments acquired through a broker.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... means of magnetic media, machine readable document, or any other medium, provided that the notice... her social security number.) (2) You failed to certify, under penalties of perjury, that you are not...

  12. 26 CFR 31.3406(d)-4 - Special rules for readily tradable instruments acquired through a broker.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... means of magnetic media, machine readable document, or any other medium, provided that the notice... her social security number.) (2) You failed to certify, under penalties of perjury, that you are not...

  13. 26 CFR 31.3406(d)-4 - Special rules for readily tradable instruments acquired through a broker.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... means of magnetic media, machine readable document, or any other medium, provided that the notice... her social security number.) (2) You failed to certify, under penalties of perjury, that you are not...

  14. 76 FR 45794 - Notice of Public Information Collection(s) Being Reviewed by the Federal Communications...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-08-01

    ... must be submitted electronically in machine-readable format. PDF images created by scanning a paper document may not be submitted, except in cases in which a word- processing version of the document is not...

  15. 26 CFR 31.3406(d)-4 - Special rules for readily tradable instruments acquired through a broker.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... means of magnetic media, machine readable document, or any other medium, provided that the notice... her social security number.) (2) You failed to certify, under penalties of perjury, that you are not...

  16. 12 CFR 602.3 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 12 Banks and Banking 6 2011-01-01 2011-01-01 false Definitions. 602.3 Section 602.3 Banks and Banking FARM CREDIT ADMINISTRATION ADMINISTRATIVE PROVISIONS RELEASING INFORMATION Availability of Records... means all documentary materials, such as books, papers, maps, photographs, and machine-readable...

  17. MachineProse: an Ontological Framework for Scientific Assertions

    PubMed Central

    Dinakarpandian, Deendayal; Lee, Yugyung; Vishwanath, Kartik; Lingambhotla, Rohini

    2006-01-01

    Objective: The idea of testing a hypothesis is central to the practice of biomedical research. However, the results of testing a hypothesis are published mainly in the form of prose articles. Encoding the results as scientific assertions that are both human and machine readable would greatly enhance the synergistic growth and dissemination of knowledge. Design: We have developed MachineProse (MP), an ontological framework for the concise specification of scientific assertions. MP is based on the idea of an assertion constituting a fundamental unit of knowledge. This is in contrast to current approaches that use discrete concept terms from domain ontologies for annotation and assertions are only inferred heuristically. Measurements: We use illustrative examples to highlight the advantages of MP over the use of the Medical Subject Headings (MeSH) system and keywords in indexing scientific articles. Results: We show how MP makes it possible to carry out semantic annotation of publications that is machine readable and allows for precise search capabilities. In addition, when used by itself, MP serves as a knowledge repository for emerging discoveries. A prototype for proof of concept has been developed that demonstrates the feasibility and novel benefits of MP. As part of the MP framework, we have created an ontology of relationship types with about 100 terms optimized for the representation of scientific assertions. Conclusion: MachineProse is a novel semantic framework that we believe may be used to summarize research findings, annotate biomedical publications, and support sophisticated searches. PMID:16357355

  18. Environmental Research: Communication Studies and Information Sources.

    ERIC Educational Resources Information Center

    Ercegovac, Zorana

    1992-01-01

    Reviews literature on environmental information since 1986, with special emphasis on machine-readable sources as reported in the published literature. Highlights include a new model for studying environmental issues; environmental communication studies, including user studies; and environmental information sources, including pollution media and…

  19. 36 CFR 902.82 - Fee schedule.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... operating duplicating machinery. Not included in direct costs are overhead expenses such as costs of space... form of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g... programs of scholarly research. (5) Non-commercial scientific institution means an institution that is not...

  20. 36 CFR 902.82 - Fee schedule.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... operating duplicating machinery. Not included in direct costs are overhead expenses such as costs of space... form of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g... programs of scholarly research. (5) Non-commercial scientific institution means an institution that is not...

  1. 5 CFR 294.102 - General definitions.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false General definitions. 294.102 Section 294.102 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS AVAILABILITY OF... the forms that such copies can take are paper, microform, audiovisual materials, or machine readable...

  2. 5 CFR 294.102 - General definitions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 5 Administrative Personnel 1 2011-01-01 2011-01-01 false General definitions. 294.102 Section 294.102 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS AVAILABILITY OF... the forms that such copies can take are paper, microform, audiovisual materials, or machine readable...

  3. On the Application of Syntactic Methodologies in Automatic Text Analysis.

    ERIC Educational Resources Information Center

    Salton, Gerard; And Others

    1990-01-01

    Summarizes various linguistic approaches proposed for document analysis in information retrieval environments. Topics discussed include syntactic analysis; use of machine-readable dictionary information; knowledge base construction; the PLNLP English Grammar (PEG) system; phrase normalization; and statistical and syntactic phrase evaluation used…

  4. 36 CFR 1275.16 - Definitions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... pictures, sound and video recordings, machine-readable media, plats, maps, models, pictures, works of art... retained or appropriate for retention as evidence of or information about these powers or duties. Included... Activities or the Watergate Special Prosecution Force; or (2) Are circumscribed in the Articles of...

  5. 36 CFR 1275.16 - Definitions.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... pictures, sound and video recordings, machine-readable media, plats, maps, models, pictures, works of art... retained or appropriate for retention as evidence of or information about these powers or duties. Included... Activities or the Watergate Special Prosecution Force; or (2) Are circumscribed in the Articles of...

  6. 36 CFR 1275.16 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... pictures, sound and video recordings, machine-readable media, plats, maps, models, pictures, works of art... retained or appropriate for retention as evidence of or information about these powers or duties. Included... Activities or the Watergate Special Prosecution Force; or (2) Are circumscribed in the Articles of...

  7. 36 CFR 1275.16 - Definitions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... pictures, sound and video recordings, machine-readable media, plats, maps, models, pictures, works of art... retained or appropriate for retention as evidence of or information about these powers or duties. Included... Activities or the Watergate Special Prosecution Force; or (2) Are circumscribed in the Articles of...

  8. Comparison of Document Data Bases

    ERIC Educational Resources Information Center

    Schipma, Peter B.; And Others

    This paper presents a detailed analysis of the content and format of seven machine-readable bibliographic data bases: Chemical Abstracts Service Condensates, Chemical and Biological Activities, and Polymer Science and Technology, Biosciences Information Service's BA Previews including Biological Abstracts and BioResearch Index, Institute for…

  9. AUTOCLASSIFICATION OF THE VARIABLE 3XMM SOURCES USING THE RANDOM FOREST MACHINE LEARNING ALGORITHM

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Farrell, Sean A.; Murphy, Tara; Lo, Kitty K., E-mail: s.farrell@physics.usyd.edu.au

    In the current era of large surveys and massive data sets, autoclassification of astrophysical sources using intelligent algorithms is becoming increasingly important. In this paper we present the catalog of variable sources in the Third XMM-Newton Serendipitous Source catalog (3XMM) autoclassified using the Random Forest machine learning algorithm. We used a sample of manually classified variable sources from the second data release of the XMM-Newton catalogs (2XMMi-DR2) to train the classifier, obtaining an accuracy of ∼92%. We also evaluated the effectiveness of identifying spurious detections using a sample of spurious sources, achieving an accuracy of ∼95%. Manual investigation of amore » random sample of classified sources confirmed these accuracy levels and showed that the Random Forest machine learning algorithm is highly effective at automatically classifying 3XMM sources. Here we present the catalog of classified 3XMM variable sources. We also present three previously unidentified unusual sources that were flagged as outlier sources by the algorithm: a new candidate supergiant fast X-ray transient, a 400 s X-ray pulsar, and an eclipsing 5 hr binary system coincident with a known Cepheid.« less

  10. Xeml Lab: a tool that supports the design of experiments at a graphical interface and generates computer-readable metadata files, which capture information about genotypes, growth conditions, environmental perturbations and sampling strategy.

    PubMed

    Hannemann, Jan; Poorter, Hendrik; Usadel, Björn; Bläsing, Oliver E; Finck, Alex; Tardieu, Francois; Atkin, Owen K; Pons, Thijs; Stitt, Mark; Gibon, Yves

    2009-09-01

    Data mining depends on the ability to access machine-readable metadata that describe genotypes, environmental conditions, and sampling times and strategy. This article presents Xeml Lab. The Xeml Interactive Designer provides an interactive graphical interface at which complex experiments can be designed, and concomitantly generates machine-readable metadata files. It uses a new eXtensible Mark-up Language (XML)-derived dialect termed XEML. Xeml Lab includes a new ontology for environmental conditions, called Xeml Environment Ontology. However, to provide versatility, it is designed to be generic and also accepts other commonly used ontology formats, including OBO and OWL. A review summarizing important environmental conditions that need to be controlled, monitored and captured as metadata is posted in a Wiki (http://www.codeplex.com/XeO) to promote community discussion. The usefulness of Xeml Lab is illustrated by two meta-analyses of a large set of experiments that were performed with Arabidopsis thaliana during 5 years. The first reveals sources of noise that affect measurements of metabolite levels and enzyme activities. The second shows that Arabidopsis maintains remarkably stable levels of sugars and amino acids across a wide range of photoperiod treatments, and that adjustment of starch turnover and the leaf protein content contribute to this metabolic homeostasis.

  11. 32 CFR 299.2 - Definitions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 32 National Defense 2 2014-07-01 2014-07-01 false Definitions. 299.2 Section 299.2 National Defense Department of Defense (Continued) OFFICE OF THE SECRETARY OF DEFENSE (CONTINUED) FREEDOM OF... compilation, such as all books, papers, maps, and photographs, machine readable materials, including those in...

  12. Machine Readable Bibliographic Records: Criteria and Creation.

    ERIC Educational Resources Information Center

    Bregzis, Ritvars

    The centrality of bibliographic records in library automation, objectives of the bibliographic record file and elemental factors involved in bibliographic record creation are discussed. The practical work of creating bibliographic records involves: (1) data base environment, (2) technical aspects, (3) cost and (4) operational methodology. The…

  13. 36 CFR § 902.82 - Fee schedule.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... operating duplicating machinery. Not included in direct costs are overhead expenses such as costs of space... form of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g... programs of scholarly research. (5) Non-commercial scientific institution means an institution that is not...

  14. 36 CFR § 1275.16 - Definitions.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... pictures, sound and video recordings, machine-readable media, plats, maps, models, pictures, works of art... retained or appropriate for retention as evidence of or information about these powers or duties. Included... Activities or the Watergate Special Prosecution Force; or (2) Are circumscribed in the Articles of...

  15. 8 CFR 217.2 - Eligibility.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... Nationality DEPARTMENT OF HOMELAND SECURITY IMMIGRATION REGULATIONS VISA WAIVER PROGRAM § 217.2 Eligibility... must present a machine-readable passport in order to be granted admission under the Visa Waiver Program. Round trip ticket means any return trip transportation ticket in the name of an arriving Visa Waiver...

  16. Management and Technology Division. Papers.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    Two papers on copyright and privacy considerations of international information transfer were presented at the 1982 International Federation of Library Associations (IFLA) conference. In "Findings of the IFLA International Study on the Copyright of Bibliographic Records in Machine-Readable Form," Dennis D. McDonald, Eleanor Jo Rodger,…

  17. Public Data Set: Radially Scanning Magnetic Probes to Study Local Helicity Injection Dynamics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Richner, Nathan J; Bongard, Michael W; Fonck, Raymond J

    This data set contains openly-documented, machine readable digital research data corresponding to figures published in N.J. Richner et al., 'Radially Scanning Magnetic Probes to Study Local Helicity Injection Dynamics,' accepted for publication in Rev. Sci. Instrum (2018).

  18. GlycoRDF: an ontology to standardize glycomics data in RDF

    PubMed Central

    Ranzinger, Rene; Aoki-Kinoshita, Kiyoko F.; Campbell, Matthew P.; Kawano, Shin; Lütteke, Thomas; Okuda, Shujiro; Shinmachi, Daisuke; Shikanai, Toshihide; Sawaki, Hiromichi; Toukach, Philip; Matsubara, Masaaki; Yamada, Issaku; Narimatsu, Hisashi

    2015-01-01

    Motivation: Over the last decades several glycomics-based bioinformatics resources and databases have been created and released to the public. Unfortunately, there is no common standard in the representation of the stored information or a common machine-readable interface allowing bioinformatics groups to easily extract and cross-reference the stored information. Results: An international group of bioinformatics experts in the field of glycomics have worked together to create a standard Resource Description Framework (RDF) representation for glycomics data, focused on glycan sequences and related biological source, publications and experimental data. This RDF standard is defined by the GlycoRDF ontology and will be used by database providers to generate common machine-readable exports of the data stored in their databases. Availability and implementation: The ontology, supporting documentation and source code used by database providers to generate standardized RDF are available online (http://www.glycoinfo.org/GlycoRDF/). Contact: rene@ccrc.uga.edu or kkiyoko@soka.ac.jp Supplementary information: Supplementary data are available at Bioinformatics online. PMID:25388145

  19. GlycoRDF: an ontology to standardize glycomics data in RDF.

    PubMed

    Ranzinger, Rene; Aoki-Kinoshita, Kiyoko F; Campbell, Matthew P; Kawano, Shin; Lütteke, Thomas; Okuda, Shujiro; Shinmachi, Daisuke; Shikanai, Toshihide; Sawaki, Hiromichi; Toukach, Philip; Matsubara, Masaaki; Yamada, Issaku; Narimatsu, Hisashi

    2015-03-15

    Over the last decades several glycomics-based bioinformatics resources and databases have been created and released to the public. Unfortunately, there is no common standard in the representation of the stored information or a common machine-readable interface allowing bioinformatics groups to easily extract and cross-reference the stored information. An international group of bioinformatics experts in the field of glycomics have worked together to create a standard Resource Description Framework (RDF) representation for glycomics data, focused on glycan sequences and related biological source, publications and experimental data. This RDF standard is defined by the GlycoRDF ontology and will be used by database providers to generate common machine-readable exports of the data stored in their databases. The ontology, supporting documentation and source code used by database providers to generate standardized RDF are available online (http://www.glycoinfo.org/GlycoRDF/). © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  20. Engineering design knowledge recycling in near-real-time

    NASA Technical Reports Server (NTRS)

    Leifer, Larry; Baya, Vinod; Toye, George; Baudin, Catherine; Underwood, Jody Gevins

    1994-01-01

    It is hypothesized that the capture and reuse of machine readable design records is cost beneficial. This informal engineering notebook design knowledge can be used to model the artifact and the design process. Design rationale is, in part, preserved and available for examination. Redesign cycle time is significantly reduced (Baya et al, 1992). These factors contribute to making it less costly to capture and reuse knowledge than to recreate comparable knowledge (current practice). To test the hypothesis, we have focused on validation of the concept and tools in two 'real design' projects this past year: (1) a short (8 month) turnaround project for NASA life science bioreactor researchers was done by a team of three mechanical engineering graduate students at Stanford University (in a class, ME210abc 'Mechatronic Systems Design and Methodology' taught by one of the authors, Leifer); and (2) a long range (8 to 20 year) international consortium project for NASA's Space Science program (STEP: satellite test of the equivalence principle). Design knowledge capture was supported this year by assigning the use of a Team-Design PowerBook. Design records were cataloged in near-real time. These records were used to qualitatively model the artifact design as it evolved. Dedal, an 'intelligent librarian' developed at NASA-ARC, was used to navigate and retrieve captured knowledge for reuse.

  1. Explosive Growth and Advancement of the NASA/IPAC Extragalactic Database (NED)

    NASA Astrophysics Data System (ADS)

    Mazzarella, Joseph M.; Ogle, P. M.; Fadda, D.; Madore, B. F.; Ebert, R.; Baker, K.; Chan, H.; Chen, X.; Frayer, C.; Helou, G.; Jacobson, J. D.; LaGue, C.; Lo, T. M.; Pevunova, O.; Schmitz, M.; Terek, S.; Steer, I.

    2014-01-01

    The NASA/IPAC Extragalactic Database (NED) is continuing to evolve in lock-step with the explosive growth of astronomical data and advancements in information technology. A new methodology is being used to fuse data from very large surveys. Selected parameters are first loaded into a new database layer and made available in areal searches before they are cross-matched with prior NED objects. Then a programmed, rule-based statistical approach is used to identify new objects and compute cross-identifications with existing objects where possible; otherwise associations between objects are derived based on positional uncertainties or spatial resolution differences. Approximately 62 million UV sources from the GALEX All-Sky Survey and Medium Imaging Survey catalogs have been integrated into NED using this new process. The December 2013 release also contains nearly half a billion sources from the 2MASS Point Source Catalog accessible in cone searches, while the large scale cross-matching is in progress. Forthcoming updates will fuse data from All-WISE, SDSS DR12, and other very large catalogs. This work is progressing in parallel with the equally important integration of data from the literature, which is also growing rapidly. Recent updates have also included H I and CO channel maps (data cubes), as well as substantial growth in redshifts, classifications, photometry, spectra and redshift-independent distances. The By Parameters search engine now incorporates a simplified form for entry of constraints, and support for long-running queries with machine-readable output. A new tool for exploring the environments of galaxies with measured radial velocities includes informative graphics and a method to assess the incompleteness of redshift measurements. The NED user interface is also undergoing a major transformation, providing more streamlined navigation and searching, and a modern development framework for future enhancements. For further information, please visit our poster (Fadda et al. 2014) and stop by the NED exhibit for a demo. NED is operated by the Jet Propulsion Laboratory, California Institute of Technology, under contract with the National Aeronautics and Space Administration.

  2. 77 FR 72337 - Apps for Vehicles Challenge

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-12-05

    ... innovation while rigorously protecting privacy. The primary fuel for the Energy Data Initiative is open data. Open data can take many forms but generally includes information that is machine-readable, freely accessible and in an industry-standard format. In particular, open data from the private sector made...

  3. 32 CFR 518.7 - FOIA terms defined.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... books, papers, maps, photographs, and machine readable materials, inclusive of those in electronic form... create or compile a record to satisfy a FOIA request. (3) Hard copy or electronic records that are... conduct. (h) Electronic record. Records (including e-mail) that are created, stored, and retrievable by...

  4. 46 CFR 503.43 - Fees for services.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... Information Act or other request. Such copies can take the form of paper or machine readable documentation (e..., which operates a program or programs of scholarly research. (6) Non-commercial scientific institution... research the results of which are not intended to promote any particular product or industry. (7...

  5. 46 CFR 503.43 - Fees for services.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... Information Act or other request. Such copies can take the form of paper or machine readable documentation (e..., which operates a program or programs of scholarly research. (6) Non-commercial scientific institution... research the results of which are not intended to promote any particular product or industry. (7...

  6. Public Data Set: H-mode Plasmas at Very Low Aspect Ratio on the Pegasus Toroidal Experiment

    DOE Data Explorer

    Thome, Kathreen E. [University of Wisconsin-Madison; Oak Ridge Associated Universities] (ORCID:0000000248013922); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Barr, Jayson L. [University of Wisconsin-Madison] (ORCID:0000000177685931); Bodner, Grant M. [University of Wisconsin-Madison] (ORCID:0000000324979172); Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Kriete, David M. [University of Wisconsin-Madison] (ORCID:0000000236572911); Perry, Justin M. [University of Wisconsin-Madison] (ORCID:0000000171228609); Reusch, Joshua A. [University of Wisconsin-Madison] (ORCID:0000000284249422); Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448)

    2016-09-30

    This data set contains openly-documented, machine readable digital research data corresponding to figures published in K.E. Thome et al., 'H-mode Plasmas at Very Low Aspect Ratio on the Pegasus Toroidal Experiment,' Nucl. Fusion 57, 022018 (2017).

  7. Recent Developments in Social Science Research

    ERIC Educational Resources Information Center

    Jenness, David

    1978-01-01

    In this discussion of recent theoretical and methodological developments in three selected areas of the social sciences--social indicators, social experimentation and evaluation, and longitudinal studies, attention is given to the growing availability of machine-readable data sets and to special archives and resources of interest to librarians.…

  8. 41 CFR 51-8.3 - Definitions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... request. Such copies can take the form of paper copy, audio-visual materials, or machine readable materials (e.g., magnetic tape or disk), among others. (g) The term search includes all time spent looking... time spent resolving general legal or policy issues regarding the application of exemptions. [54 FR...

  9. 41 CFR 51-8.3 - Definitions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... request. Such copies can take the form of paper copy, audio-visual materials, or machine readable materials (e.g., magnetic tape or disk), among others. (g) The term search includes all time spent looking... time spent resolving general legal or policy issues regarding the application of exemptions. [54 FR...

  10. Public Data Set: Impedance of an Intense Plasma-Cathode Electron Source for Tokamak Plasma Startup

    DOE Data Explorer

    Hinson, Edward T. [University of Wisconsin-Madison] (ORCID:000000019713140X); Barr, Jayson L. [University of Wisconsin-Madison] (ORCID:0000000177685931); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Perry, Justin M. [University of Wisconsin-Madison] (ORCID:0000000171228609)

    2016-05-31

    This data set contains openly-documented, machine readable digital research data corresponding to figures published in E.T. Hinson et al., 'Impedance of an Intense Plasma-Cathode Electron Source for Tokamak Plasma Startup,' Physics of Plasmas 23, 052515 (2016).

  11. 48 CFR 252.211-7003 - Item identification and valuation.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ..., used to retrieve data encoded on machine-readable media. Concatenated unique item identifier means— (1... Defense Logistics Information System (DLIS) Commercial and Government Entity (CAGE) Code). Issuing agency... identifier. Item means a single hardware article or a single unit formed by a grouping of subassemblies...

  12. Public Data Set: Control and Automation of the Pegasus Multi-point Thomson Scattering System

    DOE Data Explorer

    Bodner, Grant M. [University of Wisconsin-Madison] (ORCID:0000000324979172); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Reusch, Joshua A. [University of Wisconsin-Madison] (ORCID:0000000284249422); Rodriguez Sanchez, Cuauhtemoc [University of Wisconsin-Madison] (ORCID:0000000334712586); Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448)

    2016-08-12

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in G.M. Bodner et al., 'Control and Automation of the Pegasus Multi-point Thomson Scattering System,' Rev. Sci. Instrum. 87, 11E523 (2016).

  13. 23 CFR Appendix A to Part 1313 - Tamper Resistant Driver's License

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ...) Block graphics. (15) Security fonts and graphics with known hidden flaws. (16) Card stock, layer with colors. (17) Micro-graphics. (18) Retroflective security logos. (19) Machine readable technologies such... permit that has one or more of the following security features: (1) Ghost image. (2) Ghost graphic. (3...

  14. Developing a Large Lexical Database for Information Retrieval, Parsing, and Text Generation Systems.

    ERIC Educational Resources Information Center

    Conlon, Sumali Pin-Ngern; And Others

    1993-01-01

    Important characteristics of lexical databases and their applications in information retrieval and natural language processing are explained. An ongoing project using various machine-readable sources to build a lexical database is described, and detailed designs of individual entries with examples are included. (Contains 66 references.) (EAM)

  15. MONTO: A Machine-Readable Ontology for Teaching Word Problems in Mathematics

    ERIC Educational Resources Information Center

    Lalingkar, Aparna; Ramnathan, Chandrashekar; Ramani, Srinivasan

    2015-01-01

    The Indian National Curriculum Framework has as one of its objectives the development of mathematical thinking and problem solving ability. However, recent studies conducted in Indian metros have expressed concern about students' mathematics learning. Except in some private coaching academies, regular classroom teaching does not include problem…

  16. Use of PL/1 in a Bibliographic Information Retrieval System.

    ERIC Educational Resources Information Center

    Schipma, Peter B.; And Others

    The Information Sciences section of ITT Research Institute (IITRI) has developed a Computer Search Center and is currently conducting a research project to explore computer searching of a variety of machine-readable data bases. The Center provides Selective Dissemination of Information services to academic, industrial and research organizations…

  17. Public Data Set: A Power-Balance Model for Local Helicity Injection Startup in a Spherical Tokamak

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Barr, Jayson L.; Bongard, Michael W.; Burke, Marcus G.

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in J.L. Barr et. al, 'A Power-Balance Model for Local Helicity Injection Startup in a Spherical Tokamak,' Nuclear Fusion 58, 076011 (2018).

  18. 21 CFR 1401.3 - Definitions.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 21 Food and Drugs 9 2010-04-01 2010-04-01 false Definitions. 1401.3 Section 1401.3 Food and Drugs OFFICE OF NATIONAL DRUG CONTROL POLICY PUBLIC AVAILABILITY OF INFORMATION § 1401.3 Definitions. For the... paper, microform, audio-visual materials, or machine-readable documentation. ONDCP will provide a copy...

  19. An Operational System for Subject Switching between Controlled Vocabularies: A Computational Linguistics Approach.

    ERIC Educational Resources Information Center

    Silvester, June P.; And Others

    This report describes a new automated process that pioneers full-scale operational use of subject switching by the NASA (National Aeronautics and Space Administration) Scientific and Technical Information (STI) Facility. The subject switching process routinely translates machine-readable subject terms from one controlled vocabulary into the…

  20. A Model-Driven Approach to e-Course Management

    ERIC Educational Resources Information Center

    Savic, Goran; Segedinac, Milan; Milenkovic, Dušica; Hrin, Tamara; Segedinac, Mirjana

    2018-01-01

    This paper presents research on using a model-driven approach to the development and management of electronic courses. We propose a course management system which stores a course model represented as distinct machine-readable components containing domain knowledge of different course aspects. Based on this formally defined platform-independent…

  1. Method and system for enabling real-time speckle processing using hardware platforms

    NASA Technical Reports Server (NTRS)

    Ortiz, Fernando E. (Inventor); Kelmelis, Eric (Inventor); Durbano, James P. (Inventor); Curt, Peterson F. (Inventor)

    2012-01-01

    An accelerator for the speckle atmospheric compensation algorithm may enable real-time speckle processing of video feeds that may enable the speckle algorithm to be applied in numerous real-time applications. The accelerator may be implemented in various forms, including hardware, software, and/or machine-readable media.

  2. Common Bibliographic Standards for Baylor University Libraries. Revised.

    ERIC Educational Resources Information Center

    Scott, Sharon; And Others

    Developed by a Baylor University (Texas) Task Force, the revised policies of bibliographic standards for the university libraries provide formats for: (1) archives and manuscript control; (2) audiovisual media; (3) books; (4) machine-readable data files; (5) maps; (6) music scores; (7) serials; and (8) sound recordings. The task force assumptions…

  3. 14 CFR 221.500 - Transmission of electronic tariffs to subscribers.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... TRANSPORTATION (AVIATION PROCEEDINGS) ECONOMIC REGULATIONS TARIFFS Electronically Filed Tariffs § 221.500... to any subscriber to the on-line tariff database, including access to the justification required by... machine-readable data (raw tariff data) of all daily transactions made to its on-line tariff database. The...

  4. 19 CFR 201.20 - Fees.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... requesters, subject to the limitations of paragraph (c) of this section. For a paper photocopy of a record... overhead expenses such as costs of space and heating or lighting of the facility in which the records are... of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g., magnetic...

  5. 19 CFR 201.20 - Fees.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... requesters, subject to the limitations of paragraph (c) of this section. For a paper photocopy of a record... overhead expenses such as costs of space and heating or lighting of the facility in which the records are... of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g., magnetic...

  6. 19 CFR 201.20 - Fees.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... requesters, subject to the limitations of paragraph (c) of this section. For a paper photocopy of a record... overhead expenses such as costs of space and heating or lighting of the facility in which the records are... of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g., magnetic...

  7. 19 CFR 201.20 - Fees.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... requesters, subject to the limitations of paragraph (c) of this section. For a paper photocopy of a record... overhead expenses such as costs of space and heating or lighting of the facility in which the records are... of paper copy, microform, audio-visual materials, or machine-readable documentation (e.g., magnetic...

  8. 32 CFR 1662.6 - Fee schedule; waiver of fees.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... as costs of space, and heating or lighting the facility in which the records are stored. (2) The term... copies may take the form of paper copy, microform, audio-visual materials, or machine readable... institution of vocational education, which operates a program or programs of scholarly research. (7) The term...

  9. 32 CFR 1662.6 - Fee schedule; waiver of fees.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... as costs of space, and heating or lighting the facility in which the records are stored. (2) The term... copies may take the form of paper copy, microform, audio-visual materials, or machine readable... institution of vocational education, which operates a program or programs of scholarly research. (7) The term...

  10. 77 FR 64409 - Designation of Taiwan for the Visa Waiver Program

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-22

    ... and passport holders from designated Visa Waiver Program countries \\1\\ may apply for admission to the... for nationals of the country; (2) a government certification that it issues machine-readable passports... about the theft or loss of passports; (5) the government acceptance for repatriation any citizen, former...

  11. DOBIS and NOTIS: A Contrast in Design.

    ERIC Educational Resources Information Center

    Juergens, Bonnie; Blake, Ruth

    1987-01-01

    Compares and contrasts two systems designed for library automation applications--NOTIS, which was developed in the United States, and DOBIS, which was developed in Europe. The differences in the systems are discussed in terms of the availability or absence of machine readable bibliographic sharing capacities in the countries of origin. (CLB)

  12. 5 CFR 294.102 - General definitions.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... the forms that such copies can take are paper, microform, audiovisual materials, or machine readable... that is necessary to excise them and otherwise prepare them for release. Review does not include time... the time spent looking for material that is responsive to a request, including page-by-page or line-by...

  13. ARL Statement on Unlimited Use and Exchange of Bibliographic Records.

    ERIC Educational Resources Information Center

    Association of Research Libraries, Washington, DC.

    The Association of Research Libraries is fully committed to the principle of unrestricted access to and dissemination of ideas, i.e., member libraries must have unlimited access to the machine-readable bibliographic records which are created by member libraries and maintained in bibliographic utilities. Coordinated collection development programs…

  14. Optical Scanning for Retrospective Conversion of Information.

    ERIC Educational Resources Information Center

    Hein, Morten

    1986-01-01

    This discussion of the use of optical scanning and computer formatting for retrospective conversion focuses on a series of applications known as Optical Scanning for Creation of Information Databases (OSCID). Prior research in this area and the usefulness of OSCID for creating low-priced machine-readable data representing older materials are…

  15. Retrospective Conversion: A Question of Time, Standards, and Purpose.

    ERIC Educational Resources Information Center

    Valentine, Phyllis A.; McDonald, David R.

    1986-01-01

    Examines the factors that determine the cost of retrospective conversion (definition of conversion, standards of acceptance, method of conversion, hit rate, standards for creation of machine-readable records for nonhits); reports results of cost study at University of Michigan library; and introduces an alternative strategy for discussion. Seven…

  16. Union Listing via OCLC's Serials Control Subsystem.

    ERIC Educational Resources Information Center

    O'Malley, Terrence J.

    1984-01-01

    Describes library use of Conversion of Serials Project's (CONSER) online national machine-readable database for serials to create online union lists of serials via OCLC's Serial Control Subsystem. Problems in selection of appropriate, accurate, and authenticated records and prospects for the future are discussed. Twenty sources and sample records…

  17. Visualization of Learning Scenarios with UML4LD

    ERIC Educational Resources Information Center

    Laforcade, Pierre

    2007-01-01

    Present Educational Modelling Languages are used to formally specify abstract learning scenarios in a machine-interpretable format. Current tooling does not provide teachers/designers with some graphical facilities to help them in reusing existent scenarios. They need human-readable representations. This paper discusses the UML4LD experimental…

  18. COMPENDEX/TEXT-PAC: RETROSPECTIVE SEARCH.

    ERIC Educational Resources Information Center

    Standera, Oldrich

    The Text-Pac System is capable of generating indexes and bulletins to provide a current information service without the selectivity feature. Indexes of the accumulated data base may also be used as a basis for manual retrospective searching. The manual search involves searching computer-prepared indexes from a machine readable data base produced…

  19. Provision of Information to the Research Staff.

    ERIC Educational Resources Information Center

    Williams, Martha E.

    The Information Sciences section at Illinois Institute of Technology Research Institute (IITRI) is now operating a Computer Search Center (CSC) for handling numerous machine-readable data bases. The computer programs are generalized in the sense that they will handle any incoming data base. This is accomplished by means of a preprocessor system…

  20. The Bibliographical Control of Early Books.

    ERIC Educational Resources Information Center

    Cameron, William J.

    Examples are given of the kinds of machine-readable data bases that should be developed in order to extend attempts at universal bibliographical control into neglected areas, the results of which can be used by researchers in the humanities, specifically those using books printed before 1801. The principles of bibliographical description,…

  1. The IHMC CmapTools software in research and education: a multi-level use case in Space Meteorology

    NASA Astrophysics Data System (ADS)

    Messerotti, Mauro

    2010-05-01

    The IHMC (Institute for Human and Machine Cognition, Florida University System, USA) CmapTools software is a powerful multi-platform tool for knowledge modelling in graphical form based on concept maps. In this work we present its application for the high-level development of a set of multi-level concept maps in the framework of Space Meteorology to act as the kernel of a space meteorology domain ontology. This is an example of a research use case, as a domain ontology coded in machine-readable form via e.g. OWL (Web Ontology Language) is suitable to be an active layer of any knowledge management system embedded in a Virtual Observatory (VO). Apart from being manageable at machine level, concept maps developed via CmapTools are intrinsically human-readable and can embed hyperlinks and objects of many kinds. Therefore they are suitable to be published on the web: the coded knowledge can be exploited for educational purposes by the students and the public, as the level of information can be naturally organized among linked concept maps in progressively increasing complexity levels. Hence CmapTools and its advanced version COE (Concept-map Ontology Editor) represent effective and user-friendly software tools for high-level knowledge represention in research and education.

  2. Census of the Local Universe (CLU): Classification of Galaxy Candidates in Narrowband Images Using Machine Learning

    NASA Astrophysics Data System (ADS)

    Zhang, Chaoran; Van Sistine, Anglea; Kaplan, David; Brady, Patrick; Cook, David O.; Kasliwal, Mansi

    2018-01-01

    A complete catalog of galaxies in the local universe is critical for efficient electromagnetic follow-up of gravitational wave events (EMGW). The Census of the Local Universe (CLU; Cook et al. 2017, in preparation) aims to provide a galaxy catalog out to 200 Mpc that is as complete as possible. CLU has recently completed an Hα survey of ~3π of the sky with the goal of cataloging those galaxies that are likely hosts of EMGW events. Here, we present a tool we developed using machine learning technology to classify sources extracted from the Hα narrowband images within 200Mpc. In this analysis we find we are able to recover more galaxies compared to selections based on Hα colors alone.

  3. Development of clinical contents model markup language for electronic health records.

    PubMed

    Yun, Ji-Hyun; Ahn, Sun-Ju; Kim, Yoon

    2012-09-01

    To develop dedicated markup language for clinical contents models (CCM) to facilitate the active use of CCM in electronic health record systems. Based on analysis of the structure and characteristics of CCM in the clinical domain, we designed extensible markup language (XML) based CCM markup language (CCML) schema manually. CCML faithfully reflects CCM in both the syntactic and semantic aspects. As this language is based on XML, it can be expressed and processed in computer systems and can be used in a technology-neutral way. CCML HAS THE FOLLOWING STRENGTHS: it is machine-readable and highly human-readable, it does not require a dedicated parser, and it can be applied for existing electronic health record systems.

  4. Enhancing and Archiving the APS Catalog of the POSS I

    NASA Technical Reports Server (NTRS)

    Humphreys, Roberta M.

    2003-01-01

    We have worked on two different projects: 1) Archiving the APS Catalog of the POSS I for distribution to NASA's NED at IPAC, SIMBAD in France, and individual astronomers and 2) The automated morphological classification of galaxies. We have completed archiving the Catalog into easily readable binary files. The database together with the software to read it has been distributed on DVD's to the national and international data centers and to individual astronomers. The archived Catalog contains more than 89 million objects in 632 fields in the first epoch Palomar Observatory Sky Survey. Additional image parameters not available in the original on-line version are also included in the archived version. The archived Catalog is also available and can be queried at the APS web site (URL: http://aps.umn.edu) which has been improved with a much faster and more efficient querying system. The Catalog can be downloaded as binary datafiles with the source code for reading it. It is also being integrated into the SkyQuery system which includes the Sloan Digital Sky Survey, 2MASS, and the FIRST radio sky survey. We experimented with different classification algorithms to automate the morphological classification of galaxies. This is an especially difficult problem because there are not only a large number of attributes or parameters and measurement uncertainties, but also the added complication of human disagreement about the adopted types. To solve this problem we used 837 galaxy images from nine POSS I fields at the North Galactic Pole classified by two independent astronomers for which they agree on the morphological types. The initial goal was to separate the galaxies into the three broad classes relevant to issues of large scale structure and galaxy formation and evolution: early (ellipticals and lenticulars), spirals, and late (irregulars) with an accuracy or success rate that rivals the best astronomer classifiers. We also needed to identify a set of parameters derived from the digitized images that separate the galaxies by type. The human eye can easily recognize complicated patterns in images such as spiral arms which can be spotty, blotchy affairs that are difficult for automated techniques. A galaxy image can potentially be described by hundreds of parameters, all of which may have some relation to the morphological type. In the set of initial experiments we used 624 such parameters, in two colors, blue and red. These parameters include the surface brightness and color measured at different radii, ratios of these parameters at different radii, concentration indices, Fourier transforms and wavelet decomposition coefficients. We experimented with three different classes of classification algorithms; decision trees, k-nearest neighbors, and support vector machines (SVM). A range of experiments were conducted and we eventually narrowed the parameters to 23 selected parameters. SVM consistently outperformed the other algorithms with both sets of features. By combining the results from the different algorithms in a weighted scheme we achieved an overall classification success of 86%.

  5. 26 CFR 301.6721-1 - Failure to file correct information returns.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... file timely includes a failure to file in the required manner, for example, on magnetic media or in... they fall below the 250-threshold requirement) or on magnetic media or other machine-readable form. Filers who are required to file information returns on magnetic media and who file such information...

  6. 26 CFR 1.197-2 - Amortization of goodwill and certain other intangibles.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ..., process, design, pattern, know-how, format, package design, computer software (as defined in paragraph (c... agreement that provides one of the parties to the agreement with the right to distribute, sell, or provide... any program or routine (that is, any sequence of machine-readable code) that is designed to cause a...

  7. Criteria for Labelling Prosodic Aspects of English Speech.

    ERIC Educational Resources Information Center

    Bagshaw, Paul C.; Williams, Briony J.

    A study reports a set of labelling criteria which have been developed to label prosodic events in clear, continuous speech, and proposes a scheme whereby this information can be transcribed in a machine readable format. A prosody in a syllabic domain which is synchronized with a phonemic segmentation was annotated. A procedural definition of…

  8. Facilitating knowledge discovery and visualization through mining contextual data from published studies: lessons from JournalMap

    USDA-ARS?s Scientific Manuscript database

    Valuable information on the location and context of ecological studies are locked up in publications in myriad formats that are not easily machine readable. This presents significant challenges to building geographic-based tools to search for and visualize sources of ecological knowledge. JournalMap...

  9. PASCAL Data Base: File Description and On Line Access on ESA/IRS.

    ERIC Educational Resources Information Center

    Pelissier, Denise

    This report describes the PASCAL database, a machine readable version of the French abstract journal Bulletin Signaletique, which allows use of the file for (1) batch and online retrieval of information, (2) selective dissemination of information, and (3) publishing of the 50 sections of Bulletin Signaletique. The system, which covers nine…

  10. 26 CFR 301.6721-1 - Failure to file correct information returns.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... file timely includes a failure to file in the required manner, for example, on magnetic media or in... they fall below the 250-threshold requirement) or on magnetic media or other machine-readable form. Filers who are required to file information returns on magnetic media and who file such information...

  11. Automatic Selection of Suitable Sentences for Language Learning Exercises

    ERIC Educational Resources Information Center

    Pilán, Ildikó; Volodina, Elena; Johansson, Richard

    2013-01-01

    In our study we investigated second and foreign language (L2) sentence readability, an area little explored so far in the case of several languages, including Swedish. The outcome of our research consists of two methods for sentence selection from native language corpora based on Natural Language Processing (NLP) and machine learning (ML)…

  12. MARC Data, the OPAC, and Library Professionals

    ERIC Educational Resources Information Center

    Williams, Jo

    2009-01-01

    Purpose: The purpose of this paper is to show that knowledge of the Machine-Readable Cataloguing (MARC) format is useful in all aspects of librarianship, not just for cataloguing, and how MARC knowledge can address indexing limitations of the online catalogue. Design/methodology/approach: The paper employs examples and scenarios to show the…

  13. Microcomputer-Based Access to Machine-Readable Numeric Databases.

    ERIC Educational Resources Information Center

    Wenzel, Patrick

    1988-01-01

    Describes the use of microcomputers and relational database management systems to improve access to numeric databases by the Data and Program Library Service at the University of Wisconsin. The internal records management system, in-house reference tools, and plans to extend these tools to the entire campus are discussed. (3 references) (CLB)

  14. 29 CFR 1208.6 - Schedule of fees and methods of payment for services rendered.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... included in direct costs are overhead expenses such as costs of space and heating or lighting the facility... form of paper copy, microfilm, audiovisual materials, or machine readable documentation (e.g., magnetic... scholarly research. (7) Non-commercial scientific institution refers to an institution that is not operated...

  15. 29 CFR 1208.6 - Schedule of fees and methods of payment for services rendered.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... included in direct costs are overhead expenses such as costs of space and heating or lighting the facility... form of paper copy, microfilm, audiovisual materials, or machine readable documentation (e.g., magnetic... scholarly research. (7) Non-commercial scientific institution refers to an institution that is not operated...

  16. 45 CFR 704.1 - Material available pursuant to 5 U.S.C. 552.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    .... Not included in direct costs are overhead expenses such as costs of space and heating or lighting the... of records. Such copies can take the form of paper or machine readable documentation (e.g., magnetic... vocational education that operates a program or programs of scholarly research. (vii) Noncommercial...

  17. 29 CFR 1208.6 - Schedule of fees and methods of payment for services rendered.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... included in direct costs are overhead expenses such as costs of space and heating or lighting the facility... form of paper copy, microfilm, audiovisual materials, or machine readable documentation (e.g., magnetic... scholarly research. (7) Non-commercial scientific institution refers to an institution that is not operated...

  18. 45 CFR 704.1 - Material available pursuant to 5 U.S.C. 552.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    .... Not included in direct costs are overhead expenses such as costs of space and heating or lighting the... of records. Such copies can take the form of paper or machine readable documentation (e.g., magnetic... vocational education that operates a program or programs of scholarly research. (vii) Noncommercial...

  19. 45 CFR 704.1 - Material available pursuant to 5 U.S.C. 552.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    .... Not included in direct costs are overhead expenses such as costs of space and heating or lighting the... of records. Such copies can take the form of paper or machine readable documentation (e.g., magnetic... vocational education that operates a program or programs of scholarly research. (vii) Noncommercial...

  20. 45 CFR 704.1 - Material available pursuant to 5 U.S.C. 552.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    .... Not included in direct costs are overhead expenses such as costs of space and heating or lighting the... of records. Such copies can take the form of paper or machine readable documentation (e.g., magnetic... vocational education that operates a program or programs of scholarly research. (vii) Noncommercial...

  1. Nonbibliographic Machine-Readable Data Bases in ARL Libraries. SPEC Kit 105.

    ERIC Educational Resources Information Center

    Westerman, Mel

    This document is one of ten kits distributed annually by the Systems and Procedures Exchange Center (SPEC), a clearinghouse operated by the Association of Research Libraries, Office of Management Studies (ARL/OMS) that provides a central source of timely information and materials on the management and operations of large academic and research…

  2. Public Data Set: A Novel, Cost-Effective, Multi-Point Thomson Scattering System on the Pegasus Toroidal Experiment

    DOE Data Explorer

    Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448); Bodner, Grant M. [University of Wisconsin-Madison] (ORCID:0000000324979172); Reusch, Joshua A. [University of Wisconsin-Madison] (ORCID:0000000284249422); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Rodriguez Sanchez, Cuauhtemoc [University of Wisconsin-Madison] (ORCID:0000000334712586)

    2016-09-16

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in D.J. Schlossberg et. al., 'A Novel, Cost-Effective, Multi-Point Thomson Scattering System on the Pegasus Toroidal Experiment,' Rev. Sci. Instrum. 87, 11E403 (2016).

  3. Public Data Set: High Confinement Mode and Edge Localized Mode Characteristics in a Near-Unity Aspect Ratio Tokamak

    DOE Data Explorer

    Thome, Kathreen E. [University of Wisconsin-Madison] (ORCID:0000000248013922); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Barr, Jayson L. [University of Wisconsin-Madison] (ORCID:0000000177685931); Bodner, Grant M. [University of Wisconsin-Madison] (ORCID:0000000324979172); Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Kriete, David M. [University of Wisconsin-Madison] (ORCID:0000000236572911); Perry, Justin M. [University of Wisconsin-Madison] (ORCID:0000000171228609); Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448)

    2016-04-27

    This data set contains openly-documented, machine readable digital research data corresponding to figures published in K.E. Thome et al., 'High Confinement Mode and Edge Localized Mode Characteristics in a Near-Unity Aspect Ratio Tokamak,' Phys. Rev. Lett. 116, 175001 (2016).

  4. Functional and Software Considerations for Bibliographic Data Base Utilization.

    ERIC Educational Resources Information Center

    Cadwallader, Gouverneur

    This is the fourth in a series of eight reports of a research study for the National Agricultural Library (NAL) on the effective utilization of bibliographic data bases in machine-readable form. It describes the general functional and software requirements of an NAL system using external sources of bibliographic data. Various system design…

  5. 26 CFR 301.6721-1 - Failure to file correct information returns.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... file timely includes a failure to file in the required manner, for example, on magnetic media or in... they fall below the 250-threshold requirement) or on magnetic media or other machine-readable form. Filers who are required to file information returns on magnetic media and who file such information...

  6. 26 CFR 301.6721-1 - Failure to file correct information returns.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... file timely includes a failure to file in the required manner, for example, on magnetic media or in... they fall below the 250-threshold requirement) or on magnetic media or other machine-readable form. Filers who are required to file information returns on magnetic media and who file such information...

  7. Inventory of U.S. Health Care Data Bases, 1976-1987.

    ERIC Educational Resources Information Center

    Kralovec, Peter D.; Andes, Steven M.

    This inventory contains summary abstracts of 305 current (1976-1987) non-bibliographic machine-readable databases and national health care data that have been created by public and private organizations throughout the United States. Each of the abstracts contains pertinent information on the sponsor or database, a description of the purpose and…

  8. 28 CFR 51.20 - Form of submissions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... set. A separate data dictionary file documenting the fields in the data set, the field separators or... data set. Proprietary or commercial software system data files (e.g., SAS, SPSS, dBase, Lotus 1-2-3... General will accept certain machine readable data in the following electronic media: 3.5 inch 1.4 megabyte...

  9. 28 CFR 51.20 - Form of submissions.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... set. A separate data dictionary file documenting the fields in the data set, the field separators or... data set. Proprietary or commercial software system data files (e.g., SAS, SPSS, dBase, Lotus 1-2-3... General will accept certain machine readable data in the following electronic media: 3.5 inch 1.4 megabyte...

  10. 28 CFR 51.20 - Form of submissions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... set. A separate data dictionary file documenting the fields in the data set, the field separators or... data set. Proprietary or commercial software system data files (e.g., SAS, SPSS, dBase, Lotus 1-2-3... General will accept certain machine readable data in the following electronic media: 3.5 inch 1.4 megabyte...

  11. Corpus Linguistics for Korean Language Learning and Teaching. NFLRC Technical Report No. 26

    ERIC Educational Resources Information Center

    Bley-Vroman, Robert, Ed.; Ko, Hyunsook, Ed.

    2006-01-01

    Dramatic advances in personal computer technology have given language teachers access to vast quantities of machine-readable text, which can be analyzed with a view toward improving the basis of language instruction. Corpus linguistics provides analytic techniques and practical tools for studying language in use. This volume includes both an…

  12. Public Data Set: Initiation and Sustainment of Tokamak Plasmas with Local Helicity Injection as the Majority Current Drive

    DOE Data Explorer

    Perry, Justin M. [University of Wisconsin-Madison] (ORCID:0000000171228609); Bodner, Grant M. [University of Wisconsin-Madison] (ORCID:0000000324979172); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Pachicano, Jessica L. [University of Wisconsin-Madison] (ORCID:0000000207255693); Pierren, Christopher [University of Wisconsin-Madison] (ORCID:0000000228289825); Reusch, Joshua A. [University of Wisconsin-Madison] (ORCID:0000000284249422); Rhodes, Alexander T. [University of Wisconsin-Madison] (ORCID:0000000280735714); Richner, Nathan J. [University of Wisconsin-Madison] (ORCID:0000000155443915); Rodriguez Sanchez, Cuauhtemoc [University of Wisconsin-Madison] (ORCID:0000000334712586); Schaefer, Carolyn E. [University of Wisconsin-Madison] (ORCID:0000000248848727); Weberski, Justin D. [University of Wisconsin-Madison] (ORCID:0000000256267914)

    2018-05-22

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in J.M. Perry et al., 'Initiation and Sustainment of Tokamak Plasmas with Local Helicity Injection as the Majority Current Drive,' accepted for publication in Nuclear Fusion.

  13. Public Data Set: Non-inductively Driven Tokamak Plasmas at Near-Unity βt in the Pegasus Toroidal Experiment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Reusch, Joshua A.; Bodner, Grant M.; Bongard, Michael W.

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in J.A. Reusch et al., 'Non-inductively Driven Tokamak Plasmas at Near-Unity βt in the Pegasus Toroidal Experiment,' Phys. Plasmas 25, 056101 (2018).

  14. 19 CFR 163.5 - Methods for storage of records.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... standard business practice for storage of records include, but are not limited to, machine readable data... 19 Customs Duties 2 2012-04-01 2012-04-01 false Methods for storage of records. 163.5 Section 163... THE TREASURY (CONTINUED) RECORDKEEPING § 163.5 Methods for storage of records. (a) Original records...

  15. 19 CFR 163.5 - Methods for storage of records.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... standard business practice for storage of records include, but are not limited to, machine readable data... 19 Customs Duties 2 2011-04-01 2011-04-01 false Methods for storage of records. 163.5 Section 163... THE TREASURY (CONTINUED) RECORDKEEPING § 163.5 Methods for storage of records. (a) Original records...

  16. 78 FR 28111 - Making Open and Machine Readable the New Default for Government Information

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-05-14

    ... warning systems, location-based applications, precision farming tools, and much more, improving Americans... repository of tools and best practices to assist agencies in integrating the Open Data Policy into their... needed to ensure it remains a resource to facilitate the adoption of open data practices. (b) Within 90...

  17. 3 CFR 13642 - Executive Order 13642 of May 9, 2013. Making Open and Machine Readable the New Default for...

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ..., innovation, and scientific discovery that improves Americans' lives and contributes significantly to job... tools, and much more, improving Americans' lives in countless ways and leading to economic growth and... as an asset throughout its life cycle to promote interoperability and openness, and, wherever...

  18. Documentation for the Machine-readable Version of the 0.2-A Resolution Far-ultraviolet Stellar Spectra Measured with COPERNICUS

    NASA Technical Reports Server (NTRS)

    Sheridan, W. T.; Warren, W. H., Jr.

    1981-01-01

    The spectra described represent a subset comprising data for 60 O- and B-type stars. The tape contains data in the spectral region lamda lamda 1000-1450 A with a resolution of 0.2 A. The magnetic tape version of the data is described.

  19. 26 CFR 301.6721-1 - Failure to file correct information returns.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... file timely includes a failure to file in the required manner, for example, on magnetic media or in... they fall below the 250-threshold requirement) or on magnetic media or other machine-readable form. Filers who are required to file information returns on magnetic media and who file such information...

  20. Reference Manual for Machine-Readable Bibliographic Descriptions.

    ERIC Educational Resources Information Center

    Martin, M. D., Comp.

    UNESCO, in cooperation with several other organizations, has produced a manual, the scope and purpose of which has been to define, for most types of scientific and technical literature, a set of data elements which will constitute an adequate bibliographic citation, and to define the representation of these data elements as they should appear in a…

  1. 40 CFR 85.1905 - Alternative report formats.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 19 2012-07-01 2012-07-01 false Alternative report formats. 85.1905... Alternative report formats. (a) Any manufacturer may submit a plan for making either of the reports required by §§ 85.1903 and 85.1904 on computer cards, magnetic tape or other machine readable format. The...

  2. 40 CFR 85.1905 - Alternative report formats.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 18 2011-07-01 2011-07-01 false Alternative report formats. 85.1905... Alternative report formats. (a) Any manufacturer may submit a plan for making either of the reports required by §§ 85.1903 and 85.1904 on computer cards, magnetic tape or other machine readable format. The...

  3. 40 CFR 85.1905 - Alternative report formats.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 19 2014-07-01 2014-07-01 false Alternative report formats. 85.1905... Alternative report formats. (a) Any manufacturer may submit a plan for making either of the reports required by §§ 85.1903 and 85.1904 on computer cards, magnetic tape or other machine readable format. The...

  4. 40 CFR 85.1905 - Alternative report formats.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 19 2013-07-01 2013-07-01 false Alternative report formats. 85.1905... Alternative report formats. (a) Any manufacturer may submit a plan for making either of the reports required by §§ 85.1903 and 85.1904 on computer cards, magnetic tape or other machine readable format. The...

  5. Nanopublications for exposing experimental data in the life-sciences: a Huntington's Disease case study.

    PubMed

    Mina, Eleni; Thompson, Mark; Kaliyaperumal, Rajaram; Zhao, Jun; der Horst, van Eelke; Tatum, Zuotian; Hettne, Kristina M; Schultes, Erik A; Mons, Barend; Roos, Marco

    2015-01-01

    Data from high throughput experiments often produce far more results than can ever appear in the main text or tables of a single research article. In these cases, the majority of new associations are often archived either as supplemental information in an arbitrary format or in publisher-independent databases that can be difficult to find. These data are not only lost from scientific discourse, but are also elusive to automated search, retrieval and processing. Here, we use the nanopublication model to make scientific assertions that were concluded from a workflow analysis of Huntington's Disease data machine-readable, interoperable, and citable. We followed the nanopublication guidelines to semantically model our assertions as well as their provenance metadata and authorship. We demonstrate interoperability by linking nanopublication provenance to the Research Object model. These results indicate that nanopublications can provide an incentive for researchers to expose data that is interoperable and machine-readable for future use and preservation for which they can get credits for their effort. Nanopublications can have a leading role into hypotheses generation offering opportunities to produce large-scale data integration.

  6. Word add-in for ontology recognition: semantic enrichment of scientific literature.

    PubMed

    Fink, J Lynn; Fernicola, Pablo; Chandran, Rahul; Parastatidis, Savas; Wade, Alex; Naim, Oscar; Quinn, Gregory B; Bourne, Philip E

    2010-02-24

    In the current era of scientific research, efficient communication of information is paramount. As such, the nature of scholarly and scientific communication is changing; cyberinfrastructure is now absolutely necessary and new media are allowing information and knowledge to be more interactive and immediate. One approach to making knowledge more accessible is the addition of machine-readable semantic data to scholarly articles. The Word add-in presented here will assist authors in this effort by automatically recognizing and highlighting words or phrases that are likely information-rich, allowing authors to associate semantic data with those words or phrases, and to embed that data in the document as XML. The add-in and source code are publicly available at http://www.codeplex.com/UCSDBioLit. The Word add-in for ontology term recognition makes it possible for an author to add semantic data to a document as it is being written and it encodes these data using XML tags that are effectively a standard in life sciences literature. Allowing authors to mark-up their own work will help increase the amount and quality of machine-readable literature metadata.

  7. Applying Semantic Web technologies to improve the retrieval, credibility and use of health-related web resources.

    PubMed

    Mayer, Miguel A; Karampiperis, Pythagoras; Kukurikos, Antonis; Karkaletsis, Vangelis; Stamatakis, Kostas; Villarroel, Dagmar; Leis, Angela

    2011-06-01

    The number of health-related websites is increasing day-by-day; however, their quality is variable and difficult to assess. Various "trust marks" and filtering portals have been created in order to assist consumers in retrieving quality medical information. Consumers are using search engines as the main tool to get health information; however, the major problem is that the meaning of the web content is not machine-readable in the sense that computers cannot understand words and sentences as humans can. In addition, trust marks are invisible to search engines, thus limiting their usefulness in practice. During the last five years there have been different attempts to use Semantic Web tools to label health-related web resources to help internet users identify trustworthy resources. This paper discusses how Semantic Web technologies can be applied in practice to generate machine-readable labels and display their content, as well as to empower end-users by providing them with the infrastructure for expressing and sharing their opinions on the quality of health-related web resources.

  8. Elements of a next generation time-series ASCII data file format for Earth Sciences

    NASA Astrophysics Data System (ADS)

    Webster, C. J.

    2015-12-01

    Data in ASCII comma separated value (CSV) format are recognized as the most simple, straightforward and readable type of data present in the geosciences. Many scientific workflows developed over the years rely on data using this simple format. However, there is a need for a lightweight ASCII header format standard that is easy to create and easy to work with. Current OGC grade XML standards are complex and difficult to implement for researchers with few resources. Ideally, such a format should provide the data in CSV for easy consumption by generic applications such as spreadsheets. The format should use an existing time standard. The header should be easily human readable as well as machine parsable. The metadata format should be extendable to allow vocabularies to be adopted as they are created by external standards bodies. The creation of such a format will increase the productivity of software engineers and scientists because fewer translators and checkers would be required. Data in ASCII comma separated value (CSV) format are recognized as the most simple, straightforward and readable type of data present in the geosciences. Many scientific workflows developed over the years rely on data using this simple format. However, there is a need for a lightweight ASCII header format standard that is easy to create and easy to work with. Current OGC grade XML standards are complex and difficult to implement for researchers with few resources. Ideally, such a format would provide the data in CSV for easy consumption by generic applications such as spreadsheets. The format would use existing time standard. The header would be easily human readable as well as machine parsable. The metadata format would be extendable to allow vocabularies to be adopted as they are created by external standards bodies. The creation of such a format would increase the productivity of software engineers and scientists because fewer translators would be required.

  9. Can ASCII data files be standardized for Earth Science?

    NASA Astrophysics Data System (ADS)

    Evans, K. D.; Chen, G.; Wilson, A.; Law, E.; Olding, S. W.; Krotkov, N. A.; Conover, H.

    2015-12-01

    NASA's Earth Science Data Systems Working Groups (ESDSWG) was created over 10 years ago. The role of the ESDSWG is to make recommendations relevant to NASA's Earth science data systems from user experiences. Each group works independently focusing on a unique topic. Participation in ESDSWG groups comes from a variety of NASA-funded science and technology projects, such as MEaSUREs, NASA information technology experts, affiliated contractor, staff and other interested community members from academia and industry. Recommendations from the ESDSWG groups will enhance NASA's efforts to develop long term data products. Each year, the ESDSWG has a face-to-face meeting to discuss recommendations and future efforts. Last year's (2014) ASCII for Science Data Working Group (ASCII WG) completed its goals and made recommendations on a minimum set of information that is needed to make ASCII files at least human readable and usable for the foreseeable future. The 2014 ASCII WG created a table of ASCII files and their components as a means for understanding what kind of ASCII formats exist and what components they have in common. Using this table and adding information from other ASCII file formats, we will discuss the advantages and disadvantages of a standardized format. For instance, Space Geodesy scientists have been using the same RINEX/SINEX ASCII format for decades. Astronomers mostly archive their data in the FITS format. Yet Earth scientists seem to have a slew of ASCII formats, such as ICARTT, netCDF (an ASCII dump) and the IceBridge ASCII format. The 2015 Working Group is focusing on promoting extendibility and machine readability of ASCII data. Questions have been posed, including, Can we have a standardized ASCII file format? Can it be machine-readable and simultaneously human-readable? We will present a summary of the current used ASCII formats in terms of advantages and shortcomings, as well as potential improvements.

  10. Efficient and Scalable Cross-Matching of (Very) Large Catalogs

    NASA Astrophysics Data System (ADS)

    Pineau, F.-X.; Boch, T.; Derriere, S.

    2011-07-01

    Whether it be for building multi-wavelength datasets from independent surveys, studying changes in objects luminosities, or detecting moving objects (stellar proper motions, asteroids), cross-catalog matching is a technique widely used in astronomy. The need for efficient, reliable and scalable cross-catalog matching is becoming even more pressing with forthcoming projects which will produce huge catalogs in which astronomers will dig for rare objects, perform statistical analysis and classification, or real-time transients detection. We have developed a formalism and the corresponding technical framework to address the challenge of fast cross-catalog matching. Our formalism supports more than simple nearest-neighbor search, and handles elliptical positional errors. Scalability is improved by partitioning the sky using the HEALPix scheme, and processing independently each sky cell. The use of multi-threaded two-dimensional kd-trees adapted to managing equatorial coordinates enables efficient neighbor search. The whole process can run on a single computer, but could also use clusters of machines to cross-match future very large surveys such as GAIA or LSST in reasonable times. We already achieve performances where the 2MASS (˜470M sources) and SDSS DR7 (˜350M sources) can be matched on a single machine in less than 10 minutes. We aim at providing astronomers with a catalog cross-matching service, available on-line and leveraging on the catalogs present in the VizieR database. This service will allow users both to access pre-computed cross-matches across some very large catalogs, and to run customized cross-matching operations. It will also support VO protocols for synchronous or asynchronous queries.

  11. Construction of a Calibrated Probabilistic Classification Catalog: Application to 50k Variable Sources in the All-Sky Automated Survey

    NASA Astrophysics Data System (ADS)

    Richards, Joseph W.; Starr, Dan L.; Miller, Adam A.; Bloom, Joshua S.; Butler, Nathaniel R.; Brink, Henrik; Crellin-Quick, Arien

    2012-12-01

    With growing data volumes from synoptic surveys, astronomers necessarily must become more abstracted from the discovery and introspection processes. Given the scarcity of follow-up resources, there is a particularly sharp onus on the frameworks that replace these human roles to provide accurate and well-calibrated probabilistic classification catalogs. Such catalogs inform the subsequent follow-up, allowing consumers to optimize the selection of specific sources for further study and permitting rigorous treatment of classification purities and efficiencies for population studies. Here, we describe a process to produce a probabilistic classification catalog of variability with machine learning from a multi-epoch photometric survey. In addition to producing accurate classifications, we show how to estimate calibrated class probabilities and motivate the importance of probability calibration. We also introduce a methodology for feature-based anomaly detection, which allows discovery of objects in the survey that do not fit within the predefined class taxonomy. Finally, we apply these methods to sources observed by the All-Sky Automated Survey (ASAS), and release the Machine-learned ASAS Classification Catalog (MACC), a 28 class probabilistic classification catalog of 50,124 ASAS sources in the ASAS Catalog of Variable Stars. We estimate that MACC achieves a sub-20% classification error rate and demonstrate that the class posterior probabilities are reasonably calibrated. MACC classifications compare favorably to the classifications of several previous domain-specific ASAS papers and to the ASAS Catalog of Variable Stars, which had classified only 24% of those sources into one of 12 science classes.

  12. CONSTRUCTION OF A CALIBRATED PROBABILISTIC CLASSIFICATION CATALOG: APPLICATION TO 50k VARIABLE SOURCES IN THE ALL-SKY AUTOMATED SURVEY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Richards, Joseph W.; Starr, Dan L.; Miller, Adam A.

    2012-12-15

    With growing data volumes from synoptic surveys, astronomers necessarily must become more abstracted from the discovery and introspection processes. Given the scarcity of follow-up resources, there is a particularly sharp onus on the frameworks that replace these human roles to provide accurate and well-calibrated probabilistic classification catalogs. Such catalogs inform the subsequent follow-up, allowing consumers to optimize the selection of specific sources for further study and permitting rigorous treatment of classification purities and efficiencies for population studies. Here, we describe a process to produce a probabilistic classification catalog of variability with machine learning from a multi-epoch photometric survey. In additionmore » to producing accurate classifications, we show how to estimate calibrated class probabilities and motivate the importance of probability calibration. We also introduce a methodology for feature-based anomaly detection, which allows discovery of objects in the survey that do not fit within the predefined class taxonomy. Finally, we apply these methods to sources observed by the All-Sky Automated Survey (ASAS), and release the Machine-learned ASAS Classification Catalog (MACC), a 28 class probabilistic classification catalog of 50,124 ASAS sources in the ASAS Catalog of Variable Stars. We estimate that MACC achieves a sub-20% classification error rate and demonstrate that the class posterior probabilities are reasonably calibrated. MACC classifications compare favorably to the classifications of several previous domain-specific ASAS papers and to the ASAS Catalog of Variable Stars, which had classified only 24% of those sources into one of 12 science classes.« less

  13. Inferring Planet Occurrence Rates With a Q1-Q16 Kepler Planet Candidate Catalog Produced by a Machine Learning Classifier

    NASA Astrophysics Data System (ADS)

    Catanzarite, Joseph; Jenkins, Jon Michael; Burke, Christopher J.; McCauliff, Sean D.; Kepler Science Operations Center

    2015-01-01

    NASA's Kepler Space Telescope monitored the photometric variations of over 170,000 stars within a ~100 square degree field in the constellation Cygnus, at half-hour cadence, over its four year prime mission. The Kepler SOC (Science Operations Center) pipeline calibrates the pixels of the target apertures for each star, corrects light curves for systematic error, and detects TCEs (threshold-crossing events) that may be due to transiting planets. Finally the pipeline estimates planet parameters for all TCEs and computes quantitative diagnostics that are used by the TCERT (Threshold Crossing Event Review Team) to produce a catalog containing KOIs (Kepler Objects of Interest). KOIs are TCEs that are determined to be either likely transiting planets or astrophysical false positives such as background eclipsing binary stars. Using examples from the Q1-Q16 TCERT KOI catalog as a training set, we created a machine-learning classifier that dispositions the TCEs into categories of PC (planet candidate), AFP (astrophysical false positive) and NTP (non-transiting phenomenon). The classifier uniformly and consistently applies heuristics developed by TCERT as well as other diagnostics to the Q1-Q16 TCEs to produce a more robust and reliable catalog of planet candidates than is possible with only human classification. In this work, we estimate planet occurrence rates, based on the machine-learning-produced catalog of Kepler planet candidates. Kepler was selected as the 10th mission of the Discovery Program. Funding for this mission is provided by NASA, Science Mission Directorate.

  14. Library Dream Machines: Helping Students Master Super Online Catalogs.

    ERIC Educational Resources Information Center

    Webb, T. D.

    1992-01-01

    Describes how automation has transformed the library and how super-catalogs have affected the process of doing research. Explains how faculty and librarians can work together to help students to use the available databases effectively, by teaching them Boolean logic, standard record formats, filing rules, etc. (DMM)

  15. The semantics of Chemical Markup Language (CML): dictionaries and conventions.

    PubMed

    Murray-Rust, Peter; Townsend, Joe A; Adams, Sam E; Phadungsukanan, Weerapong; Thomas, Jens

    2011-10-14

    The semantic architecture of CML consists of conventions, dictionaries and units. The conventions conform to a top-level specification and each convention can constrain compliant documents through machine-processing (validation). Dictionaries conform to a dictionary specification which also imposes machine validation on the dictionaries. Each dictionary can also be used to validate data in a CML document, and provide human-readable descriptions. An additional set of conventions and dictionaries are used to support scientific units. All conventions, dictionaries and dictionary elements are identifiable and addressable through unique URIs.

  16. The semantics of Chemical Markup Language (CML): dictionaries and conventions

    PubMed Central

    2011-01-01

    The semantic architecture of CML consists of conventions, dictionaries and units. The conventions conform to a top-level specification and each convention can constrain compliant documents through machine-processing (validation). Dictionaries conform to a dictionary specification which also imposes machine validation on the dictionaries. Each dictionary can also be used to validate data in a CML document, and provide human-readable descriptions. An additional set of conventions and dictionaries are used to support scientific units. All conventions, dictionaries and dictionary elements are identifiable and addressable through unique URIs. PMID:21999509

  17. PREFACE: Anti-counterfeit Image Analysis Methods (A Special Session of ICSXII)

    NASA Astrophysics Data System (ADS)

    Javidi, B.; Fournel, T.

    2007-06-01

    The International Congress for Stereology is dedicated to theoretical and applied aspects of stochastic tools, image analysis and mathematical morphology. A special emphasis on `anti-counterfeit image analysis methods' has been given this year for the XIIth edition (ICSXII). Facing the economic and social threat of counterfeiting, this devoted session presents recent advances and original solutions in the field. A first group of methods are related to marks located either on the product (physical marks) or on the data (hidden information) to be protected. These methods concern laser fs 3D encoding and source separation for machine-readable identification, moiré and `guilloche' engraving for visual verification and watermarking. Machine-readable travel documents are well-suited examples introducing the second group of methods which are related to cryptography. Used in passports for data authentication and identification (of people), cryptography provides some powerful tools. Opto-digital processing allows some efficient implementations described in the papers and promising applications. We would like to thank the reviewers who have contributed to a session of high quality, and the authors for their fine and hard work. We would like to address some special thanks to the invited lecturers, namely Professor Roger Hersch and Dr Isaac Amidror for their survey of moiré methods, Prof. Serge Vaudenay for his survey of existing protocols concerning machine-readable travel documents, and Dr Elisabet Pérez-Cabré for her presentation on optical encryption for multifactor authentication. We also thank Professor Dominique Jeulin, President of the International Society for Stereology, Professor Michel Jourlin, President of the organizing committee of ICSXII, for their help and advice, and Mr Graham Douglas, the Publisher of Journal of Physics: Conference Series at IOP Publishing, for his efficiency. We hope that this collection of papers will be useful as a tool to further develop a very important field. Bahram Javidi University of Connecticut (USA) Thierry Fournel University of Saint-Etienne (France) Chairs of the special session on `Anti-counterfeit image analysis methods', July 2007

  18. Annotation of rule-based models with formal semantics to enable creation, analysis, reuse and visualization.

    PubMed

    Misirli, Goksel; Cavaliere, Matteo; Waites, William; Pocock, Matthew; Madsen, Curtis; Gilfellon, Owen; Honorato-Zimmer, Ricardo; Zuliani, Paolo; Danos, Vincent; Wipat, Anil

    2016-03-15

    Biological systems are complex and challenging to model and therefore model reuse is highly desirable. To promote model reuse, models should include both information about the specifics of simulations and the underlying biology in the form of metadata. The availability of computationally tractable metadata is especially important for the effective automated interpretation and processing of models. Metadata are typically represented as machine-readable annotations which enhance programmatic access to information about models. Rule-based languages have emerged as a modelling framework to represent the complexity of biological systems. Annotation approaches have been widely used for reaction-based formalisms such as SBML. However, rule-based languages still lack a rich annotation framework to add semantic information, such as machine-readable descriptions, to the components of a model. We present an annotation framework and guidelines for annotating rule-based models, encoded in the commonly used Kappa and BioNetGen languages. We adapt widely adopted annotation approaches to rule-based models. We initially propose a syntax to store machine-readable annotations and describe a mapping between rule-based modelling entities, such as agents and rules, and their annotations. We then describe an ontology to both annotate these models and capture the information contained therein, and demonstrate annotating these models using examples. Finally, we present a proof of concept tool for extracting annotations from a model that can be queried and analyzed in a uniform way. The uniform representation of the annotations can be used to facilitate the creation, analysis, reuse and visualization of rule-based models. Although examples are given, using specific implementations the proposed techniques can be applied to rule-based models in general. The annotation ontology for rule-based models can be found at http://purl.org/rbm/rbmo The krdf tool and associated executable examples are available at http://purl.org/rbm/rbmo/krdf anil.wipat@newcastle.ac.uk or vdanos@inf.ed.ac.uk. © The Author 2015. Published by Oxford University Press.

  19. Development of Clinical Contents Model Markup Language for Electronic Health Records

    PubMed Central

    Yun, Ji-Hyun; Kim, Yoon

    2012-01-01

    Objectives To develop dedicated markup language for clinical contents models (CCM) to facilitate the active use of CCM in electronic health record systems. Methods Based on analysis of the structure and characteristics of CCM in the clinical domain, we designed extensible markup language (XML) based CCM markup language (CCML) schema manually. Results CCML faithfully reflects CCM in both the syntactic and semantic aspects. As this language is based on XML, it can be expressed and processed in computer systems and can be used in a technology-neutral way. Conclusions CCML has the following strengths: it is machine-readable and highly human-readable, it does not require a dedicated parser, and it can be applied for existing electronic health record systems. PMID:23115739

  20. High-capacity high-speed recording

    NASA Astrophysics Data System (ADS)

    Jamberdino, A. A.

    1981-06-01

    Continuing advances in wideband communications and information handling are leading to extremely large volume digital data systems for which conventional data storage techniques are becoming inadequate. The paper presents an assessment of alternative recording technologies for the extremely wideband, high capacity storage and retrieval systems currently under development. Attention is given to longitudinal and rotary head high density magnetic recording, laser holography in human readable/machine readable devices and a wideband recorder, digital optical disks, and spot recording in microfiche formats. The electro-optical technologies considered are noted to be capable of providing data bandwidths up to 1000 megabits/sec and total data storage capacities in the 10 to the 11th to 10 to the 12th bit range, an order of magnitude improvement over conventional technologies.

  1. An Evaluation Of Holograms In Training And As Job Performance Aids

    NASA Astrophysics Data System (ADS)

    Frey, Allan H.

    1986-08-01

    Experimentation was carried out to evaluate holograms for use in training and as job aids. Holograms were compared against line drawings and photographs as methods of presenting visual information needed to accomplish a number of tasks. The dependent variables were assembly speed and assembly errors with people unstressed, assembly speed and assembly errors with people stressed, the percentage of discovered errors in assemblies, the number of correct assemblies misidentified as erroneous, and information extraction. Holograms generally were as good as or better visual aids than either photographs or line drawings. The use of holograms tends to reduce errors rather than speed assembly time in the assembly tasks used in these experiments. They also enhance the discovery of errors when the subject is attempting to locate assembly errors in a construction. The results of this experimentation suggest that serious consideration should be given to the use of holography in the development of job aids and in training. Besides these advantages for job aids, other advantages we found are that when page formated information is stored in man-readable holograms they are still useable when scratched or damaged even when similarly damaged microfilm is unuseable. Holography can also be used to store man and machine readable data simultaneously. Such storage would provide simplified backup in the event of machine failure, and it would permit the development of compatible machine and manual systems for job aid applications.

  2. Simplifying the Reuse and Interoperability of Geoscience Data Sets and Models with Semantic Metadata that is Human-Readable and Machine-actionable

    NASA Astrophysics Data System (ADS)

    Peckham, S. D.

    2017-12-01

    Standardized, deep descriptions of digital resources (e.g. data sets, computational models, software tools and publications) make it possible to develop user-friendly software systems that assist scientists with the discovery and appropriate use of these resources. Semantic metadata makes it possible for machines to take actions on behalf of humans, such as automatically identifying the resources needed to solve a given problem, retrieving them and then automatically connecting them (despite their heterogeneity) into a functioning workflow. Standardized model metadata also helps model users to understand the important details that underpin computational models and to compare the capabilities of different models. These details include simplifying assumptions on the physics, governing equations and the numerical methods used to solve them, discretization of space (the grid) and time (the time-stepping scheme), state variables (input or output), model configuration parameters. This kind of metadata provides a "deep description" of a computational model that goes well beyond other types of metadata (e.g. author, purpose, scientific domain, programming language, digital rights, provenance, execution) and captures the science that underpins a model. A carefully constructed, unambiguous and rules-based schema to address this problem, called the Geoscience Standard Names ontology will be presented that utilizes Semantic Web best practices and technologies. It has also been designed to work across science domains and to be readable by both humans and machines.

  3. Educational and Commercial Utilization of a Chemical Information Center, Four Year Summary.

    ERIC Educational Resources Information Center

    Williams, Martha E.; And Others

    The major objective of the IITRI Computer Search Center is to educate and link industry, academia, and government institutions to chemical and other scientific information systems and sources. The Center was developed to meet this objective and is in full operation providing services to users from a variety of machine-readable data bases with…

  4. 26 CFR 1.197-2 - Amortization of goodwill and certain other intangibles.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ..., process, design, pattern, know-how, format, package design, computer software (as defined in paragraph (c... section 1253(b)(1) and includes any agreement that provides one of the parties to the agreement with the... any program or routine (that is, any sequence of machine-readable code) that is designed to cause a...

  5. Intersystem Compatibility and Convertibility of Subject Vocabularies.

    ERIC Educational Resources Information Center

    Wall, E.; Barnes, J.

    This is the fifth in a series of eight reports of a research study for the National Agricultural Library (NAL) on the effective utilization of bibliographic data bases in machine readable form. NAL desires ultimately to develop techniques of interacting with other data bases so that queries put to NAL may be answered with documents or document…

  6. Four-Year Summary, Educational and Commercial Utilization of a Chemical Information Center. Part I.

    ERIC Educational Resources Information Center

    Schipma, Peter B., Ed.

    The major objective of the Illinois Institute of Technology (IIT) Computer Search Center (CSC) is to educate and link industry, academia, and government institutions to chemical and other scientific information systems and sources. The CSC is in full operation providing services to users from a variety of machine-readable data bases with minimal…

  7. An Evaluation of Implementing Koha in a Chinese Language Environment

    ERIC Educational Resources Information Center

    Chang, Naicheng; Tsai, Yuchin; Hopkinson, Alan

    2010-01-01

    Purpose: The purpose of this paper is to evaluate issues of different scripts in the same record (in MARC21 and Chinese machine-readable cataloguing (CMARC)) and Chinese internal codes (i.e. double-byte character set) when implementing Koha. It also discusses successful efforts in promoting the adoption of Koha in Taiwan, particularly the…

  8. Adolescent Fertility: National File [Machine-Readable Data File].

    ERIC Educational Resources Information Center

    Moore, Kristin A.; And Others

    This computer file contains recent cross sectional data on adolescent fertility in the United States for 1960, 1965, 1970, 1975 and 1980-85. The following variables are included: (1) births; (2) birth rates; (3) abortions; (4) non-marital childbearing; (5) infant mortality; and (6) low birth weight. Data for both teenagers and women aged 20-24 are…

  9. Adolescent Fertility: State File [Machine-Readable Data File].

    ERIC Educational Resources Information Center

    Moore, Kristin A.; And Others

    This computer file contains recent cross sectional data on adolescent fertility by state for 1960, 1965, 1970, 1975 and 1980-85. The following variables are included: (1) births; (2) birth rates; (3) abortions; (4) non-marital childbearing; (5) infant mortality; and (6) low birth weight. Data for both teenagers and women aged 20-24 years are…

  10. Experiences Using an Open Source Software Library to Teach Computer Vision Subjects

    ERIC Educational Resources Information Center

    Cazorla, Miguel; Viejo, Diego

    2015-01-01

    Machine vision is an important subject in computer science and engineering degrees. For laboratory experimentation, it is desirable to have a complete and easy-to-use tool. In this work we present a Java library, oriented to teaching computer vision. We have designed and built the library from the scratch with emphasis on readability and…

  11. Supporting Open Access to European Academic Courses: The ASK-CDM-ECTS Tool

    ERIC Educational Resources Information Center

    Sampson, Demetrios G.; Zervas, Panagiotis

    2013-01-01

    Purpose: This paper aims to present and evaluate a web-based tool, namely ASK-CDM-ECTS, which facilitates authoring and publishing on the web descriptions of (open) academic courses in machine-readable format using an application profile of the Course Description Metadata (CDM) specification, namely CDM-ECTS. Design/methodology/approach: The paper…

  12. High School and Beyond. 1980 Sophomore Cohort. First Follow-Up (1982). [machine-readable data file].

    ERIC Educational Resources Information Center

    National Center for Education Statistics (ED), Washington, DC.

    The High School and Beyond 1980 Sophomore Cohort First Follow-Up (1982) data file is presented. The First Follow-Up Sophomore Cohort data tape consists of four related data files: (1) the student data file (including data availability flags, weights, questionnaire data, and composite variables); (2) Statistical Analysis System (SAS) control cards…

  13. Public Data Set: Continuous, Edge Localized Ion Heating During Non-Solenoidal Plasma Startup and Sustainment in a Low Aspect Ratio Tokamak

    DOE Data Explorer

    Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Barr, Jayson L. [University of Wisconsin-Madison] (ORCID:0000000177685931); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Hinson, Edward T. [University of Wisconsin-Madison] (ORCID:000000019713140X); Perry, Justin M. [University of Wisconsin-Madison] (ORCID:0000000171228609); Reusch, Joshua A. [University of Wisconsin-Madison] (ORCID:0000000284249422); Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448)

    2017-05-16

    This public data set contains openly-documented, machine readable digital research data corresponding to figures published in M.G. Burke et. al., 'Continuous, Edge Localized Ion Heating During Non-Solenoidal Plasma Startup and Sustainment in a Low Aspect Ratio Tokamak,' Nucl. Fusion 57, 076010 (2017).

  14. Handling of Varied Data Bases in an Information Center Environment.

    ERIC Educational Resources Information Center

    Williams, Martha E.

    Information centers exist to provide information from machine-readable data bases to users in industry, universities and other organizations. The computer Search Center of the IIT Research Institute was designed with a number of variables and uncertainties before it. In this paper, the author discusses how the Center was designed to enable it to…

  15. Merged Federal Files [Academic Year] 1978-79 [machine-readable data file].

    ERIC Educational Resources Information Center

    National Center for Education Statistics (ED), Washington, DC.

    The Merged Federal File for 1978-79 contains school district level data from the following six source files: (1) the Census of Governments' Survey of Local Government Finances--School Systems (F-33) (with 16,343 records merged); (2) the National Center for Education Statistics Survey of School Systems (School District Universe) (with 16,743…

  16. NaturalReader: A New Generation Text Reader

    ERIC Educational Resources Information Center

    Flood, Jacqueline

    2007-01-01

    NaturalReader (http://www.naturalreaders.com/) is a new generation text reader, which means that it reads any machine readable text using synthesized speech without having to copy and paste the selected text into the NaturalReader application window. It installs a toolbar directly into all of the Microsoft Office[TM] programs and uses a mini-board…

  17. VizieR Online Data Catalog: Positions of 502 Stars in Pleiades Region (Eichhorn+ 1970)

    NASA Astrophysics Data System (ADS)

    Eichhorn, H.; Googe, W. D.; Lukac, C. F.; Murphy, J. K.

    1996-01-01

    The catalog contains the positions (equinox B1900.0 and epoch B1955.0) of 502 stars in a region of about 1.5 degrees square in the Pleiades cluster, centered on Eta Tau. These coordinates have been derived from measurements of stellar images obtained with 65 exposures of various durations on 14 photographic plates with two telescopes at McCormick Observatory and Van Vleck Observatory. The plates were reduced by the plate overlap method, which resulted in a high degree of systematic accuracy in the final positions. Data in the machine version include Hertzsprung number, color index, photovisual magnitude, right ascension and declination and their standard errors, proper motion, and differences between the present position and previous works. Data for exposures, plates, and images measured, present in the published catalog, are not included in the machine version. (1 data file).

  18. VizieR Online Data Catalog: 231 AGN candidates from the 2FGL catalog (Doert+, 2014)

    NASA Astrophysics Data System (ADS)

    Doert, M.; Errando, M.

    2016-01-01

    The second Fermi-LAT source catalog (2FGL; Nolan et al. 2012, cat. J/ApJS/199/31) is the deepest all-sky survey available in the gamma-ray band. It contains 1873 sources, of which 576 remain unassociated. The Large Area Telescope (LAT) on board the Fermi Gamma-ray Space Telescope started operations in 2008. In this work, machine-learning algorithms are used to identify unassociated sources in the 2FGL catalog with properties similar to gamma-ray-emitting Active Galactic Nuclei (AGN). This analysis finds 231 high-confidence AGN candidates (see Table3). (1 data file).

  19. National Collegiate Software Clearinghouse Software for the Humanities and Social Sciences. Summer 1989 Catalog.

    ERIC Educational Resources Information Center

    National Collegiate Software Clearinghouse, Durham, NC.

    Over 250 microcomputer software packages, intended for use on MS-DOS machines by scholars and teachers in the humanities and social sciences, are included in this catalog. The clearinghouse's first Macintosh listing is included, with many more Macintosh programs and data sets being planned and tested for future inclusion. Most programs were…

  20. U.S. Government Films, 1969. A Catalog of Motion Pictures and Filmstrips for Sale by the National Audiovisual Center.

    ERIC Educational Resources Information Center

    National Archives and Records Service (GSA), Washington, DC. National Audiovisual Center.

    Approximately 3,000 films and filmstrips which document the functions and operations of Federal agencies are referenced in this annotated sales catalog. Each entry is listed according to one of 20 areas: agriculture, automotive, aviation, business, education and culture, electricity, electronics, health and medical, human relations, machining,…

  1. User-Based Information Retrieval System Interface Evaluation: An Examination of an On-Line Public Access Catalog.

    ERIC Educational Resources Information Center

    Hert, Carol A.; Nilan, Michael S.

    1991-01-01

    Presents preliminary data that characterizes the relationship between what users say they are trying to accomplish when using an online public access catalog (OPAC) and their perceptions of what input to give the system. Human-machine interaction is discussed, and appropriate methods for evaluating information retrieval systems are considered. (18…

  2. Requirements-Based Conformance Testing of ARINC 653 Real-Time Operating Systems

    NASA Astrophysics Data System (ADS)

    Maksimov, Andrey

    2010-08-01

    Requirements-based testing is emphasized in avionics certification documents because this strategy has been found to be the most effective at revealing errors. This paper describes the unified requirements-based approach to the creation of conformance test suites for mission-critical systems. The approach uses formal machine-readable specifications of requirements and finite state machine model for test sequences generation on-the-fly. The paper also presents the test system for automated test generation for ARINC 653 services built on this approach. Possible application of the presented approach to various areas of avionics embedded systems testing is discussed.

  3. Light at Night Markup Language (LANML): XML Technology for Light at Night Monitoring Data

    NASA Astrophysics Data System (ADS)

    Craine, B. L.; Craine, E. R.; Craine, E. M.; Crawford, D. L.

    2013-05-01

    Light at Night Markup Language (LANML) is a standard, based upon XML, useful in acquiring, validating, transporting, archiving and analyzing multi-dimensional light at night (LAN) datasets of any size. The LANML standard can accommodate a variety of measurement scenarios including single spot measures, static time-series, web based monitoring networks, mobile measurements, and airborne measurements. LANML is human-readable, machine-readable, and does not require a dedicated parser. In addition LANML is flexible; ensuring future extensions of the format will remain backward compatible with analysis software. The XML technology is at the heart of communicating over the internet and can be equally useful at the desktop level, making this standard particularly attractive for web based applications, educational outreach and efficient collaboration between research groups.

  4. Actionable, long-term stable and semantic web compatible identifiers for access to biological collection objects

    PubMed Central

    Hyam, Roger; Hagedorn, Gregor; Chagnoux, Simon; Röpert, Dominik; Casino, Ana; Droege, Gabi; Glöckler, Falko; Gödderz, Karsten; Groom, Quentin; Hoffmann, Jana; Holleman, Ayco; Kempa, Matúš; Koivula, Hanna; Marhold, Karol; Nicolson, Nicky; Smith, Vincent S.; Triebel, Dagmar

    2017-01-01

    With biodiversity research activities being increasingly shifted to the web, the need for a system of persistent and stable identifiers for physical collection objects becomes increasingly pressing. The Consortium of European Taxonomic Facilities agreed on a common system of HTTP-URI-based stable identifiers which is now rolled out to its member organizations. The system follows Linked Open Data principles and implements redirection mechanisms to human-readable and machine-readable representations of specimens facilitating seamless integration into the growing semantic web. The implementation of stable identifiers across collection organizations is supported with open source provider software scripts, best practices documentations and recommendations for RDF metadata elements facilitating harmonized access to collection information in web portals. Database URL: http://cetaf.org/cetaf-stable-identifiers PMID:28365724

  5. EduCard. Adult Education Access Card. Policy Option Paper on Strategic Recommendation 4. First Edition.

    ERIC Educational Resources Information Center

    Porter, Dennis

    One recommendation of the 1989 California Strategic Plan for Adult Education is the use of EduCard. EduCard, the Adult Education Access Card, is a means of giving learners access to information about educational opportunities and providing administrators with machine-readable information on learners' prior education and traiing. Three models are:…

  6. The Application of Clustering Techniques to Citation Data. Research Reports Series B No. 6.

    ERIC Educational Resources Information Center

    Arms, William Y.; Arms, Caroline

    This report describes research carried out as part of the Design of Information Systems in the Social Sciences (DISISS) project. Cluster analysis techniques were applied to a machine readable file of bibliographic data in the form of cited journal titles in order to identify groupings which could be used to structure bibliographic files. Practical…

  7. Demographic Profile of U.S. Children: National File [Machine-Readable Data File].

    ERIC Educational Resources Information Center

    Peterson, J. L.; White, R. N.

    These two computer files contain social and demographic data about U.S. children and their families taken from the March 1985 Current Population Survey of the U.S. Census. One file is for all children; the second file is for black children. The following column variables are included: (1) family structure; (2) parent educational attainment; (3)…

  8. The Nation's Memory: The United States National Archives and Records Administration. An Interview with Don W. Wilson, Archivist of the United States, National Archives and Records Administration.

    ERIC Educational Resources Information Center

    Brodhead, Michael J.; Zink, Steven D.

    1993-01-01

    Discusses the National Archives and Records Administration (NARA) through an interview with the Archivist of the United States, Don Wilson. Topics addressed include archival independence and congressional relations; national information policy; expansion plans; machine-readable archival records; preservation activities; and relations with other…

  9. The Role of Mechanized Services in the Provision of Information with Special Reference to the University Environment.

    ERIC Educational Resources Information Center

    Heim, Kathleen M.

    The use, history, and role of machine-readable data base technology is discussed. First the development of data base technology is traced from its beginnings as a special resource for science and technology to its broader use in universities, with descriptions of some specific services. Next the current status of mechanized information services in…

  10. High School and Beyond. 1980 Senior Cohort. First Follow-Up (1982). [machine-readable data file].

    ERIC Educational Resources Information Center

    National Center for Education Statistics (ED), Washington, DC.

    The High School and Beyond 1980 Senior Cohort First Follow-Up (1982) Data File is presented. The First Follow-Up Senior Cohort data tape consists of four related data files: (1) the student data file (including data availability flags, weights, questionnaire data, and composite variables); (2) Statistical Analysis System (SAS) control cards for…

  11. CAMPUS-MINNESOTA User Information Manual. Project PRIME Report, Number 12.

    ERIC Educational Resources Information Center

    Andrew, Gary M.

    The purpose of this report is to aid the use of the computer simulation model, CAMPUS-M, in 4 specific areas: (1) the conceptual modeling of the institution; (2) the preparation of machine readable input data; (3) the preparation of simulation and report commands for the model; and (4) the actual running of the program on a CDC 6600 computer.…

  12. Second International Mathematics Study; Longitudinal, Classroom Process Surveys for Population A: Students, Teachers, and Schools, 1981-1982 [machine-readable data file].

    ERIC Educational Resources Information Center

    Wolfe, Richard G.

    The Second International Mathematics Study (SIMS) of the International Association for the Evaluation of Educational Achievement (IEA) was conducted in 20 countries on two sampled populations: Population A of 13-year-olds and Population B of students studying mathematics in their final year of secondary school. Mathematics achievement was measured…

  13. A possible extension to the RInChI as a means of providing machine readable process data.

    PubMed

    Jacob, Philipp-Maximilian; Lan, Tian; Goodman, Jonathan M; Lapkin, Alexei A

    2017-04-11

    The algorithmic, large-scale use and analysis of reaction databases such as Reaxys is currently hindered by the absence of widely adopted standards for publishing reaction data in machine readable formats. Crucial data such as yields of all products or stoichiometry are frequently not explicitly stated in the published papers and, hence, not reported in the database entry for those reactions, limiting their usefulness for algorithmic analysis. This paper presents a possible extension to the IUPAC RInChI standard via an auxiliary layer, termed ProcAuxInfo, which is a standardised, extensible form in which to report certain key reaction parameters such as declaration of all products and reactants as well as auxiliaries known in the reaction, reaction stoichiometry, amounts of substances used, conversion, yield and operating conditions. The standard is demonstrated via creation of the RInChI including the ProcAuxInfo layer based on three published reactions and demonstrates accurate data recoverability via reverse translation of the created strings. Implementation of this or another method of reporting process data by the publishing community would ensure that databases, such as Reaxys, would be able to abstract crucial data for big data analysis of their contents.

  14. Virtual Machine Language

    NASA Technical Reports Server (NTRS)

    Grasso, Christopher; Page, Dennis; O'Reilly, Taifun; Fteichert, Ralph; Lock, Patricia; Lin, Imin; Naviaux, Keith; Sisino, John

    2005-01-01

    Virtual Machine Language (VML) is a mission-independent, reusable software system for programming for spacecraft operations. Features of VML include a rich set of data types, named functions, parameters, IF and WHILE control structures, polymorphism, and on-the-fly creation of spacecraft commands from calculated values. Spacecraft functions can be abstracted into named blocks that reside in files aboard the spacecraft. These named blocks accept parameters and execute in a repeatable fashion. The sizes of uplink products are minimized by the ability to call blocks that implement most of the command steps. This block approach also enables some autonomous operations aboard the spacecraft, such as aerobraking, telemetry conditional monitoring, and anomaly response, without developing autonomous flight software. Operators on the ground write blocks and command sequences in a concise, high-level, human-readable programming language (also called VML ). A compiler translates the human-readable blocks and command sequences into binary files (the operations products). The flight portion of VML interprets the uplinked binary files. The ground subsystem of VML also includes an interactive sequence- execution tool hosted on workstations, which runs sequences at several thousand times real-time speed, affords debugging, and generates reports. This tool enables iterative development of blocks and sequences within times of the order of seconds.

  15. Word add-in for ontology recognition: semantic enrichment of scientific literature

    PubMed Central

    2010-01-01

    Background In the current era of scientific research, efficient communication of information is paramount. As such, the nature of scholarly and scientific communication is changing; cyberinfrastructure is now absolutely necessary and new media are allowing information and knowledge to be more interactive and immediate. One approach to making knowledge more accessible is the addition of machine-readable semantic data to scholarly articles. Results The Word add-in presented here will assist authors in this effort by automatically recognizing and highlighting words or phrases that are likely information-rich, allowing authors to associate semantic data with those words or phrases, and to embed that data in the document as XML. The add-in and source code are publicly available at http://www.codeplex.com/UCSDBioLit. Conclusions The Word add-in for ontology term recognition makes it possible for an author to add semantic data to a document as it is being written and it encodes these data using XML tags that are effectively a standard in life sciences literature. Allowing authors to mark-up their own work will help increase the amount and quality of machine-readable literature metadata. PMID:20181245

  16. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hajian, Amir; Alvarez, Marcelo A.; Bond, J. Richard, E-mail: ahajian@cita.utoronto.ca, E-mail: malvarez@cita.utoronto.ca, E-mail: bond@cita.utoronto.ca

    Making mock simulated catalogs is an important component of astrophysical data analysis. Selection criteria for observed astronomical objects are often too complicated to be derived from first principles. However the existence of an observed group of objects is a well-suited problem for machine learning classification. In this paper we use one-class classifiers to learn the properties of an observed catalog of clusters of galaxies from ROSAT and to pick clusters from mock simulations that resemble the observed ROSAT catalog. We show how this method can be used to study the cross-correlations of thermal Sunya'ev-Zeldovich signals with number density maps ofmore » X-ray selected cluster catalogs. The method reduces the bias due to hand-tuning the selection function and is readily scalable to large catalogs with a high-dimensional space of astrophysical features.« less

  17. Terminal Ailments Need Not Be Fatal: A Speculative Assessment of the Impact of Online Public Access Catalogs in Academic Settings.

    ERIC Educational Resources Information Center

    Sandler, Mark

    1985-01-01

    Discusses several concerns about nature of online public access catalogs (OPAC) that have particular import to reference librarians: user passivity and loss of control growing out of "human-machine interface" and the larger social context; and the tendency of computerized bibliographic systems to obfuscate human origins of library…

  18. Automatic Inference of Cryptographic Key Length Based on Analysis of Proof Tightness

    DTIC Science & Technology

    2016-06-01

    within an attack tree structure, then expand attack tree methodology to include cryptographic reductions. We then provide the algorithms for...maintaining and automatically reasoning about these expanded attack trees . We provide a software tool that utilizes machine-readable proof and attack metadata...and the attack tree methodology to provide rapid and precise answers regarding security parameters and effective security. This eliminates the need

  19. Demographic Profile of U.S. Children: States in 1980/1, 1985/6 [Machine-Readable Data File].

    ERIC Educational Resources Information Center

    Peterson, J. L.

    These six computer files contain social and demographic data about children and their families in the following states: (1) California; (2) Florida; (3) Illinois; (4) New York; (5) Pennsylvania; and (6) Texas. Data for 1980/81 and 1985/86 are reported. Data will eventually be provided for the 11 largest states. One file is for all children; the…

  20. About machine-readable travel documents

    NASA Astrophysics Data System (ADS)

    Vaudenay, S.; Vuagnoux, M.

    2007-07-01

    Passports are documents that help immigration officers to identify people. In order to strongly authenticate their data and to automatically identify people, they are now equipped with RFID chips. These contain private information, biometrics, and a digital signature by issuing authorities. Although they substantially increase security at the border controls, they also come with new security and privacy issues. In this paper, we survey existing protocols and their weaknesses.

  1. Bibliography On Multiprocessors And Distributed Processing

    NASA Technical Reports Server (NTRS)

    Miya, Eugene N.

    1988-01-01

    Multiprocessor and Distributed Processing Bibliography package consists of large machine-readable bibliographic data base, which in addition to usual keyword searches, used for producing citations, indexes, and cross-references. Data base contains UNIX(R) "refer" -formatted ASCII data and implemented on any computer running under UNIX(R) operating system. Easily convertible to other operating systems. Requires approximately one megabyte of secondary storage. Bibliography compiled in 1985.

  2. Investigation of the Public Library as a Linking Agent to Major Scientific, Educational, Social and Environmental Data Bases. Two-Year Interim Report.

    ERIC Educational Resources Information Center

    Summit, Roger K.; Firschein, Oscar

    Eight public libraries participated in a two-year experiment to investigate the potential of the public library as a "linking agent" between the public and the many machine-readable data bases currently accessible using on line computer terminals. The investigation covered users of the service, impact on the library, conditions for…

  3. Investigation of the Public Library as a Linking Agent to Major Scientific, Educational, Social, and Environmental Data Bases. Final Report.

    ERIC Educational Resources Information Center

    Lockheed Research Lab., Palo Alto, CA.

    The DIALIB Project was a 3-year experiment that investigated the potential of the public library as a "linking agent" between the public and the many machine-readable data bases currently accessible via the telephone using online terminals. The study investigated the following questions: (1) Is online search of use to the patrons of a…

  4. The Automatic Measuring Machines and Ground-Based Astrometry

    NASA Astrophysics Data System (ADS)

    Sergeeva, T. P.

    The introduction of the automatic measuring machines into the astronomical investigations a little more then a quarter of the century ago has increased essentially the range and the scale of projects which the astronomers could capable to realize since then. During that time, there have been dozens photographic sky surveys, which have covered all of the sky more then once. Due to high accuracy and speed of automatic measuring machines the photographic astrometry has obtained the opportunity to create the high precision catalogs such as CpC2. Investigations of the structure and kinematics of the stellar components of our Galaxy has been revolutionized in the last decade by the advent of automated plate measuring machines. But in an age of rapidly evolving electronic detectors and space-based catalogs, expected soon, one could think that the twilight hours of astronomical photography have become. On opposite of that point of view such astronomers as D.Monet (U.S.N.O.), L.G.Taff (STScI), M.K.Tsvetkov (IA BAS) and some other have contended the several ways of the photographic astronomy evolution. One of them sounds as: "...special efforts must be taken to extract useful information from the photographic archives before the plates degrade and the technology required to measure them disappears". Another is the minimization of the systematic errors of ground-based star catalogs by employment of certain reduction technology and a dense enough and precise space-based star reference catalogs. In addition to that the using of the higher resolution and quantum efficiency emulsions such as Tech Pan and some of the new methods of processing of the digitized information hold great promise for future deep (B<25) surveys (Bland-Hawthorn et al. 1993, AJ, 106, 2154). Thus not only the hard working of all existing automatic measuring machines is apparently needed but the designing, development and employment of a new generation of portable, mobile scanners is very necessary. The classification, main parameters of some modern automatic measuring machines, developed with them scientific researches and some of the used methods of high accuracy, reliability and certainly ensuring are reported in that paper. This work are supported by Grant N U4I000 from International Science Foundation.

  5. Robust System for Automated Identification of Martian Impact Craters

    NASA Astrophysics Data System (ADS)

    Stepinski, T. F.; Mendenhall, M. P.

    2006-12-01

    Detailed analysis of the number and morphology of impact craters on Mars provides the worth of information about the geologic history of its surface. Global catalogs of Martian craters have been compiled (for example, the Barlow catalog) but they are not comprehensive, especially for small craters. Existing methods for machine detection of craters from images suffer from low efficiency and are not practical for global surveys. We have developed a robust two-stage system for an automated cataloging of craters from digital topography data (DEM). In the first stage an innovative crater-finding transform is performed on a DEM to identify centers of potential craters, their extents, and their basic characteristics. This stage produces a preliminary catalog. In the second stage a machine learning methods are employed to eliminate false positives. Using the MOLA derived DEMs with resolution of 1/128 degrees/pixel, we have applied our system to six ~ 106 km2 sites. The system has identified 3217 craters, 43% more than are present in the Barlow catalog. The extra finds are predominantly small craters that are most difficult to account for in manual surveys. Because our automated survey is DEM-based, the resulting catalog lists craters' depths in addition to their positions, sizes, and measures of shape. This feature significantly increases the scientific utility of any catalog generated using our system. Our initial calculations yield a training set that will be used to identify craters over the entire Martian surface with estimated accuracy of 95%. Moreover, because our method is pixel-based and scale- independent, the present training set may be used to identify craters in higher resolution DEMs derived from Mars Express HRSC images. It also can be applied to future topography data from Mars and other planets. For example, it may be utilized to catalog craters on Mercury and the Moon using altimetry data to be gathered by Messenger and Lunar Reconnaissance Orbiter spacecrafts.

  6. Name-calling in the hippocampus (and beyond): coming to terms with neuron types and properties.

    PubMed

    Hamilton, D J; Wheeler, D W; White, C M; Rees, C L; Komendantov, A O; Bergamino, M; Ascoli, G A

    2017-03-01

    Widely spread naming inconsistencies in neuroscience pose a vexing obstacle to effective communication within and across areas of expertise. This problem is particularly acute when identifying neuron types and their properties. Hippocampome.org is a web-accessible neuroinformatics resource that organizes existing data about essential properties of all known neuron types in the rodent hippocampal formation. Hippocampome.org links evidence supporting the assignment of a property to a type with direct pointers to quotes and figures. Mining this knowledge from peer-reviewed reports reveals the troubling extent of terminological ambiguity and undefined terms. Examples span simple cases of using multiple synonyms and acronyms for the same molecular biomarkers (or other property) to more complex cases of neuronal naming. New publications often use different terms without mapping them to previous terms. As a result, neurons of the same type are assigned disparate names, while neurons of different types are bestowed the same name. Furthermore, non-unique properties are frequently used as names, and several neuron types are not named at all. In order to alleviate this nomenclature confusion regarding hippocampal neuron types and properties, we introduce a new functionality of Hippocampome.org: a fully searchable, curated catalog of human and machine-readable definitions, each linked to the corresponding neuron and property terms. Furthermore, we extend our robust approach to providing each neuron type with an informative name and unique identifier by mapping all encountered synonyms and homonyms.

  7. Scientific Knowledge Discovery in Complex Semantic Networks of Geophysical Systems

    NASA Astrophysics Data System (ADS)

    Fox, P.

    2012-04-01

    The vast majority of explorations of the Earth's systems are limited in their ability to effectively explore the most important (often most difficult) problems because they are forced to interconnect at the data-element, or syntactic, level rather than at a higher scientific, or semantic, level. Recent successes in the application of complex network theory and algorithms to climate data, raise expectations that more general graph-based approaches offer the opportunity for new discoveries. In the past ~ 5 years in the natural sciences there has substantial progress in providing both specialists and non-specialists the ability to describe in machine readable form, geophysical quantities and relations among them in meaningful and natural ways, effectively breaking the prior syntax barrier. The corresponding open-world semantics and reasoning provide higher-level interconnections. That is, semantics provided around the data structures, using semantically-equipped tools, and semantically aware interfaces between science application components allowing for discovery at the knowledge level. More recently, formal semantic approaches to continuous and aggregate physical processes are beginning to show promise and are soon likely to be ready to apply to geoscientific systems. To illustrate these opportunities, this presentation presents two application examples featuring domain vocabulary (ontology) and property relations (named and typed edges in the graphs). First, a climate knowledge discovery pilot encoding and exploration of CMIP5 catalog information with the eventual goal to encode and explore CMIP5 data. Second, a multi-stakeholder knowledge network for integrated assessments in marine ecosystems, where the data is highly inter-disciplinary.

  8. Photographic observations of six comets

    NASA Astrophysics Data System (ADS)

    de Sanctis, G.; Ferreri, W.; Zappala, V.

    Sixty-nine positions of six comets are given as obtained from photographic observations made at the Observatory of Torino from October 1980 to September 1982. Positions are given for Comets Encke, Stephan-Oterma, Meier, Bradfield, Panther, and Austin. Plates were measured with a Zeiss two-coordinate measuring machine. The AGK3 catalog was used to obtain the positions of reference stars and the coordinates of an additional cataloged star near the position of the comet on the plate. The mean values of the differences between the cataloged positions were found to be 0.72 arcsec and 0.52 arcsec in right ascension and declination, respectively.

  9. There Are (super)Giants in the Sky: Searching for Misidentified Massive Stars in Algorithmically-Selected Quasar Catalogs

    NASA Astrophysics Data System (ADS)

    Dorn-Wallenstein, Trevor Z.; Levesque, Emily

    2017-11-01

    Thanks to incredible advances in instrumentation, surveys like the Sloan Digital Sky Survey have been able to find and catalog billions of objects, ranging from local M dwarfs to distant quasars. Machine learning algorithms have greatly aided in the effort to classify these objects; however, there are regimes where these algorithms fail, where interesting oddities may be found. We present here an X-ray bright quasar misidentified as a red supergiant/X-ray binary, and a subsequent search of the SDSS quasar catalog for X-ray bright stars misidentified as quasars.

  10. A Study with Computer-Based Circulation Data of the Non-Use and Use of a Large Academic Library. Final Report.

    ERIC Educational Resources Information Center

    Lubans, John, Jr.; And Others

    Computer-based circulation systems, it is widely believed, can be utilized to provide data for library use studies. The study described in this report involves using such a data base to analyze aspects of library use and non-use and types of users. Another major objective of this research was the testing of machine-readable circulation data…

  11. Public Data Set: Erratum: "Multi-point, high-speed passive ion velocity distribution diagnostic on the Pegasus Toroidal Experiment" [Rev. Sci. Instrum. 83, 10D516 (2012)

    DOE Data Explorer

    Burke, Marcus G. [University of Wisconsin-Madison] (ORCID:0000000176193724); Fonck, Raymond J. [University of Wisconsin-Madison] (ORCID:0000000294386762); Bongard, Michael W. [University of Wisconsin-Madison] (ORCID:0000000231609746); Schlossberg, David J. [University of Wisconsin-Madison] (ORCID:0000000287139448); Winz, Gregory R. [University of Wisconsin-Madison] (ORCID:0000000177627184)

    2016-07-18

    This data set contains openly-documented, machine readable digital research data corresponding to figures published in M.G. Burke et al., 'Erratum: "Multi-point, high-speed passive ion velocity distribution diagnostic on the Pegasus Toroidal Experiment" [Rev. Sci. Instrum. 83, 10D516 (2012)],' Rev. Sci. Instrum. 87, 079902 (2016).

  12. Minding the Gap: The Growing Divide Between Privacy and Surveillance Technology

    DTIC Science & Technology

    2013-06-01

    MORIS Mobile Offender Recognition and Information System MRZ Machine Readable Zone NIJ National Institute of Justice NSTC National Science and...in an increasingly mobile world, the practical result is that the most restrictive state law controls a criminal surveillance investigation...1065–1066; Smith, 2011, pp. 1–3). GPS is also used for a variety of administrative purposes ranging from mobile asset tracking (Thomas, 2007) to

  13. Description of Merged Data Base: Appendix F. The Development of Institutions of Higher Education. Theory and Assessment of Impact of Four Possible Areas of Federal Intervention.

    ERIC Educational Resources Information Center

    Jackson, Gregory A.

    Data in this appendix were used in the Office of Education-supported research study of the development of higher education institutions. Machine-readable quantitative data were gathered from the National Center for Education Statistics, the Office of Civil Rights, the Council on Financial Aid to Education, the National Education Data Library, and…

  14. Enhancing the Retrieval Effectiveness of Large Information Systems. Final Report for the Period 1 June 1975-31 December 1976.

    ERIC Educational Resources Information Center

    Becker, David S.; Pyrce, Sharon R.

    The goal of this project was to find ways of enhancing the efficiency of searching machine readable data bases. Ways are sought to transfer to the computer some of the tasks that are normally performed by the user, i.e., to further automate information retrieval. Four experiments were conducted to test the feasibility of a sequential processing…

  15. The Integration of Information Science into the Library School Curriculum at the University of Western Ontario.

    ERIC Educational Resources Information Center

    Svenonius, Elaine

    The integration of Information Science into the library school at the University of Western Ontario was the theme of a talk delivered to ASIS in October 1976 and AALS in January 1977. Two problems arise in the pursuit of integration: (1) information exists in both book form and in some other form; e.g., machine readable form, and (2) theory and…

  16. Documentation for the machine-readable version of the University of Michigan Catalogue of two-dimensional spectral types for the HD stars. Volume 2: Declinations minus 53 deg to minus 40 deg

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.

    1981-01-01

    The magnetic tape version of Volume 2 of the University of Michigan systematic reclassification program for the Henry Draper Catalogue (HD) stars is described. Volume 2 contains all HD stars in the declination range -53 degrees to 40 degrees and also exists in printed form.

  17. The Text Retrieval Conferences (TRECs)

    DTIC Science & Technology

    1998-10-01

    per- form a monolingual run in the target language to act as a baseline. Thirteen groups participated in the TREC-6 CLIR track. Three major...language; the use of machine-readable bilingual dictionaries or other existing linguistic re- sources; and the use of corpus resources to train or...formance for each method. In general, the best cross- language performance was between 50%-75% as ef- fective as a quality monolingual run. The TREC-7

  18. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shettel, D.L. Jr.; Langfeldt, S.L.; Youngquist, C.A.

    This report presents a Hydrogeochemical and Stream Sediment Reconnaissance of the Christian NTMS Quadrangle, Alaska. In addition to this abbreviated data release, more complete data are available to the public in machine-readable form. These machine-readable data, as well as quarterly or semiannual program progress reports containing further information on the HSSR program in general, or on the Los Alamos National Laboratory portion of the program in particular, are available from DOE's Technical Library at its Grand Junction Area Office. Presented in this data release are location data, field analyses, and laboratory analyses of several different sample media. For the sakemore » of brevity, many field site observations have not been included in this volume; these data are, however, available on the magnetic tape. Appendices A through D describe the sample media and summarize the analytical results for each medium. The data have been subdivided by one of the Los Alamos National Laboratory sorting programs of Zinkl and others (1981a) into groups of stream-sediment, lake-sediment, stream-water, lake-water, and ground-water samples. For each group which contains a sufficient number of observations, statistical tables, tables of raw data, and 1:1,000,000 scale maps of pertinent elements have been included in this report. Also included are maps showing results of multivariate statistical analyses.« less

  19. Multivariate statistical analysis software technologies for astrophysical research involving large data bases

    NASA Technical Reports Server (NTRS)

    Djorgovski, S. George

    1994-01-01

    We developed a package to process and analyze the data from the digital version of the Second Palomar Sky Survey. This system, called SKICAT, incorporates the latest in machine learning and expert systems software technology, in order to classify the detected objects objectively and uniformly, and facilitate handling of the enormous data sets from digital sky surveys and other sources. The system provides a powerful, integrated environment for the manipulation and scientific investigation of catalogs from virtually any source. It serves three principal functions: image catalog construction, catalog management, and catalog analysis. Through use of the GID3* Decision Tree artificial induction software, SKICAT automates the process of classifying objects within CCD and digitized plate images. To exploit these catalogs, the system also provides tools to merge them into a large, complete database which may be easily queried and modified when new data or better methods of calibrating or classifying become available. The most innovative feature of SKICAT is the facility it provides to experiment with and apply the latest in machine learning technology to the tasks of catalog construction and analysis. SKICAT provides a unique environment for implementing these tools for any number of future scientific purposes. Initial scientific verification and performance tests have been made using galaxy counts and measurements of galaxy clustering from small subsets of the survey data, and a search for very high redshift quasars. All of the tests were successful, and produced new and interesting scientific results. Attachments to this report give detailed accounts of the technical aspects for multivariate statistical analysis of small and moderate-size data sets, called STATPROG. The package was tested extensively on a number of real scientific applications, and has produced real, published results.

  20. A catalog of galaxy morphology and photometric redshift

    NASA Astrophysics Data System (ADS)

    Paul, Nicholas; Shamir, Lior

    2018-01-01

    Morphology carries important information about the physical characteristics of a galaxy. Here we used machine learning to produce a catalog of ~3,000,000 SDSS galaxies classified by their broad morphology into spiral and elliptical galaxies. Comparison of the catalog to Galaxy Zooshows that the catalog contains a subset of 1.7*10^6 galaxies classified with the same level of consistency as the debiased “superclean” sub-sample. In addition to the morphology, we also computed the photometric redshifts of the galaxies. Several pattern recognition algorithms and variable selection strategies were tested, and the best accuracy of mean absolute error of ~0.0062 was achieved by using random forest with a combination of manually and automatically selected variables. The catalog shows that for redshift lower than 0.085 galaxies that visually look spiral become more prevalent as the redshift gets higher. For redshift greater than 0.085 galaxies thatvisually look elliptical become more prevalent. The catalog as well as the source code used to produce it is publicly available athttps://figshare.com/articles/Morphology_and_photometric_redshift_catalog/4833593 .

  1. Parts Quality Management: Direct Part Marking via Data Matrix Symbols for Mission Assurance

    NASA Technical Reports Server (NTRS)

    Moss, Chantrice

    2013-01-01

    A United States Government Accountability Office (GAO) review of twelve NASA programs found widespread parts quality problems contributing to significant cost overruns, schedule delays, and reduced system reliability. Direct part-marking with Data Matrix symbols could significantly improve the quality of inventory control and parts lifecycle management. This paper examines the feasibility of using 15 marking technologies for use in future NASA programs. A structural analysis is based on marked material type, operational environment (e.g., ground, suborbital, orbital), durability of marks, ease of operation, reliability, and affordability. A cost-benefits analysis considers marking technology (data plates, label printing, direct part marking) and marking types (two-dimensional machine-readable, human-readable). Previous NASA parts marking efforts and historical cost data are accounted for, including in-house vs. outsourced marking. Some marking methods are still under development. While this paper focuses on NASA programs, results may be applicable to a variety of industrial environments.

  2. Applying Machine Learning to Star Cluster Classification

    NASA Astrophysics Data System (ADS)

    Fedorenko, Kristina; Grasha, Kathryn; Calzetti, Daniela; Mahadevan, Sridhar

    2016-01-01

    Catalogs describing populations of star clusters are essential in investigating a range of important issues, from star formation to galaxy evolution. Star cluster catalogs are typically created in a two-step process: in the first step, a catalog of sources is automatically produced; in the second step, each of the extracted sources is visually inspected by 3-to-5 human classifiers and assigned a category. Classification by humans is labor-intensive and time consuming, thus it creates a bottleneck, and substantially slows down progress in star cluster research.We seek to automate the process of labeling star clusters (the second step) through applying supervised machine learning techniques. This will provide a fast, objective, and reproducible classification. Our data is HST (WFC3 and ACS) images of galaxies in the distance range of 3.5-12 Mpc, with a few thousand star clusters already classified by humans as a part of the LEGUS (Legacy ExtraGalactic UV Survey) project. The classification is based on 4 labels (Class 1 - symmetric, compact cluster; Class 2 - concentrated object with some degree of asymmetry; Class 3 - multiple peak system, diffuse; and Class 4 - spurious detection). We start by looking at basic machine learning methods such as decision trees. We then proceed to evaluate performance of more advanced techniques, focusing on convolutional neural networks and other Deep Learning methods. We analyze the results, and suggest several directions for further improvement.

  3. VizieR Online Data Catalog: Cordoba Carte du Ciel-Astrographic Catalog, CCAC (Orellana+, 2010)

    NASA Astrophysics Data System (ADS)

    Orellana, R. B.; de Biasi, M. S.; Bustos Fierro, I. H.; Calderon, J. H.

    2010-07-01

    This is Cordoba Carte du Ciel-Astrographic Catalog (CCAC) constructed from four Carte du Ciel and one Astrographic Catalog photographic plates for first epoch positions in the region of the open cluster Collinder 132. The plates were digitized using the MAMA measuring machine from the Paris Observatory. Stars from Tycho-2 catalogue (Hog et al., 2000, Cat. I/259) were used as reference stars. Every plate was reduced independently from the others adopting a first order polynomial in the measured coordinates. Proper motions were calculated using the CCAC positions as first epoch, and as second epoch the positions given by UCAC2 (Zacharias et al., 2004, Cat. I/289) and USNO-B1.0 (Monet et al., 2003, Cat. I/284). (2 data files).

  4. Studying the Sky/Planets Can Drown You in Images: Machine Learning Solutions at JPL/Caltech

    NASA Technical Reports Server (NTRS)

    Fayyad, U. M.

    1995-01-01

    JPL is working to develop a domain-independent system capable of small-scale object recognition in large image databases for science analysis. Two applications discussed are the cataloging of three billion sky objects in the Sky Image Cataloging and Analysis Tool (SKICAT) and the detection of possibly one million small volcanoes visible in the Magellan synthetic aperture radar images of Venus (JPL Adaptive Recognition Tool, JARTool).

  5. Units in the VO Version 1.0

    NASA Astrophysics Data System (ADS)

    Derriere, Sebastien; Gray, Norman; Demleitner, Markus; Louys, Mireille; Ochsenbein, Francois; Derriere, Sebastien; Gray, Norman

    2014-05-01

    This document describes a recommended syntax for writing the string representation of unit labels ("VOUnits"). In addition, it describes a set of recognised and deprecated units, which is as far as possible consistent with other relevant standards (BIPM, ISO/IEC and the IAU). The intention is that units written to conform to this specification will likely also be parsable by other well-known parsers. To this end, we include machine-readable grammars for other units syntaxes.

  6. OCTANET--an electronic library network: I. Design and development.

    PubMed Central

    Johnson, M F; Pride, R B

    1983-01-01

    The design and development of the OCTANET system for networking among medical libraries in the midcontinental region is described. This system's features and configuration may be attributed, at least in part, to normal evolution of technology in library networking, remote access to computers, and development of machine-readable data bases. Current functions and services of the system are outlined and implications for future developments in computer-based networking are discussed. PMID:6860825

  7. A bill to extend the period during which Iraqis who were employed by the United States Government in Iraq may be granted special immigrant status and to temporarily increase the fee or surcharge for processing machine-readable nonimmigrant visas.

    THOMAS, 113th Congress

    Sen. Shaheen, Jeanne [D-NH

    2013-09-30

    House - 10/01/2013 Held at the desk. (All Actions) Notes: For further action, see H.R.3233, which became Public Law 113-42 on 10/4/2013. Tracker: This bill has the status Passed SenateHere are the steps for Status of Legislation:

  8. Identification of 1.4 Million Active Galactic Nuclei In the Mid-Infrared Using WISE Data

    DTIC Science & Technology

    2015-11-01

    galaxies – infrared: stars – galaxies : active – quasars: general Supporting material: machine-readable table 1. INTRODUCTION The International Celestial...AGN-dominated galaxies , optical emission is thought to originate from the compact accretion disk surrounding the supermassive black hole (SMBH), while... galaxies , an optical centroid can be shifted relative to the radio position because of contamination from the host galaxy . Depending on the distance to

  9. Oil and gas field code master list, 1993

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    This document contains data collected through October 1993 and provides standardized field name spellings and codes for all identified oil and/or gas fields in the United States. Other Federal and State government agencies, as well as industry, use the EIA Oil and Gas Field Code Master List as the standard for field identification. A machine-readable version of the Oil and Gas Field Code Master List is available from the National Technical Information Service.

  10. BioImg.org: A Catalog of Virtual Machine Images for the Life Sciences

    PubMed Central

    Dahlö, Martin; Haziza, Frédéric; Kallio, Aleksi; Korpelainen, Eija; Bongcam-Rudloff, Erik; Spjuth, Ola

    2015-01-01

    Virtualization is becoming increasingly important in bioscience, enabling assembly and provisioning of complete computer setups, including operating system, data, software, and services packaged as virtual machine images (VMIs). We present an open catalog of VMIs for the life sciences, where scientists can share information about images and optionally upload them to a server equipped with a large file system and fast Internet connection. Other scientists can then search for and download images that can be run on the local computer or in a cloud computing environment, providing easy access to bioinformatics environments. We also describe applications where VMIs aid life science research, including distributing tools and data, supporting reproducible analysis, and facilitating education. BioImg.org is freely available at: https://bioimg.org. PMID:26401099

  11. BioImg.org: A Catalog of Virtual Machine Images for the Life Sciences.

    PubMed

    Dahlö, Martin; Haziza, Frédéric; Kallio, Aleksi; Korpelainen, Eija; Bongcam-Rudloff, Erik; Spjuth, Ola

    2015-01-01

    Virtualization is becoming increasingly important in bioscience, enabling assembly and provisioning of complete computer setups, including operating system, data, software, and services packaged as virtual machine images (VMIs). We present an open catalog of VMIs for the life sciences, where scientists can share information about images and optionally upload them to a server equipped with a large file system and fast Internet connection. Other scientists can then search for and download images that can be run on the local computer or in a cloud computing environment, providing easy access to bioinformatics environments. We also describe applications where VMIs aid life science research, including distributing tools and data, supporting reproducible analysis, and facilitating education. BioImg.org is freely available at: https://bioimg.org.

  12. PDB explorer -- a web based algorithm for protein annotation viewer and 3D visualization.

    PubMed

    Nayarisseri, Anuraj; Shardiwal, Rakesh Kumar; Yadav, Mukesh; Kanungo, Neha; Singh, Pooja; Shah, Pratik; Ahmed, Sheaza

    2014-12-01

    The PDB file format, is a text format characterizing the three dimensional structures of macro molecules available in the Protein Data Bank (PDB). Determined protein structure are found in coalition with other molecules or ions such as nucleic acids, water, ions, Drug molecules and so on, which therefore can be described in the PDB format and have been deposited in PDB database. PDB is a machine generated file, it's not human readable format, to read this file we need any computational tool to understand it. The objective of our present study is to develop a free online software for retrieval, visualization and reading of annotation of a protein 3D structure which is available in PDB database. Main aim is to create PDB file in human readable format, i.e., the information in PDB file is converted in readable sentences. It displays all possible information from a PDB file including 3D structure of that file. Programming languages and scripting languages like Perl, CSS, Javascript, Ajax, and HTML have been used for the development of PDB Explorer. The PDB Explorer directly parses the PDB file, calling methods for parsed element secondary structure element, atoms, coordinates etc. PDB Explorer is freely available at http://www.pdbexplorer.eminentbio.com/home with no requirement of log-in.

  13. The NATO thesaurus project

    NASA Technical Reports Server (NTRS)

    Krueger, Jonathan

    1990-01-01

    This document describes functionality to be developed to support the NATO technical thesaurus. Described are the specificity of the thesaurus structure and function; the distinction between the thesaurus information and its representation in a given online, machine readable, or printed form; the enhancement of the thesaurus with the assignment of COSATI codes (fields and groups) to posting terms, the integration of DTIC DRIT and NASA thesauri related terminology, translation of posting terms into French; and the provision of a basis for system design.

  14. An atlas of stellar spectra between 2.00 and 2.45 micrometers (Arnaud, Gilmore, and Collier Cameron 1989)

    NASA Technical Reports Server (NTRS)

    Warren, Wayne N., Jr.

    1990-01-01

    The machine-readable version of the atlas, as it is currently being distributed from the Astronomical Data Center, is described. The atlas represent a collection of spectra in the wavelength range 2.00 to 2.45 micros having a resolution of approximately 0.02 micron. The sample of 73 stars includes a supergiant, giants, dwarfs, and subdwarfs with a chemical abundance range of about -2 to +0.5 dex.

  15. Passport examination by a confocal-type laser profile microscope.

    PubMed

    Sugawara, Shigeru

    2008-06-10

    The author proposes a nondestructive and highly precise method of measuring the thickness of a film pasted on a passport using a confocal-type laser profile microscope. The effectiveness of this method in passport examination is demonstrated. A confocal-type laser profile microscope is used to create profiles of the film surface and film-paper interface; these profiles are used to calculate the film thickness by employing an algorithm developed by the author. The film thicknesses of the passport samples--35 genuine and 80 counterfeit Japanese passports--are measured nondestructively. The intra-sample standard deviation of the film thicknesses of the genuine and counterfeit Japanese passports was of the order of 1 microm The intersample standard deviations of the film thicknesses of passports forged using the same tools and techniques are expected to be of the order of 1 microm. The thickness values of the films on the machine-readable genuine passports ranged between 31.95 microm and 36.95 microm. The likelihood ratio of this method in the authentication of machine-readable Japanese genuine passports is 11.7. Therefore, this method is effective for the authentification of genuine passports. Since the distribution of the film thickness of all forged passports was considerably larger than the accuracy of this method, this method is considered effective also for revealing the relation among the forged passports and acquiring proof of the crime.

  16. Index to selected machine-readable geohydrologic data for Precambrian through Cretaceous rocks in Kansas

    USGS Publications Warehouse

    Spinazola, J.M.; Hansen, C.V.; Underwood, E.J.; Kenny, J.F.; Wolf, R.J.

    1987-01-01

    Machine-readable geohydrologic data for Precambrian through Cretaceous rocks in Kansas were compiled as part of the USGS Central Midwest Regional Aquifer System Analysis. The geohydrologic data include log, water quality, water level, hydraulics, and water use information. The log data consist of depths to the top of selected geologic formations determined from about 275 sites with geophysical logs and formation lithologies from about 190 sites with lithologic logs. The water quality data consist of about 10,800 analyses, of which about 1 ,200 are proprietary. The water level data consist of about 4 ,480 measured water levels and about 4,175 equivalent freshwater hydraulic heads, of which about 3,745 are proprietary. The hydraulics data consist of results from about 30 specific capacity tests and about 20 aquifer tests, and interpretations of about 285 drill stem tests (of which about 60 are proprietary) and about 75 core-sample analyses. The water use data consist of estimates of freshwater withdrawals from Precambrian through Cretaceous geohydrologic units for each of the 105 counties in Kansas. Average yearly withdrawals were estimated for each decade from 1940 to 1980. All the log and water use data and the nonproprietary parts of the water quality , water level, and hydraulics data are available on magnetic tape from the USGS office in Lawrence, Kansas. (Author 's abstract)

  17. The JPL Library Information Retrieval System

    ERIC Educational Resources Information Center

    Walsh, Josephine

    1975-01-01

    The development, capabilities, and products of the computer-based retrieval system of the Jet Propulsion Laboratory Library are described. The system handles books and documents, produces a book catalog, and provides a machine search capability. (Author)

  18. Multivariate Statistical Analysis Software Technologies for Astrophysical Research Involving Large Data Bases

    NASA Technical Reports Server (NTRS)

    Djorgovski, S. G.

    1994-01-01

    We developed a package to process and analyze the data from the digital version of the Second Palomar Sky Survey. This system, called SKICAT, incorporates the latest in machine learning and expert systems software technology, in order to classify the detected objects objectively and uniformly, and facilitate handling of the enormous data sets from digital sky surveys and other sources. The system provides a powerful, integrated environment for the manipulation and scientific investigation of catalogs from virtually any source. It serves three principal functions: image catalog construction, catalog management, and catalog analysis. Through use of the GID3* Decision Tree artificial induction software, SKICAT automates the process of classifying objects within CCD and digitized plate images. To exploit these catalogs, the system also provides tools to merge them into a large, complex database which may be easily queried and modified when new data or better methods of calibrating or classifying become available. The most innovative feature of SKICAT is the facility it provides to experiment with and apply the latest in machine learning technology to the tasks of catalog construction and analysis. SKICAT provides a unique environment for implementing these tools for any number of future scientific purposes. Initial scientific verification and performance tests have been made using galaxy counts and measurements of galaxy clustering from small subsets of the survey data, and a search for very high redshift quasars. All of the tests were successful and produced new and interesting scientific results. Attachments to this report give detailed accounts of the technical aspects of the SKICAT system, and of some of the scientific results achieved to date. We also developed a user-friendly package for multivariate statistical analysis of small and moderate-size data sets, called STATPROG. The package was tested extensively on a number of real scientific applications and has produced real, published results.

  19. Artillery ammunition marking tests

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Weil, B.S.; Lewis, J.C.

    1995-04-01

    This report describes the testing results of two approaches being considered for marking ink artillery ammunition with machine-readable data symbols. The first approach used ink-jet printing directly onto projectiles, and the second approach employed thermal-transfer printing onto self-adhesive labels that are subsequently applied automatically to projectiles. The objectives of this evaluation for each marking technology were to (1) determine typical system performance characteristics using the best commercially available equipment and (2) identify any special requirements necessary for handling ammunition when these technologies are employed.

  20. Method of modifying a volume mesh using sheet extraction

    DOEpatents

    Borden, Michael J [Albuquerque, NM; Shepherd, Jason F [Albuquerque, NM

    2007-02-20

    A method and machine-readable medium provide a technique to modify a hexahedral finite element volume mesh using dual generation and sheet extraction. After generating a dual of a volume stack (mesh), a predetermined algorithm may be followed to modify the volume mesh of hexahedral elements. The predetermined algorithm may include the steps of determining a sheet of hexahedral mesh elements, generating nodes for merging, and merging the nodes to delete the sheet of hexahedral mesh elements and modify the volume mesh.

  1. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Finnell, Joshua Eugene

    US President Barack Obama issued Executive Order 13642-Making Open and Machine Readable the New Default for Government - on May, 9 2013, mandating, wherever legally permissible and possible, that US Government information be made open to the public.[1] This edict accelerated the construction of and framework for data repositories, and data citation principles and practices, such as data.gov. As a corollary, researchers across the country's national laboratories found themselves creating data management plans, applying data set metadata standards, and ensuring the long-term access of data for federally funded scientific research.

  2. Social Ontology Documentation for Knowledge Externalization

    NASA Astrophysics Data System (ADS)

    Aranda-Corral, Gonzalo A.; Borrego-Díaz, Joaquín; Jiménez-Mavillard, Antonio

    Knowledge externalization and organization is a major challenge that companies must face. Also, they have to ask whether is possible to enhance its management. Mechanical processing of information represents a chance to carry out these tasks, as well as to turn intangible knowledge assets into real assets. Machine-readable knowledge provides a basis to enhance knowledge management. A promising approach is the empowering of Knowledge Externalization by the community (users, employees). In this paper, a social semantic tool (called OntoxicWiki) for enhancing the quality of knowledge is presented.

  3. Documentation for the machine-readable version of the thirteen color photometry of 1380 bright stars

    NASA Technical Reports Server (NTRS)

    Warren, W. H., Jr.; Roman, N. G.

    1981-01-01

    The magnetic tape version of the catalogue of thirteen-color photometry of 1380 bright stars, containing data on the 13 color medium narrow band photometric system is described. Observations of essentially all stars brighter than fifth visual magnitude north of delta = -20 deg and brighter than fourth visual magnitude south of delta = -20 deg are included. It is intended to enable users to read and process the tape without the common difficulties and uncertainties.

  4. Parts quality management: Direct part marking of data matrix symbol for mission assurance

    NASA Astrophysics Data System (ADS)

    Moss, Chantrice; Chakrabarti, Suman; Scott, David W.

    A United States Government Accountability Office (GAO) review of twelve NASA programs found widespread parts quality problems contributing to significant cost overruns, schedule delays, and reduced system reliability. Direct part marking with Data Matrix symbols could significantly improve the quality of inventory control and parts lifecycle management. This paper examines the feasibility of using direct part marking technologies for use in future NASA programs. A structural analysis is based on marked material type, operational environment (e.g., ground, suborbital, Low Earth Orbit), durability of marks, ease of operation, reliability, and affordability. A cost-benefits analysis considers marking technology (label printing, data plates, and direct part marking) and marking types (two-dimensional machine-readable, human-readable). Previous NASA parts marking efforts and historical cost data are accounted for, including in-house vs. outsourced marking. Some marking methods are still under development. While this paper focuses on NASA programs, results may be applicable to a variety of industrial environments.

  5. Parts Quality Management: Direct Part Marking of Data Matrix Symbol for Mission Assurance

    NASA Technical Reports Server (NTRS)

    Moss, Chantrice; Chakrabarti, Suman; Scott, David W.

    2013-01-01

    A United States Government Accountability Office (GAO) review of twelve NASA programs found widespread parts quality problems contributing to significant cost overruns, schedule delays, and reduced system reliability. Direct part marking with Data Matrix symbols could significantly improve the quality of inventory control and parts lifecycle management. This paper examines the feasibility of using direct part marking technologies for use in future NASA programs. A structural analysis is based on marked material type, operational environment (e.g., ground, suborbital, Low Earth Orbit), durability of marks, ease of operation, reliability, and affordability. A cost-benefits analysis considers marking technology (label printing, data plates, and direct part marking) and marking types (two-dimensional machine-readable, human-readable). Previous NASA parts marking efforts and historical cost data are accounted for, including inhouse vs. outsourced marking. Some marking methods are still under development. While this paper focuses on NASA programs, results may be applicable to a variety of industrial environments.

  6. THE MILKY WAY PROJECT: LEVERAGING CITIZEN SCIENCE AND MACHINE LEARNING TO DETECT INTERSTELLAR BUBBLES

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Beaumont, Christopher N.; Williams, Jonathan P.; Goodman, Alyssa A.

    We present Brut, an algorithm to identify bubbles in infrared images of the Galactic midplane. Brut is based on the Random Forest algorithm, and uses bubbles identified by >35,000 citizen scientists from the Milky Way Project to discover the identifying characteristics of bubbles in images from the Spitzer Space Telescope. We demonstrate that Brut's ability to identify bubbles is comparable to expert astronomers. We use Brut to re-assess the bubbles in the Milky Way Project catalog, and find that 10%-30% of the objects in this catalog are non-bubble interlopers. Relative to these interlopers, high-reliability bubbles are more confined to themore » mid-plane, and display a stronger excess of young stellar objects along and within bubble rims. Furthermore, Brut is able to discover bubbles missed by previous searches—particularly bubbles near bright sources which have low contrast relative to their surroundings. Brut demonstrates the synergies that exist between citizen scientists, professional scientists, and machine learning techniques. In cases where ''untrained' citizens can identify patterns that machines cannot detect without training, machine learning algorithms like Brut can use the output of citizen science projects as input training sets, offering tremendous opportunities to speed the pace of scientific discovery. A hybrid model of machine learning combined with crowdsourced training data from citizen scientists can not only classify large quantities of data, but also address the weakness of each approach if deployed alone.« less

  7. Semantic Technologies and Bio-Ontologies.

    PubMed

    Gutierrez, Fernando

    2017-01-01

    As information available through data repositories constantly grows, the need for automated mechanisms for linking, querying, and sharing data has become a relevant factor both in research and industry. This situation is more evident in research fields such as the life sciences, where new experiments by different research groups are constantly generating new information regarding a wide variety of related study objects. However, current methods for representing information and knowledge are not suited for machine processing. The Semantic Technologies are a set of standards and protocols that intend to provide methods for representing and handling data that encourages reusability of information and is machine-readable. In this chapter, we will provide a brief introduction to Semantic Technologies, and how these protocols and standards have been incorporated into the life sciences to facilitate dissemination and access to information.

  8. A Low-Cost, Efficient, Machine-Assisted Manual Circulation System

    ERIC Educational Resources Information Center

    Stangl, Peter

    1975-01-01

    A circulation system uses plastic embossed user cards, an addressograph electric imprinter, a copy of the catalog card as a book card, and a pocket imprinted by the user's card and holding the book card during circulation. (LS)

  9. Scalable Machine Learning for Massive Astronomical Datasets

    NASA Astrophysics Data System (ADS)

    Ball, Nicholas M.; Gray, A.

    2014-04-01

    We present the ability to perform data mining and machine learning operations on a catalog of half a billion astronomical objects. This is the result of the combination of robust, highly accurate machine learning algorithms with linear scalability that renders the applications of these algorithms to massive astronomical data tractable. We demonstrate the core algorithms kernel density estimation, K-means clustering, linear regression, nearest neighbors, random forest and gradient-boosted decision tree, singular value decomposition, support vector machine, and two-point correlation function. Each of these is relevant for astronomical applications such as finding novel astrophysical objects, characterizing artifacts in data, object classification (including for rare objects), object distances, finding the important features describing objects, density estimation of distributions, probabilistic quantities, and exploring the unknown structure of new data. The software, Skytree Server, runs on any UNIX-based machine, a virtual machine, or cloud-based and distributed systems including Hadoop. We have integrated it on the cloud computing system of the Canadian Astronomical Data Centre, the Canadian Advanced Network for Astronomical Research (CANFAR), creating the world's first cloud computing data mining system for astronomy. We demonstrate results showing the scaling of each of our major algorithms on large astronomical datasets, including the full 470,992,970 objects of the 2 Micron All-Sky Survey (2MASS) Point Source Catalog. We demonstrate the ability to find outliers in the full 2MASS dataset utilizing multiple methods, e.g., nearest neighbors. This is likely of particular interest to the radio astronomy community given, for example, that survey projects contain groups dedicated to this topic. 2MASS is used as a proof-of-concept dataset due to its convenience and availability. These results are of interest to any astronomical project with large and/or complex datasets that wishes to extract the full scientific value from its data.

  10. Scalable Machine Learning for Massive Astronomical Datasets

    NASA Astrophysics Data System (ADS)

    Ball, Nicholas M.; Astronomy Data Centre, Canadian

    2014-01-01

    We present the ability to perform data mining and machine learning operations on a catalog of half a billion astronomical objects. This is the result of the combination of robust, highly accurate machine learning algorithms with linear scalability that renders the applications of these algorithms to massive astronomical data tractable. We demonstrate the core algorithms kernel density estimation, K-means clustering, linear regression, nearest neighbors, random forest and gradient-boosted decision tree, singular value decomposition, support vector machine, and two-point correlation function. Each of these is relevant for astronomical applications such as finding novel astrophysical objects, characterizing artifacts in data, object classification (including for rare objects), object distances, finding the important features describing objects, density estimation of distributions, probabilistic quantities, and exploring the unknown structure of new data. The software, Skytree Server, runs on any UNIX-based machine, a virtual machine, or cloud-based and distributed systems including Hadoop. We have integrated it on the cloud computing system of the Canadian Astronomical Data Centre, the Canadian Advanced Network for Astronomical Research (CANFAR), creating the world's first cloud computing data mining system for astronomy. We demonstrate results showing the scaling of each of our major algorithms on large astronomical datasets, including the full 470,992,970 objects of the 2 Micron All-Sky Survey (2MASS) Point Source Catalog. We demonstrate the ability to find outliers in the full 2MASS dataset utilizing multiple methods, e.g., nearest neighbors, and the local outlier factor. 2MASS is used as a proof-of-concept dataset due to its convenience and availability. These results are of interest to any astronomical project with large and/or complex datasets that wishes to extract the full scientific value from its data.

  11. The NEOUCOM Cooperative Cataloging Service: development and review of the first four years.

    PubMed Central

    Miller, D R

    1983-01-01

    The Basic Medical Sciences Library of the Northeastern Ohio Universities College of Medicine (NEOUCOM) provided a Cooperative Cataloging Service to fourteen of its affiliated hospitals' libraries since March 1978, using the OCLC system. Analysis of the first four years of service showed that the hospital libraries spent almost $30,000 to catalog more than 18,000 titles. Personnel expenses and other costs eclipsed the savings from a 31.3% duplication rate. Centralized bibliographic control control and the principal by-product of the service, a uniform, machine-related data base, provided the foundation for an on-line integrated library system to serve the consortium. The hospital libraries contributed 44% of the unique titles in this data base, which emphasis the need to share resources and continue cooperation. PMID:6860826

  12. The NEOUCOM Cooperative Cataloging Service: development and review of the first four years.

    PubMed

    Miller, D R

    1983-04-01

    The Basic Medical Sciences Library of the Northeastern Ohio Universities College of Medicine (NEOUCOM) provided a Cooperative Cataloging Service to fourteen of its affiliated hospitals' libraries since March 1978, using the OCLC system. Analysis of the first four years of service showed that the hospital libraries spent almost $30,000 to catalog more than 18,000 titles. Personnel expenses and other costs eclipsed the savings from a 31.3% duplication rate. Centralized bibliographic control control and the principal by-product of the service, a uniform, machine-related data base, provided the foundation for an on-line integrated library system to serve the consortium. The hospital libraries contributed 44% of the unique titles in this data base, which emphasis the need to share resources and continue cooperation.

  13. Automatic Earthquake Detection by Active Learning

    NASA Astrophysics Data System (ADS)

    Bergen, K.; Beroza, G. C.

    2017-12-01

    In recent years, advances in machine learning have transformed fields such as image recognition, natural language processing and recommender systems. Many of these performance gains have relied on the availability of large, labeled data sets to train high-accuracy models; labeled data sets are those for which each sample includes a target class label, such as waveforms tagged as either earthquakes or noise. Earthquake seismologists are increasingly leveraging machine learning and data mining techniques to detect and analyze weak earthquake signals in large seismic data sets. One of the challenges in applying machine learning to seismic data sets is the limited labeled data problem; learning algorithms need to be given examples of earthquake waveforms, but the number of known events, taken from earthquake catalogs, may be insufficient to build an accurate detector. Furthermore, earthquake catalogs are known to be incomplete, resulting in training data that may be biased towards larger events and contain inaccurate labels. This challenge is compounded by the class imbalance problem; the events of interest, earthquakes, are infrequent relative to noise in continuous data sets, and many learning algorithms perform poorly on rare classes. In this work, we investigate the use of active learning for automatic earthquake detection. Active learning is a type of semi-supervised machine learning that uses a human-in-the-loop approach to strategically supplement a small initial training set. The learning algorithm incorporates domain expertise through interaction between a human expert and the algorithm, with the algorithm actively posing queries to the user to improve detection performance. We demonstrate the potential of active machine learning to improve earthquake detection performance with limited available training data.

  14. Uniform Slavic Transliteration Alphabet (USTA).

    ERIC Educational Resources Information Center

    Dekleva, Borut

    The Uniform Slavic Transliteration Alphabet (USTA) was designed primarily with the following objectives: to aid librarians (catalogers and bibliographers), information scientists, transliterators, and editors of bibliographic works of the many Slavic tongues; and to serve as original research for the further development of a machine-readable…

  15. A Comparison of Costs of Searching the Machine-Readable Data Bases ERIC and "Psychological Abstracts" in an Annual Subscription Rate System Against Costs Estimated for the Same Searches Done in the Lockheed DIALOG System and the System Development Corporation for ERIC, and the Lockheed DIALOG System and PASAT for "Psychological Abstracts."

    ERIC Educational Resources Information Center

    Palmer, Crescentia

    A comparison of costs for computer-based searching of Psychological Abstracts and Educational Resources Information Center (ERIC) systems by the New York State Library at Albany was produced by combining data available from search request forms and from bills from the contract subscription service, the State University of New…

  16. Ink-constrained halftoning with application to QR codes

    NASA Astrophysics Data System (ADS)

    Bayeh, Marzieh; Compaan, Erin; Lindsey, Theodore; Orlow, Nathan; Melczer, Stephen; Voller, Zachary

    2014-01-01

    This paper examines adding visually significant, human recognizable data into QR codes without affecting their machine readability by utilizing known methods in image processing. Each module of a given QR code is broken down into pixels, which are halftoned in such a way as to keep the QR code structure while revealing aspects of the secondary image to the human eye. The loss of information associated to this procedure is discussed, and entropy values are calculated for examples given in the paper. Numerous examples of QR codes with embedded images are included.

  17. Method of generating a surface mesh

    DOEpatents

    Shepherd, Jason F [Albuquerque, NM; Benzley, Steven [Provo, UT; Grover, Benjamin T [Tracy, CA

    2008-03-04

    A method and machine-readable medium provide a technique to generate and modify a quadrilateral finite element surface mesh using dual creation and modification. After generating a dual of a surface (mesh), a predetermined algorithm may be followed to generate and modify a surface mesh of quadrilateral elements. The predetermined algorithm may include the steps of generating two-dimensional cell regions in dual space, determining existing nodes in primal space, generating new nodes in the dual space, and connecting nodes to form the quadrilateral elements (faces) for the generated and modifiable surface mesh.

  18. Atlas of Vega: 3850-6860 Å

    NASA Astrophysics Data System (ADS)

    Kim, Hyun-Sook; Han, Inwoo; Valyavin, G.; Lee, Byeong-Cheol; Shimansky, V.; Galazutdinov, G. A.

    2009-10-01

    We present a high resolving power (λ/Δλ = 90,000) and high signal-to-noise ratio (˜700) spectral atlas of Vega covering the 3850-6860 Å wavelength range. The atlas is a result of averaging of spectra recorded with the aid of the echelle spectrograph BOES fed by the 1.8 m telescope at Bohyunsan Observatory (Korea). The atlas is provided only in machine-readable form (electronic data file) and will be available in the SIMBAD database upon publication. Based on data collected with the 1.8 m telescope operated at BOAO Observatory, Korea.

  19. Readability versus Leveling.

    ERIC Educational Resources Information Center

    Fry, Edward

    2002-01-01

    Shows some similarities and differences between readability formulas and leveling procedures and reports some current large-scale uses of readability formulas. Presents a dictionary definition of readability and leveling, and a history and background of readability and leveling. Discusses what goes into determining readability and leveling scores.…

  20. A Modular Framework for Transforming Structured Data into HTML with Machine-Readable Annotations

    NASA Astrophysics Data System (ADS)

    Patton, E. W.; West, P.; Rozell, E.; Zheng, J.

    2010-12-01

    There is a plethora of web-based Content Management Systems (CMS) available for maintaining projects and data, i.a. However, each system varies in its capabilities and often content is stored separately and accessed via non-uniform web interfaces. Moving from one CMS to another (e.g., MediaWiki to Drupal) can be cumbersome, especially if a large quantity of data must be adapted to the new system. To standardize the creation, display, management, and sharing of project information, we have assembled a framework that uses existing web technologies to transform data provided by any service that supports the SPARQL Protocol and RDF Query Language (SPARQL) queries into HTML fragments, allowing it to be embedded in any existing website. The framework utilizes a two-tier XML Stylesheet Transformation (XSLT) that uses existing ontologies (e.g., Friend-of-a-Friend, Dublin Core) to interpret query results and render them as HTML documents. These ontologies can be used in conjunction with custom ontologies suited to individual needs (e.g., domain-specific ontologies for describing data records). Furthermore, this transformation process encodes machine-readable annotations, namely, the Resource Description Framework in attributes (RDFa), into the resulting HTML, so that capable parsers and search engines can extract the relationships between entities (e.g, people, organizations, datasets). To facilitate editing of content, the framework provides a web-based form system, mapping each query to a dynamically generated form that can be used to modify and create entities, while keeping the native data store up-to-date. This open framework makes it easy to duplicate data across many different sites, allowing researchers to distribute their data in many different online forums. In this presentation we will outline the structure of queries and the stylesheets used to transform them, followed by a brief walkthrough that follows the data from storage to human- and machine-accessible web page. We conclude with a discussion on content caching and steps toward performing queries across multiple domains.

  1. Data Publication and Interoperability for Long Tail Researchers via the Open Data Repository's (ODR) Data Publisher.

    NASA Astrophysics Data System (ADS)

    Stone, N.; Lafuente, B.; Bristow, T.; Keller, R.; Downs, R. T.; Blake, D. F.; Fonda, M.; Pires, A.

    2016-12-01

    Working primarily with astrobiology researchers at NASA Ames, the Open Data Repository (ODR) has been conducting a software pilot to meet the varying needs of this multidisciplinary community. Astrobiology researchers often have small communities or operate individually with unique data sets that don't easily fit into existing database structures. The ODR constructed its Data Publisher software to allow researchers to create databases with common metadata structures and subsequently extend them to meet their individual needs and data requirements. The software accomplishes these tasks through a web-based interface that allows collaborative creation and revision of common metadata templates and individual extensions to these templates for custom data sets. This allows researchers to search disparate datasets based on common metadata established through the metadata tools, but still facilitates distinct analyses and data that may be stored alongside the required common metadata. The software produces web pages that can be made publicly available at the researcher's discretion so that users may search and browse the data in an effort to make interoperability and data discovery a human-friendly task while also providing semantic data for machine-based discovery. Once relevant data has been identified, researchers can utilize the built-in application programming interface (API) that exposes the data for machine-based consumption and integration with existing data analysis tools (e.g. R, MATLAB, Project Jupyter - http://jupyter.org). The current evolution of the project has created the Astrobiology Habitable Environments Database (AHED)[1] which provides an interface to databases connected through a common metadata core. In the next project phase, the goal is for small research teams and groups to be self-sufficient in publishing their research data to meet funding mandates and academic requirements as well as fostering increased data discovery and interoperability through human-readable and machine-readable interfaces. This project is supported by the Science-Enabling Research Activity (SERA) and NASA NNX11AP82A, MSL. [1] B. Lafuente et al. (2016) AGU, submitted.

  2. Effects of compression and individual variability on face recognition performance

    NASA Astrophysics Data System (ADS)

    McGarry, Delia P.; Arndt, Craig M.; McCabe, Steven A.; D'Amato, Donald P.

    2004-08-01

    The Enhanced Border Security and Visa Entry Reform Act of 2002 requires that the Visa Waiver Program be available only to countries that have a program to issue to their nationals machine-readable passports incorporating biometric identifiers complying with applicable standards established by the International Civil Aviation Organization (ICAO). In June 2002, the New Technologies Working Group of ICAO unanimously endorsed the use of face recognition (FR) as the globally interoperable biometric for machine-assisted identity confirmation with machine-readable travel documents (MRTDs), although Member States may elect to use fingerprint and/or iris recognition as additional biometric technologies. The means and formats are still being developed through which biometric information might be stored in the constrained space of integrated circuit chips embedded within travel documents. Such information will be stored in an open, yet unalterable and very compact format, probably as digitally signed and efficiently compressed images. The objective of this research is to characterize the many factors that affect FR system performance with respect to the legislated mandates concerning FR. A photograph acquisition environment and a commercial face recognition system have been installed at Mitretek, and over 1,400 images have been collected of volunteers. The image database and FR system are being used to analyze the effects of lossy image compression, individual differences, such as eyeglasses and facial hair, and the acquisition environment on FR system performance. Images are compressed by varying ratios using JPEG2000 to determine the trade-off points between recognition accuracy and compression ratio. The various acquisition factors that contribute to differences in FR system performance among individuals are also being measured. The results of this study will be used to refine and test efficient face image interchange standards that ensure highly accurate recognition, both for automated FR systems and human inspectors. Working within the M1-Biometrics Technical Committee of the InterNational Committee for Information Technology Standards (INCITS) organization, a standard face image format will be tested and submitted to organizations such as ICAO.

  3. Educational Aids for Visually Handicapped.

    ERIC Educational Resources Information Center

    American Printing House for the Blind, Louisville, KY.

    Listings specifying source and cost are provided of tactile aids and materials designed for the visually handicapped. Items are presented in the following categories: supply sources and catalogs for aids; braille devices, including duplicators, reading and writing aids, reading readiness materials, and writing machines, slates, and styluses; deaf…

  4. Research in image management and access

    NASA Technical Reports Server (NTRS)

    Vondran, Raymond F.; Barron, Billy J.

    1993-01-01

    Presently, the problem of over-all library system design has been compounded by the accretion of both function and structure to a basic framework of requirements. While more device power has led to increased functionality, opportunities for reducing system complexity at the user interface level have not always been pursued with equal zeal. The purpose of this book is therefore to set forth and examine these opportunities, within the general framework of human factors research in man-machine interfaces. Human factors may be viewed as a series of trade-off decisions among four polarized objectives: machine resources and user specifications; functionality and user requirements. In the past, a limiting factor was the availability of systems. However, in the last two years, over one hundred libraries supported by many different software configurations have been added to the Internet. This document includes a statistical analysis of human responses to five Internet library systems by key features, development of the ideal online catalog system, and ideal online catalog systems for libraries and information centers.

  5. Collaborative Planning of Robotic Exploration

    NASA Technical Reports Server (NTRS)

    Norris, Jeffrey; Backes, Paul; Powell, Mark; Vona, Marsette; Steinke, Robert

    2004-01-01

    The Science Activity Planner (SAP) software system includes an uplink-planning component, which enables collaborative planning of activities to be undertaken by an exploratory robot on a remote planet or on Earth. Included in the uplink-planning component is the SAP-Uplink Browser, which enables users to load multiple spacecraft activity plans into a single window, compare them, and merge them. The uplink-planning component includes a subcomponent that implements the Rover Markup Language Activity Planning format (RML-AP), based on the Extensible Markup Language (XML) format that enables the representation, within a single document, of planned spacecraft and robotic activities together with the scientific reasons for the activities. Each such document is highly parseable and can be validated easily. Another subcomponent of the uplink-planning component is the Activity Dictionary Markup Language (ADML), which eliminates the need for two mission activity dictionaries - one in a human-readable format and one in a machine-readable format. Style sheets that have been developed along with the ADML format enable users to edit one dictionary in a user-friendly environment without compromising

  6. Samples: The Story That They Tell and Our Role in Better Connecting Their Physical and Data Lifecycles.

    NASA Astrophysics Data System (ADS)

    Stall, S.

    2016-12-01

    The story of a sample starts with a proposal, a data management plan, and funded research. The sample is created, given a unique identifier (IGSN) and properly cared for during its journey to an appropriate storage location. Through its metadata, and publication information, the sample can become well known and shared with other researchers. Ultimately, a valuable sample can tell its entire story through its IGSN, associated ORCIDs, associated publication DOIs, and DOIs of data generated from sample analysis. This journey, or workflow, is in many ways still manual. Tools exist to generate IGSNs for the sample and subsamples. Publishers are committed to making IGSNs machine readable in their journals, but the connection back to the IGSN management system, specifically the System for Earth Sample Registration (SESAR) is not fully complete. Through encouragement of publishers, like AGU, and improved data management practices, such as those promoted by AGU's Data Management Assessment program, the complete lifecycle of a sample can and will be told through the journey it takes from creation, documentation (metadata), analysis, subsamples, publication, and sharing. Publishers and data facilities are using efforts like the Coalition for Publishing Data in the Earth and Space Sciences (COPDESS) to "implement and promote common policies and procedures for the publication and citation of data across Earth Science journals", including IGSNs. As our community improves its data management practices and publishers adopt and enforce machine readable use of unique sample identifiers, the ability to tell the entire story of a sample is close at hand. Better Data Management results in Better Science.

  7. The JPL Library information retrieval system

    NASA Technical Reports Server (NTRS)

    Walsh, J.

    1975-01-01

    The development, capabilities, and products of the computer-based retrieval system of the Jet Propulsion Laboratory Library are described. The system handles books and documents, produces a book catalog, and provides a machine search capability. Programs and documentation are available to the public through NASA's computer software dissemination program.

  8. Analyzing readability of medicines information material in Slovenia

    PubMed Central

    Kasesnik, Karin; Kline, Mihael

    2011-01-01

    Objective: Readability has been claimed to be an important factor for understanding texts describing health symptoms and medications. Such texts may be a factor which indirectly affects the health of the population. Despite the expertise of physicians, the readability of information sources may be important for acquiring essential treatment information. The aim of this study was to measure the readability level of medicines promotion material in Slovenia. Methods: The Flesch readability formula was modified to comply with Slovene texts. On the basis of determining the Slovene readability algorithm, the readability ease related to the readability grade level of different Slovene texts was established. In order to estimate an adjustment of the texts to the recommended readability grade level of the targeted population, readability values of English texts were set. One sample t-test and standard deviations from the arithmetic mean values were used as statistical tests. Results: The results of the research showed low readability scores of the Slovene texts. Difficult readability values were seen in different types of examined texts: in patient information leaflets, in the summaries of product characteristics, in promotional materials, while describing over-the-counter medications and in the materials for creating disease awareness. Especially low readability values were found within the texts belonging to promotional materials intended for the physicians. None of researched items, not even for the general public, were close to primary school grade readability levels and therefore could not be described as easily readable. Conclusion: This study provides an understanding of the level of readability of selected Slovene medicines information material. It was concluded that health-related texts were not compliant with general public or with healthcare professional needs. PMID:23093886

  9. The readability of American Academy of Pediatrics patient education brochures.

    PubMed

    Freda, Margaret Comerford

    2005-01-01

    The purpose of this study was to evaluate the readability of American Academy of Pediatrics (AAP) patient education brochures. Seventy-four brochures were analyzed using two readability formulas. Mean readability for all 74 brochures was grade 7.94 using the Flesch-Kincaid formula, and grade 10.1 with SMOG formula (P = .001). Using the SMOG formula, no brochures were of acceptably low (< or =8th grade) readability levels (range 8.3 to 12.7). Using the Flesch-Kincaid formula, 41 of the 74 had acceptable readability levels (< or =8th grade). The SMOG formula routinely assessed brochures 2 to 3 grade levels higher than did the Flesch-Kincaid formula. Some AAP patient education brochures have acceptably low levels of readability, but at least half are written at higher than acceptable readability levels for the general public. This study also demonstrated statistically significant variability between the two different readability formulas; had only the SMOG formula been used, all of the brochures would have had unacceptably high readability levels. Readability is an essential concept for patient education materials. Professional associations that develop and market patient education materials should test for readability and publish those readability levels on each piece of patient education so health care providers will know if the materials are appropriate for their patients.

  10. Evaluation of the readability of ACOG patient education pamphlets. The American College of Obstetricians and Gynecologists.

    PubMed

    Freda, M C; Damus, K; Merkatz, I R

    1999-05-01

    To evaluate whether ACOG's patient education pamphlets comply with the recommended readability level for health education materials intended for the general public. All 100 English-language pamphlets available during 1997 (created or revised between 1988 and 1997) were evaluated using four standard readability formulas. Mean readability levels of ACOG's pamphlets were between grade 7.0 to grade 9.3, depending on the formula used. Analysis of readability over the 10 years showed a trend toward lower readability levels. Analysis by category of pamphlet found that the lowest readability levels were in "Especially for teens" pamphlets. Our data suggested that most of ACOG's patient education pamphlets currently available are written at a higher readability level than recommended for the general public. The readability of those pamphlets improved in the 10 years since the organization published its first pamphlet, but the goal of sixth-grade readability level has not been reached.

  11. Mining the Galaxy Zoo Database: Machine Learning Applications

    NASA Astrophysics Data System (ADS)

    Borne, Kirk D.; Wallin, J.; Vedachalam, A.; Baehr, S.; Lintott, C.; Darg, D.; Smith, A.; Fortson, L.

    2010-01-01

    The new Zooniverse initiative is addressing the data flood in the sciences through a transformative partnership between professional scientists, volunteer citizen scientists, and machines. As part of this project, we are exploring the application of machine learning techniques to data mining problems associated with the large and growing database of volunteer science results gathered by the Galaxy Zoo citizen science project. We will describe the basic challenge, some machine learning approaches, and early results. One of the motivators for this study is the acquisition (through the Galaxy Zoo results database) of approximately 100 million classification labels for roughly one million galaxies, yielding a tremendously large and rich set of training examples for improving automated galaxy morphological classification algorithms. In our first case study, the goal is to learn which morphological and photometric features in the Sloan Digital Sky Survey (SDSS) database correlate most strongly with user-selected galaxy morphological class. As a corollary to this study, we are also aiming to identify which galaxy parameters in the SDSS database correspond to galaxies that have been the most difficult to classify (based upon large dispersion in their volunter-provided classifications). Our second case study will focus on similar data mining analyses and machine leaning algorithms applied to the Galaxy Zoo catalog of merging and interacting galaxies. The outcomes of this project will have applications in future large sky surveys, such as the LSST (Large Synoptic Survey Telescope) project, which will generate a catalog of 20 billion galaxies and will produce an additional astronomical alert database of approximately 100 thousand events each night for 10 years -- the capabilities and algorithms that we are exploring will assist in the rapid characterization and classification of such massive data streams. This research has been supported in part through NSF award #0941610.

  12. Social Science Data Archives and Libraries: A View to the Future.

    ERIC Educational Resources Information Center

    Clark, Barton M.

    1982-01-01

    Discusses factors militating against integration of social science data archives and libraries in near future, noting usage of materials, access requisite skills of librarians, economic stability of archives, existing structures which manage social science data archives. Role of librarians, data access tools, and cataloging of machine-readable…

  13. LIBRARY INFORMATION PROCESSING USING AN ON-LINE, REAL-TIME COMPUTER SYSTEM.

    ERIC Educational Resources Information Center

    HOLZBAUR, FREDERICK W.; FARRIS, EUGENE H.

    DIRECT MAN-MACHINE COMMUNICATION IS NOW POSSIBLE THROUGH ON-LINE, REAL-TIME TYPEWRITER TERMINALS DIRECTLY CONNECTED TO COMPUTERS. THESE TERMINAL SYSTEMS PERMIT THE OPERATOR, WHETHER ORDER CLERK, CATALOGER, REFERENCE LIBRARIAN OR TYPIST, TO INTERACT WITH THE COMPUTER IN MANIPULATING DATA STORED WITHIN IT. THE IBM ADMINISTRATIVE TERMINAL SYSTEM…

  14. Modeling Stochastic Kinetics of Molecular Machines at Multiple Levels: From Molecules to Modules

    PubMed Central

    Chowdhury, Debashish

    2013-01-01

    A molecular machine is either a single macromolecule or a macromolecular complex. In spite of the striking superficial similarities between these natural nanomachines and their man-made macroscopic counterparts, there are crucial differences. Molecular machines in a living cell operate stochastically in an isothermal environment far from thermodynamic equilibrium. In this mini-review we present a catalog of the molecular machines and an inventory of the essential toolbox for theoretically modeling these machines. The tool kits include 1), nonequilibrium statistical-physics techniques for modeling machines and machine-driven processes; and 2), statistical-inference methods for reverse engineering a functional machine from the empirical data. The cell is often likened to a microfactory in which the machineries are organized in modular fashion; each module consists of strongly coupled multiple machines, but different modules interact weakly with each other. This microfactory has its own automated supply chain and delivery system. Buoyed by the success achieved in modeling individual molecular machines, we advocate integration of these models in the near future to develop models of functional modules. A system-level description of the cell from the perspective of molecular machinery (the mechanome) is likely to emerge from further integrations that we envisage here. PMID:23746505

  15. CANFAR + Skytree: Mining Massive Datasets as an Essential Part of the Future of Astronomy

    NASA Astrophysics Data System (ADS)

    Ball, Nicholas M.

    2013-01-01

    The future study of large astronomical datasets, consisting of hundreds of millions to billions of objects, will be dominated by large computing resources, and by analysis tools of the necessary scalability and sophistication to extract useful information. Significant effort will be required to fulfil their potential as a provider of the next generation of science results. To-date, computing systems have allowed either sophisticated analysis of small datasets, e.g., most astronomy software, or simple analysis of large datasets, e.g., database queries. At the Canadian Astronomy Data Centre, we have combined our cloud computing system, the Canadian Advanced Network for Astronomical Research (CANFAR), with the world's most advanced machine learning software, Skytree, to create the world's first cloud computing system for data mining in astronomy. This allows the full sophistication of the huge fields of data mining and machine learning to be applied to the hundreds of millions of objects that make up current large datasets. CANFAR works by utilizing virtual machines, which appear to the user as equivalent to a desktop. Each machine is replicated as desired to perform large-scale parallel processing. Such an arrangement carries far more flexibility than other cloud systems, because it enables the user to immediately install and run the same code that they already utilize for science on their desktop. We demonstrate the utility of the CANFAR + Skytree system by showing science results obtained, including assigning photometric redshifts with full probability density functions (PDFs) to a catalog of approximately 133 million galaxies from the MegaPipe reductions of the Canada-France-Hawaii Telescope Legacy Wide and Deep surveys. Each PDF is produced nonparametrically from 100 instances of the photometric parameters for each galaxy, generated by perturbing within the errors on the measurements. Hence, we produce, store, and assign redshifts to, a catalog of over 13 billion object instances. This catalog is comparable in size to those expected from next-generation surveys, such as Large Synoptic Survey Telescope. The CANFAR+Skytree system is open for use by any interested member of the astronomical community.

  16. Advanced, Analytic, Automated (AAA) Measurement of Engagement During Learning

    PubMed Central

    D’Mello, Sidney; Dieterle, Ed; Duckworth, Angela

    2017-01-01

    It is generally acknowledged that engagement plays a critical role in learning. Unfortunately, the study of engagement has been stymied by a lack of valid and efficient measures. We introduce the advanced, analytic, and automated (AAA) approach to measure engagement at fine-grained temporal resolutions. The AAA measurement approach is grounded in embodied theories of cognition and affect, which advocate a close coupling between thought and action. It uses machine-learned computational models to automatically infer mental states associated with engagement (e.g., interest, flow) from machine-readable behavioral and physiological signals (e.g., facial expressions, eye tracking, click-stream data) and from aspects of the environmental context. We present15 case studies that illustrate the potential of the AAA approach for measuring engagement in digital learning environments. We discuss strengths and weaknesses of the AAA approach, concluding that it has significant promise to catalyze engagement research. PMID:29038607

  17. Advanced, Analytic, Automated (AAA) Measurement of Engagement During Learning.

    PubMed

    D'Mello, Sidney; Dieterle, Ed; Duckworth, Angela

    2017-01-01

    It is generally acknowledged that engagement plays a critical role in learning. Unfortunately, the study of engagement has been stymied by a lack of valid and efficient measures. We introduce the advanced, analytic, and automated (AAA) approach to measure engagement at fine-grained temporal resolutions. The AAA measurement approach is grounded in embodied theories of cognition and affect, which advocate a close coupling between thought and action. It uses machine-learned computational models to automatically infer mental states associated with engagement (e.g., interest, flow) from machine-readable behavioral and physiological signals (e.g., facial expressions, eye tracking, click-stream data) and from aspects of the environmental context. We present15 case studies that illustrate the potential of the AAA approach for measuring engagement in digital learning environments. We discuss strengths and weaknesses of the AAA approach, concluding that it has significant promise to catalyze engagement research.

  18. Computer Information Project for Monographs at the Medical Research Library of Brooklyn

    PubMed Central

    Koch, Michael S.; Kovacs, Helen

    1973-01-01

    The article describes a resource library's computer-based project that provides cataloging and other bibliographic services and promotes greater use of the book collection. A few studies are cited to show the significance of monographic literature in medical libraries. The educational role of the Medical Research Library of Brooklyn is discussed, both with regard to the parent institution and to smaller medical libraries in the same geographic area. Types of aid given to smaller libraries are enumerated. Information is given on methods for providing machine-produced catalog cards, current awareness notes, and bibliographic lists. Actualities and potentialities of the computer project are discussed. PMID:4579767

  19. Identification of stars and digital version of the catalogue of 1958 by Brodskaya and Shajn

    NASA Astrophysics Data System (ADS)

    Gorbunov, M. A.; Shlyapnikov, A. A.

    2017-12-01

    The following topics are considered: the identification of objects on search maps, the determination of their coordinates at the epoch of 2000, and converting the published version of the catalogue of 1958 by Brodskaya and Shajn into a machine-readable format. The statistics for photometric and spectral data from the original catalogue is presented. A digital version of the catalogue is described, as well as its presentation in HTML, VOTable and AJS formats and the basic principles of work in the interactive application of International Virtual Observatory - the Aladin Sky Atlas.

  20. Residential Energy Consumption Survey (RECS): National interim energy consumption survey, 1978-1979, household monthly energy consumption and expenditures. Shopper's guide

    NASA Astrophysics Data System (ADS)

    Windell, P.

    1981-08-01

    The data from the National Interim Energy Consumption Survey (NIECS) is available to the public on machine readable magnetic tapes. Brief overviews of the Residential Energy Consumption Survey as a whole and of the NIECS in particular is a brief description of each of the files included in this tape, and a list of the variables in the data set are included. A copy of the fuel supplier record form used to collect consumption and expenditure data for each of the sample households is also included.

  1. Introducing meta-services for biomedical information extraction

    PubMed Central

    Leitner, Florian; Krallinger, Martin; Rodriguez-Penagos, Carlos; Hakenberg, Jörg; Plake, Conrad; Kuo, Cheng-Ju; Hsu, Chun-Nan; Tsai, Richard Tzong-Han; Hung, Hsi-Chuan; Lau, William W; Johnson, Calvin A; Sætre, Rune; Yoshida, Kazuhiro; Chen, Yan Hua; Kim, Sun; Shin, Soo-Yong; Zhang, Byoung-Tak; Baumgartner, William A; Hunter, Lawrence; Haddow, Barry; Matthews, Michael; Wang, Xinglong; Ruch, Patrick; Ehrler, Frédéric; Özgür, Arzucan; Erkan, Güneş; Radev, Dragomir R; Krauthammer, Michael; Luong, ThaiBinh; Hoffmann, Robert; Sander, Chris; Valencia, Alfonso

    2008-01-01

    We introduce the first meta-service for information extraction in molecular biology, the BioCreative MetaServer (BCMS; ). This prototype platform is a joint effort of 13 research groups and provides automatically generated annotations for PubMed/Medline abstracts. Annotation types cover gene names, gene IDs, species, and protein-protein interactions. The annotations are distributed by the meta-server in both human and machine readable formats (HTML/XML). This service is intended to be used by biomedical researchers and database annotators, and in biomedical language processing. The platform allows direct comparison, unified access, and result aggregation of the annotations. PMID:18834497

  2. NASA Thesaurus Data File

    NASA Technical Reports Server (NTRS)

    2012-01-01

    The NASA Thesaurus contains the authorized NASA subject terms used to index and retrieve materials in the NASA Aeronautics and Space Database (NA&SD) and NASA Technical Reports Server (NTRS). The scope of this controlled vocabulary includes not only aerospace engineering, but all supporting areas of engineering and physics, the natural space sciences (astronomy, astrophysics, planetary science), Earth sciences, and the biological sciences. The NASA Thesaurus Data File contains all valid terms and hierarchical relationships, USE references, and related terms in machine-readable form. The Data File is available in the following formats: RDF/SKOS, RDF/OWL, ZThes-1.0, and CSV/TXT.

  3. DisGeNET-RDF: harnessing the innovative power of the Semantic Web to explore the genetic basis of diseases.

    PubMed

    Queralt-Rosinach, Núria; Piñero, Janet; Bravo, Àlex; Sanz, Ferran; Furlong, Laura I

    2016-07-15

    DisGeNET-RDF makes available knowledge on the genetic basis of human diseases in the Semantic Web. Gene-disease associations (GDAs) and their provenance metadata are published as human-readable and machine-processable web resources. The information on GDAs included in DisGeNET-RDF is interlinked to other biomedical databases to support the development of bioinformatics approaches for translational research through evidence-based exploitation of a rich and fully interconnected linked open data. http://rdf.disgenet.org/ support@disgenet.org. © The Author 2016. Published by Oxford University Press.

  4. How semantics can inform the geological mapping process and support intelligent queries

    NASA Astrophysics Data System (ADS)

    Lombardo, Vincenzo; Piana, Fabrizio; Mimmo, Dario

    2017-04-01

    The geologic mapping process requires the organization of data according to the general knowledge about the objects, namely the geologic units, and to the objectives of a graphic representation of such objects in a map, following an established model of geotectonic evolution. Semantics can greatly help such a process in two concerns: the provision of a terminological base to name and classify the objects of the map; on the other, the implementation of a machine-readable encoding of the geologic knowledge base supports the application of reasoning mechanisms and the derivation of novel properties and relations about the objects of the map. The OntoGeonous initiative has built a terminological base of geological knowledge in a machine-readable format, following the Semantic Web tenets and the Linked Data paradigm. The major knowledge sources of the OntoGeonous initiative are GeoScience Markup Language schemata and vocabularies (through its last version, GeoSciML 4, 2015, published by the IUGS CGI Commission) and the INSPIRE "Data Specification on Geology" directives (an operative simplification of GeoSciML, published by INSPIRE Thematic Working Group Geology of the European Commission). The Linked Data paradigm has been exploited by linking (without replicating, to avoid inconsistencies) the already existing machine-readable encoding for some specific domains, such as the lithology domain (vocabulary Simple Lithology) and the geochronologic time scale (ontology "gts"). Finally, for the upper level knowledge, shared across several geologic domains, we have resorted to NASA SWEET ontology. The OntoGeonous initiative has also produced a wiki that explains how the geologic knowledge has been encoded from shared geoscience vocabularies (https://www.di.unito.it/wikigeo/). In particular, the sections dedicated to axiomatization will support the construction of an appropriate data base schema that can be then filled with the objects of the map. This contribution will discuss how the formal encoding of the geological knowledge opens new perspectives for the analysis and representation of the geological systems. In fact, once that the major concepts are defined, the resulting formal conceptual model of the geologic system can hold across different technical and scientific communities. Furthermore, this would allow for a semi-automatic or automatic classification of the cartographic database, where a significant number of properties (attributes) of the recorded instances could be inferred through computational reasoning. So, for example, the system can be queried for showing the instances that satisfy some property (e.g., "Retrieve all the lithostratigraphic units composed of clastic sedimentary rock") or for classifying some unit according to the properties holding for that unit (e.g., "What is the class of the geologic unit composed of siltstone material?").

  5. The Readability of an Unreadable Text.

    ERIC Educational Resources Information Center

    Gordon, Robert M.

    1980-01-01

    The Dale-Chall Readability Formula and the Fry Readability Graph were used to analyze passages of Plato's "Parmenides," a notoriously difficult literary piece. The readability levels of the text ranged from fourth to eighth grade (Dale-Chall) and from sixth to tenth grade (Fry), indicating the limitations of the readability tests. (DF)

  6. Integrating a local database into the StarView distributed user interface

    NASA Technical Reports Server (NTRS)

    Silberberg, D. P.

    1992-01-01

    A distributed user interface to the Space Telescope Data Archive and Distribution Service (DADS) known as StarView is being developed. The DADS architecture consists of the data archive as well as a relational database catalog describing the archive. StarView is a client/server system in which the user interface is the front-end client to the DADS catalog and archive servers. Users query the DADS catalog from the StarView interface. Query commands are transmitted via a network and evaluated by the database. The results are returned via the network and are displayed on StarView forms. Based on the results, users decide which data sets to retrieve from the DADS archive. Archive requests are packaged by StarView and sent to DADS, which returns the requested data sets to the users. The advantages of distributed client/server user interfaces over traditional one-machine systems are well known. Since users run software on machines separate from the database, the overall client response time is much faster. Also, since the server is free to process only database requests, the database response time is much faster. Disadvantages inherent in this architecture are slow overall database access time due to the network delays, lack of a 'get previous row' command, and that refinements of a previously issued query must be submitted to the database server, even though the domain of values have already been returned by the previous query. This architecture also does not allow users to cross correlate DADS catalog data with other catalogs. Clearly, a distributed user interface would be more powerful if it overcame these disadvantages. A local database is being integrated into StarView to overcome these disadvantages. When a query is made through a StarView form, which is often composed of fields from multiple tables, it is translated to an SQL query and issued to the DADS catalog. At the same time, a local database table is created to contain the resulting rows of the query. The returned rows are displayed on the form as well as inserted into the local database table. Identical results are produced by reissuing the query to either the DADS catalog or to the local table. Relational databases do not provide a 'get previous row' function because of the inherent complexity of retrieving previous rows of multiple-table joins. However, since this function is easily implemented on a single table, StarView uses the local table to retrieve the previous row. Also, StarView issues subsequent query refinements to the local table instead of the DADS catalog, eliminating the network transmission overhead. Finally, other catalogs can be imported into the local database for cross correlation with local tables. Overall, it is believe that this is a more powerful architecture for distributed, database user interfaces.

  7. A Manual for Readable Writing.

    ERIC Educational Resources Information Center

    Klare, George R.

    One of the ways to handle the increasing demands on readers' skills is to make writing more readable. The problem has two different aspects: predicting how readable writing will be to a reader, and producing writing that is readable to that reader. Prediction is relatively simple, and can be done statistically with readability formulas. Production…

  8. Break-Even Point for a Proof Slip Operation

    ERIC Educational Resources Information Center

    Anderson, James F.

    1972-01-01

    Break-even analysis is applied to determine what magnitude of titles added per year is sufficient to utilize economically Library of Congress proof slips and a Xerox 914 copying machine in the cataloging operation of a library. A formula is derived, and an example of its use is given. (1 reference) (Author/SJ)

  9. Educational Resources in the ASCC Library

    ERIC Educational Resources Information Center

    Lin, Steven

    2006-01-01

    After two years of construction, American Samoa Community College opened its new library on September 2, 2003. The library is located on the east side of campus and is equipped with ten computer workstations, four online public access catalogs, three copying machines, and an elevator that is in compliance with the Americans with Disabilities Act.…

  10. GALAXY ZOO MORPHOLOGY AND PHOTOMETRIC REDSHIFTS IN THE SLOAN DIGITAL SKY SURVEY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Way, M. J.

    It has recently been demonstrated that one can accurately derive galaxy morphology from particular primary and secondary isophotal shape estimates in the Sloan Digital Sky Survey (SDSS) imaging catalog. This was accomplished by applying Machine Learning techniques to the Galaxy Zoo morphology catalog. Using the broad bandpass photometry of the SDSS in combination with precise knowledge of galaxy morphology should help in estimating more accurate photometric redshifts for galaxies. Using the Galaxy Zoo separation for spirals and ellipticals in combination with SDSS photometry we attempt to calculate photometric redshifts. In the best case we find that the root-mean-square error formore » luminous red galaxies classified as ellipticals is as low as 0.0118. Given these promising results we believe better photometric redshift estimates for all galaxies in the SDSS ({approx}350 million) will be feasible if researchers can also leverage their derived morphologies via Machine Learning. These initial results look to be promising for those interested in estimating weak lensing, baryonic acoustic oscillation, and other fields dependent upon accurate photometric redshifts.« less

  11. Readability, content, and quality of online patient education materials on preeclampsia.

    PubMed

    Lange, Elizabeth M S; Shah, Anuj M; Braithwaite, Brian A; You, Whitney B; Wong, Cynthia A; Grobman, William A; Toledo, Paloma

    2015-01-01

    The objective of this study was to evaluate the readability, content, and quality of patient education materials addressing preeclampsia. Websites of U.S. obstetrics and gynecology residency programs were searched for patient education materials. Readability, content, and quality were assessed. A one-sample t-test was used to evaluate mean readability level compared with the recommended 6th grade reading level. Mean readability levels were higher using all indices (p < 0.001). Content was variable with good website understandability, but poor actionability. The mean readability was above the recommended 6th grade reading level. The content, readability, and actionability of preeclampsia patient education materials should be improved.

  12. Machines first, humans second: on the importance of algorithmic interpretation of open chemistry data.

    PubMed

    Clark, Alex M; Williams, Antony J; Ekins, Sean

    2015-01-01

    The current rise in the use of open lab notebook techniques means that there are an increasing number of scientists who make chemical information freely and openly available to the entire community as a series of micropublications that are released shortly after the conclusion of each experiment. We propose that this trend be accompanied by a thorough examination of data sharing priorities. We argue that the most significant immediate benefactor of open data is in fact chemical algorithms, which are capable of absorbing vast quantities of data, and using it to present concise insights to working chemists, on a scale that could not be achieved by traditional publication methods. Making this goal practically achievable will require a paradigm shift in the way individual scientists translate their data into digital form, since most contemporary methods of data entry are designed for presentation to humans rather than consumption by machine learning algorithms. We discuss some of the complex issues involved in fixing current methods, as well as some of the immediate benefits that can be gained when open data is published correctly using unambiguous machine readable formats. Graphical AbstractLab notebook entries must target both visualisation by scientists and use by machine learning algorithms.

  13. Towards a semantic web of paleoclimatology

    NASA Astrophysics Data System (ADS)

    Emile-Geay, J.; Eshleman, J. A.

    2012-12-01

    The paleoclimate record is information-rich, yet signifiant technical barriers currently exist before it can be used to automatically answer scientific questions. Here we make the case for a universal format to structure paleoclimate data. A simple example demonstrates the scientific utility of such a self-contained way of organizing coral data and meta-data in the Matlab language. This example is generalized to a universal ontology that may form the backbone of an open-source, open-access and crowd-sourced paleoclimate database. Its key attributes are: 1. Parsability: the format is self-contained (hence machine-readable), and would therefore enable a semantic web of paleoclimate information. 2. Universality: the format is platform-independent (readable on all computer and operating systems), and language- independent (readable in major programming languages) 3. Extensibility: the format requires a minimum set of fields to appropriately define a paleoclimate record, but allows for the database to grow organically as more records are added, or - equally important - as more metadata are added to existing records. 4. Citability: The format enables the automatic citation of peer- reviewed articles as well as data citations whenever a data record is being used for analysis, making due recognition of scientific work an automatic part and foundational principle of paleoclimate data analysis. 5. Ergonomy: The format will be easy to use, update and manage. This structure is designed to enable semantic searches, and is expected to help accelerate discovery in all workflows where paleoclimate data are being used. Practical steps towards the implementation of such a system at the community level are then discussed.; Preliminary ontology describing relationships between the data and meta-data fields of the Nurhati et al. [2011] climate record. Several fields are viewed as instances of larger classes (ProxyClass,Site,Reference), which would allow computers to perform operations on all records within a specific class (e.g. if the measurement type is δ18O , or if the proxy class is 'Tree Ring Width', or if the resolution is less than 3 months, etc). All records in such a database would be bound to each other by similar links, allowing machines to automatically process any form of query involving existing information. Such a design would also allow growth, by adding records and/or additional information about each record.

  14. Semantic Metadata for Heterogeneous Spatial Planning Documents

    NASA Astrophysics Data System (ADS)

    Iwaniak, A.; Kaczmarek, I.; Łukowicz, J.; Strzelecki, M.; Coetzee, S.; Paluszyński, W.

    2016-09-01

    Spatial planning documents contain information about the principles and rights of land use in different zones of a local authority. They are the basis for administrative decision making in support of sustainable development. In Poland these documents are published on the Web according to a prescribed non-extendable XML schema, designed for optimum presentation to humans in HTML web pages. There is no document standard, and limited functionality exists for adding references to external resources. The text in these documents is discoverable and searchable by general-purpose web search engines, but the semantics of the content cannot be discovered or queried. The spatial information in these documents is geographically referenced but not machine-readable. Major manual efforts are required to integrate such heterogeneous spatial planning documents from various local authorities for analysis, scenario planning and decision support. This article presents results of an implementation using machine-readable semantic metadata to identify relationships among regulations in the text, spatial objects in the drawings and links to external resources. A spatial planning ontology was used to annotate different sections of spatial planning documents with semantic metadata in the Resource Description Framework in Attributes (RDFa). The semantic interpretation of the content, links between document elements and links to external resources were embedded in XHTML pages. An example and use case from the spatial planning domain in Poland is presented to evaluate its efficiency and applicability. The solution enables the automated integration of spatial planning documents from multiple local authorities to assist decision makers with understanding and interpreting spatial planning information. The approach is equally applicable to legal documents from other countries and domains, such as cultural heritage and environmental management.

  15. Generation of open biomedical datasets through ontology-driven transformation and integration processes.

    PubMed

    Carmen Legaz-García, María Del; Miñarro-Giménez, José Antonio; Menárguez-Tortosa, Marcos; Fernández-Breis, Jesualdo Tomás

    2016-06-03

    Biomedical research usually requires combining large volumes of data from multiple heterogeneous sources, which makes difficult the integrated exploitation of such data. The Semantic Web paradigm offers a natural technological space for data integration and exploitation by generating content readable by machines. Linked Open Data is a Semantic Web initiative that promotes the publication and sharing of data in machine readable semantic formats. We present an approach for the transformation and integration of heterogeneous biomedical data with the objective of generating open biomedical datasets in Semantic Web formats. The transformation of the data is based on the mappings between the entities of the data schema and the ontological infrastructure that provides the meaning to the content. Our approach permits different types of mappings and includes the possibility of defining complex transformation patterns. Once the mappings are defined, they can be automatically applied to datasets to generate logically consistent content and the mappings can be reused in further transformation processes. The results of our research are (1) a common transformation and integration process for heterogeneous biomedical data; (2) the application of Linked Open Data principles to generate interoperable, open, biomedical datasets; (3) a software tool, called SWIT, that implements the approach. In this paper we also describe how we have applied SWIT in different biomedical scenarios and some lessons learned. We have presented an approach that is able to generate open biomedical repositories in Semantic Web formats. SWIT is able to apply the Linked Open Data principles in the generation of the datasets, so allowing for linking their content to external repositories and creating linked open datasets. SWIT datasets may contain data from multiple sources and schemas, thus becoming integrated datasets.

  16. LiPD and CSciBox: A Case Study in Why Data Standards are Important for Paleoscience

    NASA Astrophysics Data System (ADS)

    Weiss, I.; Bradley, E.; McKay, N.; Emile-Geay, J.; de Vesine, L. R.; Anderson, K. A.; White, J. W. C.; Marchitto, T. M., Jr.

    2016-12-01

    CSciBox [1] is an integrated software system that helps geoscientists build and evaluate age models. Its user chooses from a number of built-in analysis tools, composing them into an analysis workflow and applying it to paleoclimate proxy datasets. CSciBox employs modern database technology to store both the data and the analysis results in an easily accessible and searchable form, and offers the user access to the computational toolbox, the data, and the results via a graphical user interface and a sophisticated plotter. Standards are a staple of modern life, and underlie any form of automation. Without data standards, it is difficult, if not impossible, to construct effective computer tools for paleoscience analysis. The LiPD (Linked Paleo Data) framework [2] enables the storage of both data and metadata in systematic, meaningful, machine-readable ways. LiPD has been a primary enabler of CSciBox's goals of usability, interoperability, and reproducibility. Building LiPD capabilities into CSciBox's importer, for instance, eliminated the need to ask the user about file formats, variable names, relationships between columns in the input file, etc. Building LiPD capabilities into the exporter facilitated the storage of complete details about the input data-provenance, preprocessing steps, etc.-as well as full descriptions of any analyses that were performed using the CSciBox tool, along with citations to appropriate references. This comprehensive collection of data and metadata, which is all linked together in a semantically meaningful, machine-readable way, not only completely documents the analyses and makes them reproducible. It also enables interoperability with any other software system that employs the LiPD standard. [1] www.cs.colorado.edu/ lizb/cscience.html[2] McKay & Emile-Geay, Climate of the Past 12:1093 (2016)

  17. Rosetta: Ensuring the Preservation and Usability of ASCII-based Data into the Future

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M. K.; Arms, S. C.

    2015-12-01

    Field data obtained from dataloggers often take the form of comma separated value (CSV) ASCII text files. While ASCII based data formats have positive aspects, such as the ease of accessing the data from disk and the wide variety of tools available for data analysis, there are some drawbacks, especially when viewing the situation through the lens of data interoperability and stewardship. The Unidata data translation tool, Rosetta, is a web-based service that provides an easy, wizard-based interface for data collectors to transform their datalogger generated ASCII output into Climate and Forecast (CF) compliant netCDF files following the CF-1.6 discrete sampling geometries. These files are complete with metadata describing what data are contained in the file, the instruments used to collect the data, and other critical information that otherwise may be lost in one of many README files. The choice of the machine readable netCDF data format and data model, coupled with the CF conventions, ensures long-term preservation and interoperability, and that future users will have enough information to responsibly use the data. However, with the understanding that the observational community appreciates the ease of use of ASCII files, methods for transforming the netCDF back into a CSV or spreadsheet format are also built-in. One benefit of translating ASCII data into a machine readable format that follows open community-driven standards is that they are instantly able to take advantage of data services provided by the many open-source data server tools, such as the THREDDS Data Server (TDS). While Rosetta is currently a stand-alone service, this talk will also highlight efforts to couple Rosetta with the TDS, thus allowing self-publishing of thoroughly documented datasets by the data producers themselves.

  18. Towards Web 3.0: taxonomies and ontologies for medical education -- a systematic review.

    PubMed

    Blaum, Wolf E; Jarczweski, Anne; Balzer, Felix; Stötzner, Philip; Ahlers, Olaf

    2013-01-01

    Both for curricular development and mapping, as well as for orientation within the mounting supply of learning resources in medical education, the Semantic Web ("Web 3.0") poses a low-threshold, effective tool that enables identification of content related items across system boundaries. Replacement of the currently required manual with an automatically generated link, which is based on content and semantics, requires the use of a suitably structured vocabulary for a machine-readable description of object content. Aim of this study is to compile the existing taxonomies and ontologies used for the annotation of medical content and learning resources, to compare those using selected criteria, and to verify their suitability in the context described above. Based on a systematic literature search, existing taxonomies and ontologies for the description of medical learning resources were identified. Through web searches and/or direct contact with the respective editors, each of the structured vocabularies thus identified were examined in regards to topic, structure, language, scope, maintenance, and technology of the taxonomy/ontology. In addition, suitability for use in the Semantic Web was verified. Among 20 identified publications, 14 structured vocabularies were identified, which differed rather strongly in regards to language, scope, currency, and maintenance. None of the identified vocabularies fulfilled the necessary criteria for content description of medical curricula and learning resources in the German-speaking world. While moving towards Web 3.0, a significant problem lies in the selection and use of an appropriate German vocabulary for the machine-readable description of object content. Possible solutions include development, translation and/or combination of existing vocabularies, possibly including partial translations of English vocabularies.

  19. LEMS: a language for expressing complex biological models in concise and hierarchical form and its use in underpinning NeuroML 2.

    PubMed

    Cannon, Robert C; Gleeson, Padraig; Crook, Sharon; Ganapathy, Gautham; Marin, Boris; Piasini, Eugenio; Silver, R Angus

    2014-01-01

    Computational models are increasingly important for studying complex neurophysiological systems. As scientific tools, it is essential that such models can be reproduced and critically evaluated by a range of scientists. However, published models are currently implemented using a diverse set of modeling approaches, simulation tools, and computer languages making them inaccessible and difficult to reproduce. Models also typically contain concepts that are tightly linked to domain-specific simulators, or depend on knowledge that is described exclusively in text-based documentation. To address these issues we have developed a compact, hierarchical, XML-based language called LEMS (Low Entropy Model Specification), that can define the structure and dynamics of a wide range of biological models in a fully machine readable format. We describe how LEMS underpins the latest version of NeuroML and show that this framework can define models of ion channels, synapses, neurons and networks. Unit handling, often a source of error when reusing models, is built into the core of the language by specifying physical quantities in models in terms of the base dimensions. We show how LEMS, together with the open source Java and Python based libraries we have developed, facilitates the generation of scripts for multiple neuronal simulators and provides a route for simulator free code generation. We establish that LEMS can be used to define models from systems biology and map them to neuroscience-domain specific simulators, enabling models to be shared between these traditionally separate disciplines. LEMS and NeuroML 2 provide a new, comprehensive framework for defining computational models of neuronal and other biological systems in a machine readable format, making them more reproducible and increasing the transparency and accessibility of their underlying structure and properties.

  20. LEMS: a language for expressing complex biological models in concise and hierarchical form and its use in underpinning NeuroML 2

    PubMed Central

    Cannon, Robert C.; Gleeson, Padraig; Crook, Sharon; Ganapathy, Gautham; Marin, Boris; Piasini, Eugenio; Silver, R. Angus

    2014-01-01

    Computational models are increasingly important for studying complex neurophysiological systems. As scientific tools, it is essential that such models can be reproduced and critically evaluated by a range of scientists. However, published models are currently implemented using a diverse set of modeling approaches, simulation tools, and computer languages making them inaccessible and difficult to reproduce. Models also typically contain concepts that are tightly linked to domain-specific simulators, or depend on knowledge that is described exclusively in text-based documentation. To address these issues we have developed a compact, hierarchical, XML-based language called LEMS (Low Entropy Model Specification), that can define the structure and dynamics of a wide range of biological models in a fully machine readable format. We describe how LEMS underpins the latest version of NeuroML and show that this framework can define models of ion channels, synapses, neurons and networks. Unit handling, often a source of error when reusing models, is built into the core of the language by specifying physical quantities in models in terms of the base dimensions. We show how LEMS, together with the open source Java and Python based libraries we have developed, facilitates the generation of scripts for multiple neuronal simulators and provides a route for simulator free code generation. We establish that LEMS can be used to define models from systems biology and map them to neuroscience-domain specific simulators, enabling models to be shared between these traditionally separate disciplines. LEMS and NeuroML 2 provide a new, comprehensive framework for defining computational models of neuronal and other biological systems in a machine readable format, making them more reproducible and increasing the transparency and accessibility of their underlying structure and properties. PMID:25309419

  1. Color pictorial serpentine halftone for secure embedded data

    NASA Astrophysics Data System (ADS)

    Curry, Douglas N.

    1998-04-01

    This paper introduces a new rotatable glyph shape for trusted printing applications that has excellent image rendering, data storage and counterfeit deterrence properties. Referred to as a serpentine because it tiles into a meandering line screen, it can produce high quality images independent of its ability to embed data. The hafltone cell is constructed with hyperbolic curves to enhance its dynamic range, and generates low distortion because of rotational tone invariance with its neighbors. An extension to the process allows the data to be formatted into human readable text patterns, viewable with a magnifying glass, and therefore not requiring input scanning. The resultant embedded halftone patterns can be recognized as simple numbers (0 - 9) or alphanumerics (a - z). The pattern intensity can be offset from the surrounding image field intensity, producing a watermarking effect. We have been able to embed words such as 'original' or license numbers into the background halftone pattern of images which can be readily observed in the original image, and which conveniently disappear upon copying. We have also embedded data blocks with self-clocking codes and error correction data which are machine-readable. Finally, we have successfully printed full color images with both the embedded data and text, simulating a trusted printing application.

  2. A Web-Based Information System for Field Data Management

    NASA Astrophysics Data System (ADS)

    Weng, Y. H.; Sun, F. S.

    2014-12-01

    A web-based field data management system has been designed and developed to allow field geologists to store, organize, manage, and share field data online. System requirements were analyzed and clearly defined first regarding what data are to be stored, who the potential users are, and what system functions are needed in order to deliver the right data in the right way to the right user. A 3-tiered architecture was adopted to create this secure, scalable system that consists of a web browser at the front end while a database at the back end and a functional logic server in the middle. Specifically, HTML, CSS, and JavaScript were used to implement the user interface in the front-end tier, the Apache web server runs PHP scripts, and MySQL to server is used for the back-end database. The system accepts various types of field information, including image, audio, video, numeric, and text. It allows users to select data and populate them on either Google Earth or Google Maps for the examination of the spatial relations. It also makes the sharing of field data easy by converting them into XML format that is both human-readable and machine-readable, and thus ready for reuse.

  3. Readability of HIV/AIDS educational materials: the role of the medium of communication, target audience, and producer characteristics.

    PubMed

    Wells, J A

    1994-12-01

    The reading difficulty of many HIV/AIDS brochures and pamphlets limits their effectiveness. This analysis addresses correlates of readability in 136 HIV/AIDS educational items. Readability is measured using the SMOG Index. The medium of communication is significantly related to readability: comic books and brochures are, on average, more readable than books and pamphlets (10.9 versus 11.9). The target audience also differentiates readability. Materials for HIV antibody test seekers, the general community, and sexually active adults have a more difficult reading grade, averaging 12.1, whereas materials for ethnic minorities average a more readable 9.2. The producer organization's type and location are unrelated to readability, but an AIDS-specific organizational focus correlates with better readability (grade 10.8 vs. 11.8). These findings remain significant in multivariate analysis. The results indicate that brochures and comics are more likely to be comprehended by low-literacy populations, that an understanding of the literacy of target audiences is needed to produce materials with appropriate reading levels, and that policies to influence producer organizations may result in the creation of more readable materials.

  4. Tiny, Dusty, Galactic HI Clouds: The GALFA-HI Compact Cloud Catalog

    NASA Astrophysics Data System (ADS)

    Saul, Destry R.; Putman, M. E.; Peek, J. G.

    2013-01-01

    The recently published GALFA-HI Compact Cloud Catalog contains 2000 nearby neutral hydrogen clouds under 20' in angular size detected with a machine-vision algorithm in the Galactic Arecibo L-Band Feed Array HI survey (GALFA-HI). At a distance of 1kpc, the compact clouds would typically be 1 solar mass and 1pc in size. We observe that nearly all of the compact clouds that are classified as high velocity (> 90 km/s) are near previously-identified high velocity complexes. We separate the compact clouds into populations based on velocity, linewidth, and position. We have begun to search for evidence of dust in these clouds using IRIS and have detections in several populations.

  5. Modeling stochastic kinetics of molecular machines at multiple levels: from molecules to modules.

    PubMed

    Chowdhury, Debashish

    2013-06-04

    A molecular machine is either a single macromolecule or a macromolecular complex. In spite of the striking superficial similarities between these natural nanomachines and their man-made macroscopic counterparts, there are crucial differences. Molecular machines in a living cell operate stochastically in an isothermal environment far from thermodynamic equilibrium. In this mini-review we present a catalog of the molecular machines and an inventory of the essential toolbox for theoretically modeling these machines. The tool kits include 1), nonequilibrium statistical-physics techniques for modeling machines and machine-driven processes; and 2), statistical-inference methods for reverse engineering a functional machine from the empirical data. The cell is often likened to a microfactory in which the machineries are organized in modular fashion; each module consists of strongly coupled multiple machines, but different modules interact weakly with each other. This microfactory has its own automated supply chain and delivery system. Buoyed by the success achieved in modeling individual molecular machines, we advocate integration of these models in the near future to develop models of functional modules. A system-level description of the cell from the perspective of molecular machinery (the mechanome) is likely to emerge from further integrations that we envisage here. Copyright © 2013 Biophysical Society. Published by Elsevier Inc. All rights reserved.

  6. Improving readability of informed consents for research at an academic medical institution.

    PubMed

    Hadden, Kristie B; Prince, Latrina Y; Moore, Tina D; James, Laura P; Holland, Jennifer R; Trudeau, Christopher R

    2017-12-01

    The final rule for the protection of human subjects requires that informed consent be "in language understandable to the subject" and mandates that "the informed consent must be organized in such a way that facilitates comprehension." This study assessed the readability of Institutional Review Board-approved informed consent forms at our institution, implemented an intervention to improve the readability of consent forms, and measured the first year impact of the intervention. Readability assessment was conducted on a sample of 217 Institutional Review Board-approved informed consents from 2013 to 2015. A plain language informed consent template was developed and implemented and readability was assessed again after 1 year. The mean readability of the baseline sample was 10th grade. The mean readability of the post-intervention sample (n=82) was seventh grade. Providing investigators with a plain language informed consent template and training can promote improved readability of informed consents for research.

  7. Readability of Online Health Information: A Meta-Narrative Systematic Review.

    PubMed

    Daraz, Lubna; Morrow, Allison S; Ponce, Oscar J; Farah, Wigdan; Katabi, Abdulrahman; Majzoub, Abdul; Seisa, Mohamed O; Benkhadra, Raed; Alsawas, Mouaz; Larry, Prokop; Murad, M Hassan

    2018-01-01

    Online health information should meet the reading level for the general public (set at sixth-grade level). Readability is a key requirement for information to be helpful and improve quality of care. The authors conducted a systematic review to evaluate the readability of online health information in the United States and Canada. Out of 3743 references, the authors included 157 cross-sectional studies evaluating 7891 websites using 13 readability scales. The mean readability grade level across websites ranged from grade 10 to 15 based on the different scales. Stratification by specialty, health condition, and type of organization producing information revealed the same findings. In conclusion, online health information in the United States and Canada has a readability level that is inappropriate for general public use. Poor readability can lead to misinformation and may have a detrimental effect on health. Efforts are needed to improve readability and the content of online health information.

  8. Determining Readability: How to Select and Apply Easy-to-Use Readability Formulas to Assess the Difficulty of Adult Literacy Materials

    ERIC Educational Resources Information Center

    Burke, Victoria; Greenberg, Daphne

    2010-01-01

    There are many readability tools that instructors can use to help adult learners select reading materials. We describe and compare different types of readability tools: formulas calculated by hand, tools found on the Web, tools embedded in a word processing program, and readability tools found in a commercial software program. Practitioners do not…

  9. Assessing readability formula differences with written health information materials: application, results, and recommendations.

    PubMed

    Wang, Lih-Wern; Miller, Michael J; Schmitt, Michael R; Wen, Frances K

    2013-01-01

    Readability formulas are often used to guide the development and evaluation of literacy-sensitive written health information. However, readability formula results may vary considerably as a result of differences in software processing algorithms and how each formula is applied. These variations complicate interpretations of reading grade level estimates, particularly without a uniform guideline for applying and interpreting readability formulas. This research sought to (1) identify commonly used readability formulas reported in the health care literature, (2) demonstrate the use of the most commonly used readability formulas on written health information, (3) compare and contrast the differences when applying common readability formulas to identical selections of written health information, and (4) provide recommendations for choosing an appropriate readability formula for written health-related materials to optimize their use. A literature search was conducted to identify the most commonly used readability formulas in health care literature. Each of the identified formulas was subsequently applied to word samples from 15 unique examples of written health information about the topic of depression and its treatment. Readability estimates from common readability formulas were compared based on text sample size, selection, formatting, software type, and/or hand calculations. Recommendations for their use were provided. The Flesch-Kincaid formula was most commonly used (57.42%). Readability formulas demonstrated variability up to 5 reading grade levels on the same text. The Simple Measure of Gobbledygook (SMOG) readability formula performed most consistently. Depending on the text sample size, selection, formatting, software, and/or hand calculations, the individual readability formula estimated up to 6 reading grade levels of variability. The SMOG formula appears best suited for health care applications because of its consistency of results, higher level of expected comprehension, use of more recent validation criteria for determining reading grade level estimates, and simplicity of use. To improve interpretation of readability results, reporting reading grade level estimates from any formula should be accompanied with information about word sample size, location of word sampling in the text, formatting, and method of calculation. Copyright © 2013 Elsevier Inc. All rights reserved.

  10. California State Library: Processing Center Design and Specifications. Volume I, System Description and Input Processing.

    ERIC Educational Resources Information Center

    Sherman, Don; Shoffner, Ralph M.

    The scope of the California State Library-Processing Center (CSL-PC) project is to develop the design and specifications for a computerized technical processing center to provide services to a network of participating California libraries. Immediate objectives are: (1) retrospective conversion of card catalogs to a machine-form data base,…

  11. VizieR Online Data Catalog: OCSVM anomalies (Solarz+, 2017)

    NASA Astrophysics Data System (ADS)

    Solarz, A.; Bilicki, M.; Gromadzki, M.; Pollo, A.; Durkalec, A.; Wypych, M.

    2017-07-01

    One table containing 642,353 sources selected as anomalous with one-class support vector machine algorithm in AllWISE data release. Data have AllWISE photometry in W1, W2 and W3 passband and include W3 flux correction described in Krakowski et al. (2016A&A...596A..39K). (1 data file).

  12. A STORAGE AND RETRIEVAL SYSTEM FOR DOCUMENTS IN INSTRUCTIONAL RESOURCES. REPORT NO. 13.

    ERIC Educational Resources Information Center

    DIAMOND, ROBERT M.; LEE, BERTA GRATTAN

    IN ORDER TO IMPROVE INSTRUCTION WITHIN TWO-YEAR LOWER DIVISION COURSES, A COMPREHENSIVE RESOURCE LIBRARY WAS DEVELOPED AND A SIMPLIFIED CATALOGING AND INFORMATION RETRIEVAL SYSTEM WAS APPLIED TO IT. THE ROYAL MCBEE "KEYDEX" SYSTEM, CONTAINING THREE MAJOR COMPONENTS--A PUNCH MACHINE, FILE CARDS, AND A LIGHT BOX--WAS USED. CARDS WERE HEADED WITH KEY…

  13. Operational alternatives for LANDSAT in California

    NASA Technical Reports Server (NTRS)

    Wilson, P.; Gialdini, M. J.

    1981-01-01

    Data integration is defined and examined as the means of promoting data sharing among the various governmental and private geobased information systems in California. Elements of vertical integration considered included technical factors (such as resolution and classification) and institutional factors (such as organizational control, and legal and political barriers). Attempts are made to fit the theoretical elements of vertical integration into a meaningful structure for looking at the problem from a statewide focus. Both manual (mapped) and machine readable data systems are included. Special attention is given to LANDSAT imagery because of its strong potential for integrated use and its primary in the California Integrated Remote Sensing System program.

  14. Examples of Effective Data Sharing in Scientific Publishing

    DOE PAGES

    Kitchin, John R.

    2015-05-11

    Here, we present a perspective on an approach to data sharing in scientific publications we have been developing in our group. The essence of the approach is that data can be embedded in a human-readable and machine-addressable way within the traditional publishing environment. We show this by example for both computational and experimental data. We articulate a need for new authoring tools to facilitate data sharing, and we discuss the tools we have been developing for this purpose. With these tools, data generation, analysis, and manuscript preparation can be deeply integrated, resulting in easier and better data sharing in scientificmore » publications.« less

  15. Toolsets for Airborne Data (TAD): Improving Machine Readability for ICARTT Data Files

    NASA Technical Reports Server (NTRS)

    Early, Amanda Benson; Beach, Aubrey; Northup, Emily; Wang, Dali; Kusterer, John; Quam, Brandi; Chen, Gao

    2015-01-01

    The Atmospheric Science Data Center (ASDC) at NASA Langley Research Center is responsible for the ingest, archive, and distribution of NASA Earth Science data in the areas of radiation budget, clouds, aerosols, and tropospheric chemistry. The ASDC specializes in atmospheric data that is important to understanding the causes and processes of global climate change and the consequences of human activities on the climate. The ASDC currently supports more than 44 projects and has over 1,700 archived data sets, which increase daily. ASDC customers include scientists, researchers, federal, state, and local governments, academia, industry, and application users, the remote sensing community, and the general public.

  16. Method of modifying a volume mesh using sheet insertion

    DOEpatents

    Borden, Michael J [Albuquerque, NM; Shepherd, Jason F [Albuquerque, NM

    2006-08-29

    A method and machine-readable medium provide a technique to modify a hexahedral finite element volume mesh using dual generation and sheet insertion. After generating a dual of a volume stack (mesh), a predetermined algorithm may be followed to modify (refine) the volume mesh of hexahedral elements. The predetermined algorithm may include the steps of locating a sheet of hexahedral mesh elements, determining a plurality of hexahedral elements within the sheet to refine, shrinking the plurality of elements, and inserting a new sheet of hexahedral elements adjacently to modify the volume mesh. Additionally, another predetermined algorithm using mesh cutting may be followed to modify a volume mesh.

  17. Examples of Effective Data Sharing in Scientific Publishing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kitchin, John R.

    Here, we present a perspective on an approach to data sharing in scientific publications we have been developing in our group. The essence of the approach is that data can be embedded in a human-readable and machine-addressable way within the traditional publishing environment. We show this by example for both computational and experimental data. We articulate a need for new authoring tools to facilitate data sharing, and we discuss the tools we have been developing for this purpose. With these tools, data generation, analysis, and manuscript preparation can be deeply integrated, resulting in easier and better data sharing in scientificmore » publications.« less

  18. Readability of the Written Study Information in Pediatric Research in France

    PubMed Central

    Ménoni, Véronique; Lucas, Noël; Leforestier, Jean-François; Doz, François; Chatellier, Gilles; Jacqz-Aigain, Evelyne; Giraud, Carole; Tréluyer, Jean-Marc; Chappuy, Hélène

    2011-01-01

    Background The aim was to evaluate the readability of research information leaflets (RIL) for minors asked to participate in biomedical research studies and to assess the factors influencing this readability. Methods and Findings All the pediatric protocols from three French pediatric clinical research units were included (N = 104). Three criteria were used to evaluate readability: length of the text, Flesch's readability score and presence of illustrations. We compared the readability of RIL to texts specifically written for children (school textbooks, school exams or extracts from literary works). We assessed the effect of protocol characteristics on readability. The RIL had a median length of 608 words [350 words, 25th percentile; 1005 words, 75th percentile], corresponding to two pages. The readability of the RIL, with a median Flesch score of 40 [30; 47], was much poorer than that of pediatric reference texts, with a Flesch score of 67 [60; 73]. A small proportion of RIL (13/91; 14%) were illustrated. The RIL were longer (p<0.001), more readable (p<0.001) and more likely to be illustrated (p<0.009) for industrial than for institutional sponsors. Conclusion Researchers should routinely compute the reading ease of study information sheets and make greater efforts to improve the readability of written documents for potential participants. PMID:21494689

  19. Readability of "Dear Patient" device advisory notification letters created by a device manufacturer.

    PubMed

    Mueller, Luke A; Sharma, Arjun; Ottenberg, Abigale L; Mueller, Paul S

    2013-04-01

    In 2006, the Heart Rhythm Society (HRS) recommended that cardiovascular implantable electronic device (CIED) manufacturers use advisory notification letters to communicate with affected patients. To evaluate the readability of the HRS sample "patient device advisory notification" letter and those created by 1 CIED manufacturer. The HRS sample letter and 25 Boston Scientific Corporation letters dated from 2005 through 2011 were evaluated by using 6 readability tests. Readability (Flesch-Kincaid score) of the HRS sample letter was grade level 12.5, and median readability of the device manufacturer letters was grade level 12.8 (range 10.8-18.9). Similar results were obtained by using other readability scales. No letters had readability scores at the National Work Group on Literacy and Health's recommended reading level-fifth grade; the letters' readability exceeded this recommended level by an average of 7.7 grades (95% confidence interval 6.9-8.5; P<.001). Likewise, no letters had readability scores at the average reading level of US adults-eighth grade; the letters' readability exceeded this level by an average of 4.7 grades (95% confidence interval 3.9-5.5; P< .001). The readability of the HRS sample letter and those created by a CIED manufacturer significantly exceeded the recommended and average US adults' reading skill levels. Such letters are unlikely to be informative to many patients. CIED manufacturers should ensure that advisory letters are comprehensible to most affected patients. Copyright © 2013 Heart Rhythm Society. Published by Elsevier Inc. All rights reserved.

  20. Controlled English to facilitate human/machine analytical processing

    NASA Astrophysics Data System (ADS)

    Braines, Dave; Mott, David; Laws, Simon; de Mel, Geeth; Pham, Tien

    2013-06-01

    Controlled English is a human-readable information representation format that is implemented using a restricted subset of the English language, but which is unambiguous and directly accessible by simple machine processes. We have been researching the capabilities of CE in a number of contexts, and exploring the degree to which a flexible and more human-friendly information representation format could aid the intelligence analyst in a multi-agent collaborative operational environment; especially in cases where the agents are a mixture of other human users and machine processes aimed at assisting the human users. CE itself is built upon a formal logic basis, but allows users to easily specify models for a domain of interest in a human-friendly language. In our research we have been developing an experimental component known as the "CE Store" in which CE information can be quickly and flexibly processed and shared between human and machine agents. The CE Store environment contains a number of specialized machine agents for common processing tasks and also supports execution of logical inference rules that can be defined in the same CE language. This paper outlines the basic architecture of this approach, discusses some of the example machine agents that have been developed, and provides some typical examples of the CE language and the way in which it has been used to support complex analytical tasks on synthetic data sources. We highlight the fusion of human and machine processing supported through the use of the CE language and CE Store environment, and show this environment with examples of highly dynamic extensions to the model(s) and integration between different user-defined models in a collaborative setting.

Top