Sample records for jane efficient mapping

  1. JANE: efficient mapping of prokaryotic ESTs and variable length sequence reads on related template genomes

    PubMed Central

    2009-01-01

    Background ESTs or variable sequence reads can be available in prokaryotic studies well before a complete genome is known. Use cases include (i) transcriptome studies or (ii) single cell sequencing of bacteria. Without suitable software their further analysis and mapping would have to await finalization of the corresponding genome. Results The tool JANE rapidly maps ESTs or variable sequence reads in prokaryotic sequencing and transcriptome efforts to related template genomes. It provides an easy-to-use graphics interface for information retrieval and a toolkit for EST or nucleotide sequence function prediction. Furthermore, we developed for rapid mapping an enhanced sequence alignment algorithm which reassembles and evaluates high scoring pairs provided from the BLAST algorithm. Rapid assembly on and replacement of the template genome by sequence reads or mapped ESTs is achieved. This is illustrated (i) by data from Staphylococci as well as from a Blattabacteria sequencing effort, (ii) mapping single cell sequencing reads is shown for poribacteria to sister phylum representative Rhodopirellula Baltica SH1. The algorithm has been implemented in a web-server accessible at http://jane.bioapps.biozentrum.uni-wuerzburg.de. Conclusion Rapid prokaryotic EST mapping or mapping of sequence reads is achieved applying JANE even without knowing the cognate genome sequence. PMID:19943962

  2. Walking towards the Past: Loss and Place in Jane Urquhart's "A Map of Glass"

    ERIC Educational Resources Information Center

    Moore, Susan K.

    2008-01-01

    In this article, the author examines the notions of "loss" and "place" in relation to the mourning of one's personal and historical pasts. In doing so, the author draws upon the psychoanalytical writings of Julia Kristeva in an analysis of Jane Urquhart's 2005 novel "A Map of Glass"--a story about emplacement,…

  3. Interview with a SureSmile doctor: Nicole M. Jane. Interview by Robert P. Scholz.

    PubMed

    Jane, Nicole M

    2009-04-01

    SureSmile (OraMetrix, Richardson, Tex) uses advanced technologies, including digital models, virtual simulations, and robotic wire bending to help clinicians provide customized, efficient treatment. Nicole M. Jane has been a solo orthodontic practitioner in South Lake Tahoe, California, since 2004. She has been using SureSmile technology in her practice since December 2005 and is an advisor for SureSmile in her capacity as a practicing orthodontist. Before becoming an orthodontist, Dr Jane was a general dentist in the US Navy, 1998-2001.

  4. An historical ophthalmic study of Jane Austen.

    PubMed

    Wilson, Graham A

    2012-11-01

    Today, no other classic novelist has the popularity or power of Jane Austen, and in 2013 the world will celebrate 200 years of her comic masterpiece Pride and Prejudice. Her millions of fans have an abiding fascination with all aspects of her life, including her health and the cause of her death. This historical ophthalmic study of Jane Austen, based on very incomplete medical bibliographic data, finds that she had a mild ocular surface disorder from age 23. This disorder did not significantly impact on her visual performance for writing. There are many references to eyes in her novels, but Jane's eyes and those of her characters cannot contribute further to the debate around the cause of her death at age 41.

  5. Optimizing Frozen Sample Preparation for Laser Microdissection: Assessment of CryoJane Tape-Transfer System®

    PubMed Central

    Golubeva, Yelena G.; Smith, Roberta M.; Sternberg, Lawrence R.

    2013-01-01

    efficient dissection and high quality RNA retrieval from CryoJane preparations. CryoJane technology therefore has the potential to facilitate standardization of laser microdissection slide preparation from frozen tissues. PMID:23805281

  6. Father Figures in the Novels of Jane Austen

    ERIC Educational Resources Information Center

    Odeh, Adli

    2011-01-01

    Miniaturist as Jane Austen is, she has depicted the life of a few families. In her letter to her niece, Anna Austen, she writes: "three or four families in a country village is the very thing to work on"(Chapman's Edition, 1970, P.10). Jane's knowledge about these families is, in no way shallow. It is rich in variation and contrasts.…

  7. Jane: a new tool for the cophylogeny reconstruction problem.

    PubMed

    Conow, Chris; Fielder, Daniel; Ovadia, Yaniv; Libeskind-Hadas, Ran

    2010-02-03

    This paper describes the theory and implementation of a new software tool, called Jane, for the study of historical associations. This problem arises in parasitology (associations of hosts and parasites), molecular systematics (associations of orderings and genes), and biogeography (associations of regions and orderings). The underlying problem is that of reconciling pairs of trees subject to biologically plausible events and costs associated with these events. Existing software tools for this problem have strengths and limitations, and the new Jane tool described here provides functionality that complements existing tools. The Jane software tool uses a polynomial time dynamic programming algorithm in conjunction with a genetic algorithm to find very good, and often optimal, solutions even for relatively large pairs of trees. The tool allows the user to provide rich timing information on both the host and parasite trees. In addition the user can limit host switch distance and specify multiple host switch costs by specifying regions in the host tree and costs for host switches between pairs of regions. Jane also provides a graphical user interface that allows the user to interactively experiment with modifications to the solutions found by the program. Jane is shown to be a useful tool for cophylogenetic reconstruction. Its functionality complements existing tools and it is therefore likely to be of use to researchers in the areas of parasitology, molecular systematics, and biogeography.

  8. Jane Austen and Addison's disease: an unconvincing diagnosis.

    PubMed

    White, K G

    2009-12-01

    Jane Austen's letters describe a two-year deterioration into bed-ridden exhaustion, with unusual colouring, bilious attacks and rheumatic pains. In 1964, Zachary Cope postulated tubercular Addison's to explain her symptoms and her relatively pain-free illness. Literary scholars later countered this posthumous diagnosis on grounds that are not well substantiated, while medical authors supported his conclusion. Important symptoms reported by contemporary Addison's patients-mental confusion, generalised pain and suffering, weight loss and anorexia-are absent from Jane Austen's letters. Thus, by listening to the patient's perspective, we can conclude it is unlikely that Addison's disease caused Jane Austen's demise. Disseminated bovine tuberculosis would offer a coherent explanation for her symptoms, so that Cope's original suggestion of infective tuberculosis as the cause of her illness may have been correct.

  9. Jane Austen's (1775-1817) references to headache: fact and fiction.

    PubMed

    Larner, A J

    2010-11-01

    References to headache in Jane Austen's works, both fictional and non-fictional, and in biographical works undertaken by Austen family members have been collated. These multiple references suggest that Jane Austen used headache as a narrative device to reflect not only physiological bodily processes but also psychological states, possibly based on her own experience of headache and that of female relations and acquaintances.

  10. Jane Addams, Stories, and Imagination

    ERIC Educational Resources Information Center

    Griffith, Susan C.

    2009-01-01

    Early twentieth-century social activist Jane Addams is best known for her work at Hull House, the settlement house she founded with Ellen Gates Starr in 1889. Adams was also a pacifist, storyteller, writer and philosopher. Through her actions, stories, and writing, Addams modeled a philosophy of democracy-in-action based in imagination and…

  11. Jane Cooke Wright (1919-2013): Pioneering oncologist, woman and humanitarian.

    PubMed

    Crosby, Harriet L

    2016-02-01

    Jane Wright was a fundamental researcher in cancer chemotherapy in the 1950s-1980s and was one of the first scientists to test anti-cancer drugs on humans rather than solely on mice, discovering the use of the popular antimetabolite drug methotrexate on solid tumours. From her research she was able to conclude which specific anti-cancer agents would have the greatest lethal effect on a patient's certain cancer type and she invented a method of delivering chemotherapy agents directly to an internal cancer site. During a time when the Civil Rights in the United States of America were undergoing a transformation to reduce the discrimination and segregation imposed on African Americans and the civil rights activist Martin Luther King Jr made a speech to call for an end to racism in 1963, Jane Wright became the first African American to hold such a high position at a nationally recognised institution and the first woman to be elected President of the New York Cancer Society. US President Lyndon B Johnson appointed Jane to the President's Commission of Heart Disease, Cancer and Stroke (serving 1964-1965) and the National Cancer Advisory Board (serving 1966-1970). Jane retired in 1987 by which time she had published more than 75 scientific papers, led delegations of oncologists in China, the former Soviet Union, Africa and Europe and held key positions in various international and national organisations. Jane Wright passed away on 19 February 2013 aged 93 but her legacy lives on in the name of an award from the American Association of Cancer Research. © The Author(s) 2014.

  12. Peter & Jane: A Program Showcase

    ERIC Educational Resources Information Center

    Kalinowski, Michael

    2008-01-01

    This article features the early childhood programs at Peter & Jane Kindergarten located in Petaling Java, Malaysia. The primary purpose of the programs is to lay a strong foundation for a lifelong love of learning in each child. According to principal Patricia Teh, the activities are funded by parent fees and serve children two to six years of…

  13. "A transcript of actual life": headache in the novels of Jane Austen.

    PubMed

    Larner, Andrew J

    2010-04-01

    References to headache in the novels of Jane Austen have been examined. Nine characters, all female, suffer headache at one time or another, often in association with emotionally stressful situations. As an authorial device, headache may have served Jane Austen as a culturally sanctioned form of bodily expression.

  14. Was Jane Addams a Promiscuous Pragmatist?

    ERIC Educational Resources Information Center

    Atkinson, Becky

    2013-01-01

    Contemporary pragmatist and feminist scholars have proposed the possibilities for "changing the theoretical analyses and concrete practices" of both feminism and classical American pragmatism offered by its recuperation through feminism. Particularly, scholarship on Jane Addams has reached back to retrieve her activism, ethics, and…

  15. Meet EPA Scientist Jane Gallagher, Ph.D.

    EPA Pesticide Factsheets

    Dr. Jane Gallagher is an EPA research health scientist working with expertise in both field and clinical studies. She develops, tests, and integrates new methods, approaches and biomarkers to study human health risks posed by environmental chemicals

  16. Channeling Jane Austen: how it helped me become a better psychotherapist.

    PubMed

    Carver, Christi R

    2014-08-01

    The influences on a therapist are many. This article illustrates how a 19th-century novelist, Jane Austen, informs the work of a nurse therapist in the 21st century. The characters in a Jane Austen novel provide perspectives from setting boundaries to handling feelings. Austen's characters promote an acceptance of less attractive qualities in others and in oneself that can benefit the therapy. Copyright 2014, SLACK Incorporated.

  17. A Reflective Conversation with Jane Piirto

    ERIC Educational Resources Information Center

    Sansom, Shyanne; Barnes, Bryan; Carrizales, Jason; Shaughnessy, Michael F.

    2018-01-01

    This article offers a conversation with Dr. Jane Piirto, author, and professor at Ashland University, where she teaches in the Department of Inclusive Services & Exceptional Learners and the Doctoral Program in Leadership Studies. Here she discusses what she is currently working on, how the five dimensions of overexcitability relate to ADHD,…

  18. Jane Butler Kahle: Passion, Determination, and Vision

    ERIC Educational Resources Information Center

    Scantlebury, Kathryn

    2014-01-01

    In this article, I describe how Jane Butler Kahle's intellectual curiosity, commitment to equity and her ability to use various research approaches, to establish research and networking groups comprised of scholars from diverse backgrounds, theoretical perspectives and geographic locations, to disseminate research outcomes and findings…

  19. Rhetoric and Gender in Jane Austen's "Persuasion."

    ERIC Educational Resources Information Center

    Walzer, Arthur E.

    1995-01-01

    Argues for a reading of Jane Austen's "Persuasion" that undermines Joseph Duffy's reading of the novel as a commentary on shifting social class structures, and which bolsters Nancy Armstrong's reading as a commentary on female voice and the values of the domestic household. Interprets the novel in the light of 18th-century rhetorical…

  20. Jane Austen in the High School Classroom (Open to Suggestion).

    ERIC Educational Resources Information Center

    Fritzer, Penelope

    1996-01-01

    Argues that Jane Austen's novels lend themselves to the high school curriculum, and that students will discover a leisurely, rural world in which the concerns of the young people are often similar to theirs. (SR)

  1. "Socializing Democracy": The Community Literacy Pedagogy of Jane Addams

    ERIC Educational Resources Information Center

    Wendler, Rachael

    2014-01-01

    This article reclaims Jane Addams as a community literacy pedagogue and explicates her pedagogical theory through an analysis of her social thought. Addams' goal of "socializing democracy" through education led her to both encourage immigrant students to associate across difference and to assimilate into dominant literacies--tensions…

  2. Searching for Bill and Jane: Electronic Full-Text Literature.

    ERIC Educational Resources Information Center

    Still, Julie; Kassabian, Vibiana

    1998-01-01

    Examines electronic full-text literature available on the World Wide Web and on CD-ROM. Discusses authors and genres, electronic texts, and fees. Highlights Shakespeare, Jane Austen, and nature writing. Provides a bibliography of Web guides, specialized Shakespeare pages, and pages dealing with the Shakespeare authorship debate and secondary…

  3. Induced Recall of Jane Austen's Novels: Films, Television, Videos.

    ERIC Educational Resources Information Center

    Diaz de Chumaceiro, Cora L.

    2000-01-01

    Notes that the popularity of Jane Austen adaptations in theaters, television, and videos increases the probability that patients and therapists may recall these movies in treatment. Underscores excerpts from a comparison of an Austen novel with the psychoanalytic process and highlights available film adaptations in video format. (SC)

  4. Applied Interventions in the Prevention and Treatment of Obesity Through the Research of Professor Jane Wardle.

    PubMed

    Croker, Helen; Beeken, Rebecca J

    2017-03-01

    Obesity presents a challenge for practitioners, policy makers, researchers and for those with obesity themselves. This review focuses on psychological approaches to its management and prevention in children and adults. Through exploring the work of the late Professor Jane Wardle, we look at the earliest behavioural treatment approaches and how psychological theory has been used to develop more contemporary approaches, for example incorporating genetic feedback and habit formation theory into interventions. We also explore how Jane has challenged thinking about the causal pathways of obesity in relation to eating behaviour. Beyond academic work, Jane was an advocate of developing interventions which had real-world applications. Therefore, we discuss how she not only developed new interventions but also made these widely available and the charity that she established.

  5. Jane Ellen McAllister: Pioneer for Excellence in Teacher Education.

    ERIC Educational Resources Information Center

    Williams-Burns, Winona

    1982-01-01

    Describes the career and accomplishments of Jane Ellen McAllister and highlights her years at Southern University, Virginia State, Fisk University, Miner Teachers College, Grambling State University, and Jackson State College. Also evaluates McAllister's impact on Black education in general. (GC)

  6. 4. Photocopy of photograph, JANE MOSELEY (VESSEL 53) TIED UP ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    4. Photocopy of photograph, JANE MOSELEY (VESSEL 53) TIED UP AT DOCK IN UNKNOWN LOCATION. Date and photographer unknown. (Original in Mariners Museum, Newport News, Virginia, negative #625) - Shooters Island, Ships Graveyard, Vessel No. 53, Newark Bay, Staten Island (subdivision), Richmond County, NY

  7. Cultural Contact through Musical Poetry in Clara Janes's "Kampa"

    ERIC Educational Resources Information Center

    Faszer-McMahon, Debra

    2009-01-01

    Clara Janes's "Kampa" is a love song dedicated to the renowned Czech poet Vladimir Holan. The work includes a musical and lyrical composition performed on tape, and its unconventional musical mode offers an alternative to divisions between western and non-western literary and musical forms. The poetry of "Kampa" presents musical methods of…

  8. A Response to Jane Sahi's "Dialogue as Education: Martin Buber"

    ERIC Educational Resources Information Center

    Baniwal, Vikas

    2014-01-01

    This article is inspired by Jane Sahi's commentary, "Dialogue as Education: Martin Buber," published under the feature "Classics with Commentary" in the Monsoon 2005 issue of "Contemporary Education Dialogue." I seek to further the discussion of the contributions of Martin Buber to the discourse of education through…

  9. Women Consciousness Exploration in Jane Austen and Her Works

    ERIC Educational Resources Information Center

    He, Xiaojun; Liao, Lina

    2015-01-01

    Jane Austen is one of the greatest realistic novelists in English literature of the 19th century. Austen is contrary to consciousness of man centrism and traditional values of masculine writers before her time. All her 6 novels are concerned with love and marriage; they truthfully reflect women life of her age, reveals her feminine consciousness…

  10. Interjections in the Performance of Jane Austen's "Pride and Prejudice"

    ERIC Educational Resources Information Center

    O'Connell, Daniel C.; Kowal, Sabine

    2010-01-01

    Three data sets of primary and secondary interjections were compared: (1) the original interjections written into the text of Jane Austen's (1813/1994) novel "Pride and prejudice"; (2) the interjections read aloud in commercial recordings by six professional readers of the entire text of the novel; (3) the interjections spoken by actresses and…

  11. Jane Austen on love and pedagogical power.

    PubMed

    Fessenbecker, Patrick

    2011-01-01

    This essay notes initially recent prominence of theories of pedagogy that attempt to "de-mystify" it and reveal troubling power relations, and their subsequent contention that love is impossible in the student-teacher relationship. "Pedagogical" interpretations of Jane Austen's fiction, however, see pedagogy as essential to love. I argue that this is so precisely because of the power dynamics involved; drawing on Jessica Benjamin's psychoanalytic interpretation of G. W. F. Hegel's analysis of the "Lord-Bondsman," I suggest that Austen portrays the loving relationship as inherently involving the occupation and subsequent exchange of roles as superior and inferior, incarnated as "teacher" and "student."

  12. Incest, Incorporation, and "King Lear" in Jane Smiley's "A Thousand Acres."

    ERIC Educational Resources Information Center

    Leslie, Marina

    1998-01-01

    Suggests that Jane Smiley's "A Thousand Acres" is a faithful and a "profoundly subversive" revision of Shakespeare's "King Lear." Argues that the terms in which the novel have been most frequently praised, no less than the case made for banning it, raise important questions about the relationship between the novel's…

  13. Verification of the WFAS Lightning Efficiency Map

    Treesearch

    Paul Sopko; Don Latham; Isaac Grenfell

    2007-01-01

    A Lightning Ignition Efficiency map was added to the suite of daily maps offered by the Wildland Fire Assessment System (WFAS) in 1999. This map computes a lightning probability of ignition (POI) based on the estimated fuel type, fuel depth, and 100-hour fuel moisture interpolated from the Remote Automated Weather Station (RAWS) network. An attempt to verify the...

  14. The Awakening of the Social Conscience: Jane Maud Campbell, 1869-1947

    ERIC Educational Resources Information Center

    Jones, Plummer Alston, Jr.

    2012-01-01

    Jane Maud Campbell's career demonstrated her commitment and passion for library services with immigrants and minorities as one of the first advocates for multiculturalism in librarianship. She began her career working in the Newark Public Library and soon was employed as the librarian of the Passaic Public Library. She was the first woman employed…

  15. ReactionMap: an efficient atom-mapping algorithm for chemical reactions.

    PubMed

    Fooshee, David; Andronico, Alessio; Baldi, Pierre

    2013-11-25

    Large databases of chemical reactions provide new data-mining opportunities and challenges. Key challenges result from the imperfect quality of the data and the fact that many of these reactions are not properly balanced or atom-mapped. Here, we describe ReactionMap, an efficient atom-mapping algorithm. Our approach uses a combination of maximum common chemical subgraph search and minimization of an assignment cost function derived empirically from training data. We use a set of over 259,000 balanced atom-mapped reactions from the SPRESI commercial database to train the system, and we validate it on random sets of 1000 and 17,996 reactions sampled from this pool. These large test sets represent a broad range of chemical reaction types, and ReactionMap correctly maps about 99% of the atoms and about 96% of the reactions, with a mean time per mapping of 2 s. Most correctly mapped reactions are mapped with high confidence. Mapping accuracy compares favorably with ChemAxon's AutoMapper, versions 5 and 6.1, and the DREAM Web tool. These approaches correctly map 60.7%, 86.5%, and 90.3% of the reactions, respectively, on the same data set. A ReactionMap server is available on the ChemDB Web portal at http://cdb.ics.uci.edu .

  16. STS-70 Mission Specialist Nancy Jane Currie suits up

    NASA Technical Reports Server (NTRS)

    1995-01-01

    STS-70 Mission Specialist Nancy Jane Currie is donning her launch/entry suit in the Operations and Checkout Building with help from a suit technician. Currie has flown in space once before, on STS-57. Currie and four crew mates will depart shortly for Launch Pad 39B, where the Space Shuttle Discovery is undergoing final preparations for a liftoff scheduled during a two and a half hour launch window opening at 9:41 a.m. EDT.

  17. A Conversation with Kristy and Jane | Center for Cancer Research

    Cancer.gov

    Jane has been coming to the NIH Clinical Center for treatment for neurofibromatosis type 1 (NF1) since she was three years old. She is currently enrolled in a trial that tests Selumetinib, a MEK inhibitor, and her tumor is now 30.7 percent smaller than when she first started this trial three years ago. Her diagnosis has changed the lives of her family but has also given them

  18. A Female Interrogative Reader: The Adolescent Jane Austen Reads and Rewrites (His)tory.

    ERIC Educational Resources Information Center

    Reid-Walsh, Jacqueline

    1992-01-01

    Argues that Jane Austen's unpublished juvenile work "The History of England" has considerable relevance to twentieth-century high-school English classrooms. Notes that the work humorously shows the gender bias of traditional history texts because it is a "woman-centered" rewriting. (RS)

  19. How Jane Addams Expands Our View of Education as an Ethical Enterprise

    ERIC Educational Resources Information Center

    Bruce, Bertram C.

    2017-01-01

    Jane Addams's "Democracy and Social Ethics" is more than a historical artifact describing the work of a prominent social reformer. It is also a significant contribution to philosophy, especially in the area of social ethics. Moreover, though less widely acknowledged, Addams's work is essential for anyone who seeks an ethical vision for…

  20. A Conversation with Kristy and Jane | Center for Cancer Research

    Cancer.gov

    Jane has been coming to the NIH Clinical Center for treatment for neurofibromatosis type 1 (NF1) since she was three years old. She is currently enrolled in a trial that tests Selumetinib, a MEK inhibitor, and her tumor is now 30.7 percent smaller than when she first started this trial three years ago. Her diagnosis has changed the lives of her family but has also given them new passions and perseverance. Read more...

  1. Jane Austen (1775-1817) and the cultural history of health.

    PubMed

    Biddiss, Michael

    2014-08-01

    This paper provides a review of some aspects of the life and novels of Jane Austen that have particular bearing on her approach to issues of sickness and health. It is based on a Keynote Lecture given at the Annual Congress of the British Society for the History of Medicine on 1 September 2011 at the University of Surrey. © IMechE 2013 Reprints and permissions: sagepub.co.uk/journalsPermissions.nav.

  2. Readings on Jane Austen. The Greenhaven Press Literary Companion to British Authors.

    ERIC Educational Resources Information Center

    Swisher, Clarice, Ed.

    Designed for young adults, this book on Jane Austen's novels is one of an anthology series providing accessible resources for students researching great literary lives and works. Contributing writers' essays in the book are taken from a wide variety of sources and are edited to accommodate the reading and comprehension levels of young adults; each…

  3. The I-Generation--From Toddlers to Teenagers: A Conversation with Jane M. Healy.

    ERIC Educational Resources Information Center

    Tell, Carol

    2000-01-01

    In "Failure to Connect" (1998), Jane Healy examined pros and cons of computer use, warning that good teachers, small classes, and challenging curricula trump high-tech products. Computers can impede youngsters' development. Computers enhance learning only if teachers comprehend them, use appropriate applications, and define learning…

  4. A Journey through Creativity as a Writer and Researcher: "An Interview with Jane Piirto"

    ERIC Educational Resources Information Center

    Henshon, Suzanna E.

    2006-01-01

    This article presents an interview with Jane Piirto, a Trustees' Professor at Ashland University in Ohio, Director of Talent Development Education, teaching in the departments of Curriculum and Instruction and in Educational Leadership. Her scholarly books are "Talented Children and Adults" (3 editions, latest 2007 from Prufrock Press);…

  5. The Generalist's Corner: On Following Your Bliss--An Interview with Jane S. Halonen

    ERIC Educational Resources Information Center

    Irons, Jessica G.

    2007-01-01

    Jane S. Halonen is the Dean of the College of Arts and Sciences at the University of West Florida, where she insists on teaching introductory psychology each fall. She was employed in various roles at Alverno College for 17 years and also served for 5 years as the Director of the School of Psychology at James Madison University. Past president of…

  6. JANE, A new information retrieval system for the Radiation Shielding Information Center

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Trubey, D.K.

    A new information storage and retrieval system has been developed for the Radiation Shielding Information Center (RSIC) at Oak Ridge National Laboratory to replace mainframe systems that have become obsolete. The database contains citations and abstracts of literature which were selected by RSIC analysts and indexed with terms from a controlled vocabulary. The database, begun in 1963, has been maintained continuously since that time. The new system, called JANE, incorporates automatic indexing techniques and on-line retrieval using the RSIC Data General Eclipse MV/4000 minicomputer, Automatic indexing and retrieval techniques based on fuzzy-set theory allow the presentation of results in ordermore » of Retrieval Status Value. The fuzzy-set membership function depends on term frequency in the titles and abstracts and on Term Discrimination Values which indicate the resolving power of the individual terms. These values are determined by the Cover Coefficient method. The use of a commercial database base to store and retrieve the indexing information permits rapid retrieval of the stored documents. Comparisons of the new and presently-used systems for actual searches of the literature indicate that it is practical to replace the mainframe systems with a minicomputer system similar to the present version of JANE. 18 refs., 10 figs.« less

  7. Efficient Bit-to-Symbol Likelihood Mappings

    NASA Technical Reports Server (NTRS)

    Moision, Bruce E.; Nakashima, Michael A.

    2010-01-01

    This innovation is an efficient algorithm designed to perform bit-to-symbol and symbol-to-bit likelihood mappings that represent a significant portion of the complexity of an error-correction code decoder for high-order constellations. Recent implementation of the algorithm in hardware has yielded an 8- percent reduction in overall area relative to the prior design.

  8. Barriers to Treatment of Tinea Capitis in Children Living in the Jane Finch Community of Toronto.

    PubMed

    Zur, Rebecca L; Shapero, Jonathan; Shapero, Harvey

    2015-01-01

    Tinea capitis is a common fungal infection of the scalp. If left untreated, tinea capitis infection can cause severe inflammatory reactions and the development of kerion. Tinea capitis is effectively treated with oral antifungals, but at present these are not covered under government assistance programs. To assess the potential impact of a limited use code for antifungal therapy in the treatment of childhood tinea capitis. Fourteen family physicians practicing in the Jane Finch area were surveyed on their experience treating tinea capitis in this community. Seventy-one percent of surveyed family physicians felt that cost impedes the treatment of tinea capitis in their practice, and 100% felt that a limited use code would have a positive impact on their patients. A limited use code for oral antifungal treatments of tinea capitis may provide a simple, cost-effective solution to a major problem impacting children in the Jane Finch area. © The Author(s) 2015.

  9. Jane Austen's Novels as a Guide to Social and Individual Responsibility for High School Students.

    ERIC Educational Resources Information Center

    Fritzer, Penelope

    1998-01-01

    Jane Austen's novels are particularly appropriate for adolescents. These classics deal with topics of high interest to young people. Austen grapples with the question of what the individual owes to society and what he or she is obliged to tolerate in the way of strictures on behavior. Article proposes that Austen's novels should be more widely…

  10. Multi-Depth-Map Raytracing for Efficient Large-Scene Reconstruction.

    PubMed

    Arikan, Murat; Preiner, Reinhold; Wimmer, Michael

    2016-02-01

    With the enormous advances of the acquisition technology over the last years, fast processing and high-quality visualization of large point clouds have gained increasing attention. Commonly, a mesh surface is reconstructed from the point cloud and a high-resolution texture is generated over the mesh from the images taken at the site to represent surface materials. However, this global reconstruction and texturing approach becomes impractical with increasing data sizes. Recently, due to its potential for scalability and extensibility, a method for texturing a set of depth maps in a preprocessing and stitching them at runtime has been proposed to represent large scenes. However, the rendering performance of this method is strongly dependent on the number of depth maps and their resolution. Moreover, for the proposed scene representation, every single depth map has to be textured by the images, which in practice heavily increases processing costs. In this paper, we present a novel method to break these dependencies by introducing an efficient raytracing of multiple depth maps. In a preprocessing phase, we first generate high-resolution textured depth maps by rendering the input points from image cameras and then perform a graph-cut based optimization to assign a small subset of these points to the images. At runtime, we use the resulting point-to-image assignments (1) to identify for each view ray which depth map contains the closest ray-surface intersection and (2) to efficiently compute this intersection point. The resulting algorithm accelerates both the texturing and the rendering of the depth maps by an order of magnitude.

  11. Liborg: a lidar-based robot for efficient 3D mapping

    NASA Astrophysics Data System (ADS)

    Vlaminck, Michiel; Luong, Hiep; Philips, Wilfried

    2017-09-01

    In this work we present Liborg, a spatial mapping and localization system that is able to acquire 3D models on the y using data originated from lidar sensors. The novelty of this work is in the highly efficient way we deal with the tremendous amount of data to guarantee fast execution times while preserving sufficiently high accuracy. The proposed solution is based on a multi-resolution technique based on octrees. The paper discusses and evaluates the main benefits of our approach including its efficiency regarding building and updating the map and its compactness regarding compressing the map. In addition, the paper presents a working prototype consisting of a robot equipped with a Velodyne Lidar Puck (VLP-16) and controlled by a Raspberry Pi serving as an independent acquisition platform.

  12. "Clearing the Sill of the World": Jane Eyre and the Power of Education in the Nineteenth-Century Novel

    ERIC Educational Resources Information Center

    Davis, Nancy L.; Rainey, William

    2009-01-01

    The idea of education in nineteenth-century women's writing revolves around social class, social mores, and the subtleties of the writer's imagination. Nowhere can this be seen more vividly and thoroughly than in Charlotte Bronte's novel, "Jane Eyre". The book's opening scene, striking in its symbolic detail, highlights and foreshadows the…

  13. An efficient cardiac mapping strategy for radiofrequency catheter ablation with active learning.

    PubMed

    Feng, Yingjing; Guo, Ziyan; Dong, Ziyang; Zhou, Xiao-Yun; Kwok, Ka-Wai; Ernst, Sabine; Lee, Su-Lin

    2017-07-01

    A major challenge in radiofrequency catheter ablation procedures is the voltage and activation mapping of the endocardium, given a limited mapping time. By learning from expert interventional electrophysiologists (operators), while also making use of an active-learning framework, guidance on performing cardiac voltage mapping can be provided to novice operators or even directly to catheter robots. A learning from demonstration (LfD) framework, based upon previous cardiac mapping procedures performed by an expert operator, in conjunction with Gaussian process (GP) model-based active learning, was developed to efficiently perform voltage mapping over right ventricles (RV). The GP model was used to output the next best mapping point, while getting updated towards the underlying voltage data pattern as more mapping points are taken. A regularized particle filter was used to keep track of the kernel hyperparameter used by GP. The travel cost of the catheter tip was incorporated to produce time-efficient mapping sequences. The proposed strategy was validated on a simulated 2D grid mapping task, with leave-one-out experiments on 25 retrospective datasets, in an RV phantom using the Stereotaxis Niobe ® remote magnetic navigation system, and on a tele-operated catheter robot. In comparison with an existing geometry-based method, regression error was reduced and was minimized at a faster rate over retrospective procedure data. A new method of catheter mapping guidance has been proposed based on LfD and active learning. The proposed method provides real-time guidance for the procedure, as well as a live evaluation of mapping sufficiency.

  14. Think, Jane, Think. See Jane Think. Go, Jane... Metacognition and Learning in the Library

    ERIC Educational Resources Information Center

    Jaeger, Paige

    2007-01-01

    Buzzwords are as prolific in educational circles as bunny rabbits are in spring. Over the last 10 years everyone has heard the buzz of multiculturism, multiple intelligences, learning modalities, essential questions, cultural literacy, media literacy, differentiated instruction, learning by design, curriculum alignment, curriculum mapping,…

  15. MIMO: an efficient tool for molecular interaction maps overlap

    PubMed Central

    2013-01-01

    Background Molecular pathways represent an ensemble of interactions occurring among molecules within the cell and between cells. The identification of similarities between molecular pathways across organisms and functions has a critical role in understanding complex biological processes. For the inference of such novel information, the comparison of molecular pathways requires to account for imperfect matches (flexibility) and to efficiently handle complex network topologies. To date, these characteristics are only partially available in tools designed to compare molecular interaction maps. Results Our approach MIMO (Molecular Interaction Maps Overlap) addresses the first problem by allowing the introduction of gaps and mismatches between query and template pathways and permits -when necessary- supervised queries incorporating a priori biological information. It then addresses the second issue by relying directly on the rich graph topology described in the Systems Biology Markup Language (SBML) standard, and uses multidigraphs to efficiently handle multiple queries on biological graph databases. The algorithm has been here successfully used to highlight the contact point between various human pathways in the Reactome database. Conclusions MIMO offers a flexible and efficient graph-matching tool for comparing complex biological pathways. PMID:23672344

  16. Cause of Death in “John Doe & Jane Doe”: A 5 year review

    PubMed Central

    Dasari, Harish; Singh, Amandeep

    2014-01-01

    Sometimes the opinion regarding the cause of death in “John Doe or Jane Doe” i.e. on unknown dead bodies is a test of ability of the forensic expert and on many occasions it yields little or no results. Here the identification of the body as such poses problems; rest aside the opinion regarding the cause/ manner of death. The present 5yr study was undertaken in the Department of Forensic Medicine & Toxicology, Government Medical College & Hospital, Chandigarh to find the patterns of cause of death in unknown dead bodies, as very little literature is available with regard to John Doe or Jane Doe cases as a group, in India. Unidentified bodies comprised 4 % of the total 3165 cases brought for post-mortem examination to the department. Maximum cases belonged to the age group 41 - 50 years, 30 %. Majority of the opinions regarding the cause of death were given as “no definite opinion” (31%), followed by “cranio-cerebral damage” (30 %) and coronary insufficiency/ Cardiac disease/ aortic aneurysm rupture, (8.9%). Following measures should be undertaken to increase the chances of getting these unknown bodies identified and thereby increasing the chances of arriving at a definite cause of death: drafting of additional legislation for the management of unidentified dead bodies along with streamlining of work on the part of police, use of active investigation and modern investigative techniques, fixing the accountability of the police. Internet based sites of the police like ZIPNET (Zonal Integrated Police Networking) in Northern India, should also be used. PMID:25302219

  17. Adaptive proxy map server for efficient vector spatial data rendering

    NASA Astrophysics Data System (ADS)

    Sayar, Ahmet

    2013-01-01

    The rapid transmission of vector map data over the Internet is becoming a bottleneck of spatial data delivery and visualization in web-based environment because of increasing data amount and limited network bandwidth. In order to improve both the transmission and rendering performances of vector spatial data over the Internet, we propose a proxy map server enabling parallel vector data fetching as well as caching to improve the performance of web-based map servers in a dynamic environment. Proxy map server is placed seamlessly anywhere between the client and the final services, intercepting users' requests. It employs an efficient parallelization technique based on spatial proximity and data density in case distributed replica exists for the same spatial data. The effectiveness of the proposed technique is proved at the end of the article by the application of creating map images enriched with earthquake seismic data records.

  18. c-Jun controls the efficiency of MAP kinase signaling by transcriptional repression of MAP kinase phosphatases

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sprowles, Amy; Robinson, Dan; Wu Yimi

    2005-08-15

    The mammalian JNK signaling pathway regulates the transcriptional response of cells to environmental stress, including UV irradiation. This signaling pathway is composed of a classical MAP kinase cascade; activation results in phosphorylation of the transcription factor substrates c-Jun and ATF2, and leads to changes in gene expression. The defining components of this pathway are conserved in the fission yeast S. pombe, where the genetic studies have shown that the ability of the JNK homolog Spc1 to be activated in response to UV irradiation is dependent on the presence of the transcription factor substrate Atf1. We have used genetic analysis tomore » define the role of c-Jun in activation of the mammalian JNK signaling pathway. Our results show that optimal activation of JNK requires the presence of its transcription factor substrate c-Jun. Mutational analysis shows that the ability of c-Jun to support efficient activation of JNK requires the ability of Jun to bind DNA, suggesting a transcriptional mechanism. Consistent with this, we show that c-Jun represses the expression of several MAP kinase phosphatases. In the absence of c-Jun, the increased expression of MAP kinase phosphatases leads to impaired activation of the ERK, JNK, and p38 MAP kinases after pathway activation. The results show that one function of c-Jun is to regulate the efficiency of signaling by the ERK, p38, and JNK MAP kinases, a function that is likely to affect cellular responses to many different stimuli.« less

  19. Efficient and Scalable Graph Similarity Joins in MapReduce

    PubMed Central

    Chen, Yifan; Zhang, Weiming; Tang, Jiuyang

    2014-01-01

    Along with the emergence of massive graph-modeled data, it is of great importance to investigate graph similarity joins due to their wide applications for multiple purposes, including data cleaning, and near duplicate detection. This paper considers graph similarity joins with edit distance constraints, which return pairs of graphs such that their edit distances are no larger than a given threshold. Leveraging the MapReduce programming model, we propose MGSJoin, a scalable algorithm following the filtering-verification framework for efficient graph similarity joins. It relies on counting overlapping graph signatures for filtering out nonpromising candidates. With the potential issue of too many key-value pairs in the filtering phase, spectral Bloom filters are introduced to reduce the number of key-value pairs. Furthermore, we integrate the multiway join strategy to boost the verification, where a MapReduce-based method is proposed for GED calculation. The superior efficiency and scalability of the proposed algorithms are demonstrated by extensive experimental results. PMID:25121135

  20. Efficient and scalable graph similarity joins in MapReduce.

    PubMed

    Chen, Yifan; Zhao, Xiang; Xiao, Chuan; Zhang, Weiming; Tang, Jiuyang

    2014-01-01

    Along with the emergence of massive graph-modeled data, it is of great importance to investigate graph similarity joins due to their wide applications for multiple purposes, including data cleaning, and near duplicate detection. This paper considers graph similarity joins with edit distance constraints, which return pairs of graphs such that their edit distances are no larger than a given threshold. Leveraging the MapReduce programming model, we propose MGSJoin, a scalable algorithm following the filtering-verification framework for efficient graph similarity joins. It relies on counting overlapping graph signatures for filtering out nonpromising candidates. With the potential issue of too many key-value pairs in the filtering phase, spectral Bloom filters are introduced to reduce the number of key-value pairs. Furthermore, we integrate the multiway join strategy to boost the verification, where a MapReduce-based method is proposed for GED calculation. The superior efficiency and scalability of the proposed algorithms are demonstrated by extensive experimental results.

  1. Efficient algorithms for dilated mappings of binary trees

    NASA Technical Reports Server (NTRS)

    Iqbal, M. Ashraf

    1990-01-01

    The problem is addressed to find a 1-1 mapping of the vertices of a binary tree onto those of a target binary tree such that the son of a node on the first binary tree is mapped onto a descendent of the image of that node in the second binary tree. There are two natural measures of the cost of this mapping, namely the dilation cost, i.e., the maximum distance in the target binary tree between the images of vertices that are adjacent in the original tree. The other measure, expansion cost, is defined as the number of extra nodes/edges to be added to the target binary tree in order to ensure a 1-1 mapping. An efficient algorithm to find a mapping of one binary tree onto another is described. It is shown that it is possible to minimize one cost of mapping at the expense of the other. This problem arises when designing pipelined arithmetic logic units (ALU) for special purpose computers. The pipeline is composed of ALU chips connected in the form of a binary tree. The operands to the pipeline can be supplied to the leaf nodes of the binary tree which then process and pass the results up to their parents. The final result is available at the root. As each new application may require a distinct nesting of operations, it is useful to be able to find a good mapping of a new binary tree over existing ALU tree. Another problem arises if every distinct required binary tree is known beforehand. Here it is useful to hardwire the pipeline in the form of a minimal supertree that contains all required binary trees.

  2. Efficient statistical mapping of avian count data

    USGS Publications Warehouse

    Royle, J. Andrew; Wikle, C.K.

    2005-01-01

    We develop a spatial modeling framework for count data that is efficient to implement in high-dimensional prediction problems. We consider spectral parameterizations for the spatially varying mean of a Poisson model. The spectral parameterization of the spatial process is very computationally efficient, enabling effective estimation and prediction in large problems using Markov chain Monte Carlo techniques. We apply this model to creating avian relative abundance maps from North American Breeding Bird Survey (BBS) data. Variation in the ability of observers to count birds is modeled as spatially independent noise, resulting in over-dispersion relative to the Poisson assumption. This approach represents an improvement over existing approaches used for spatial modeling of BBS data which are either inefficient for continental scale modeling and prediction or fail to accommodate important distributional features of count data thus leading to inaccurate accounting of prediction uncertainty.

  3. SModelS v1.1 user manual: Improving simplified model constraints with efficiency maps

    NASA Astrophysics Data System (ADS)

    Ambrogi, Federico; Kraml, Sabine; Kulkarni, Suchita; Laa, Ursula; Lessa, Andre; Magerl, Veronika; Sonneveld, Jory; Traub, Michael; Waltenberger, Wolfgang

    2018-06-01

    SModelS is an automatized tool for the interpretation of simplified model results from the LHC. It allows to decompose models of new physics obeying a Z2 symmetry into simplified model components, and to compare these against a large database of experimental results. The first release of SModelS, v1.0, used only cross section upper limit maps provided by the experimental collaborations. In this new release, v1.1, we extend the functionality of SModelS to efficiency maps. This increases the constraining power of the software, as efficiency maps allow to combine contributions to the same signal region from different simplified models. Other new features of version 1.1 include likelihood and χ2 calculations, extended information on the topology coverage, an extended database of experimental results as well as major speed upgrades for both the code and the database. We describe in detail the concepts and procedures used in SModelS v1.1, explaining in particular how upper limits and efficiency map results are dealt with in parallel. Detailed instructions for code usage are also provided.

  4. A highly efficient approach to protein interactome mapping based on collaborative filtering framework.

    PubMed

    Luo, Xin; You, Zhuhong; Zhou, Mengchu; Li, Shuai; Leung, Hareton; Xia, Yunni; Zhu, Qingsheng

    2015-01-09

    The comprehensive mapping of protein-protein interactions (PPIs) is highly desired for one to gain deep insights into both fundamental cell biology processes and the pathology of diseases. Finely-set small-scale experiments are not only very expensive but also inefficient to identify numerous interactomes despite their high accuracy. High-throughput screening techniques enable efficient identification of PPIs; yet the desire to further extract useful knowledge from these data leads to the problem of binary interactome mapping. Network topology-based approaches prove to be highly efficient in addressing this problem; however, their performance deteriorates significantly on sparse putative PPI networks. Motivated by the success of collaborative filtering (CF)-based approaches to the problem of personalized-recommendation on large, sparse rating matrices, this work aims at implementing a highly efficient CF-based approach to binary interactome mapping. To achieve this, we first propose a CF framework for it. Under this framework, we model the given data into an interactome weight matrix, where the feature-vectors of involved proteins are extracted. With them, we design the rescaled cosine coefficient to model the inter-neighborhood similarity among involved proteins, for taking the mapping process. Experimental results on three large, sparse datasets demonstrate that the proposed approach outperforms several sophisticated topology-based approaches significantly.

  5. A Highly Efficient Approach to Protein Interactome Mapping Based on Collaborative Filtering Framework

    PubMed Central

    Luo, Xin; You, Zhuhong; Zhou, Mengchu; Li, Shuai; Leung, Hareton; Xia, Yunni; Zhu, Qingsheng

    2015-01-01

    The comprehensive mapping of protein-protein interactions (PPIs) is highly desired for one to gain deep insights into both fundamental cell biology processes and the pathology of diseases. Finely-set small-scale experiments are not only very expensive but also inefficient to identify numerous interactomes despite their high accuracy. High-throughput screening techniques enable efficient identification of PPIs; yet the desire to further extract useful knowledge from these data leads to the problem of binary interactome mapping. Network topology-based approaches prove to be highly efficient in addressing this problem; however, their performance deteriorates significantly on sparse putative PPI networks. Motivated by the success of collaborative filtering (CF)-based approaches to the problem of personalized-recommendation on large, sparse rating matrices, this work aims at implementing a highly efficient CF-based approach to binary interactome mapping. To achieve this, we first propose a CF framework for it. Under this framework, we model the given data into an interactome weight matrix, where the feature-vectors of involved proteins are extracted. With them, we design the rescaled cosine coefficient to model the inter-neighborhood similarity among involved proteins, for taking the mapping process. Experimental results on three large, sparse datasets demonstrate that the proposed approach outperforms several sophisticated topology-based approaches significantly. PMID:25572661

  6. A Highly Efficient Approach to Protein Interactome Mapping Based on Collaborative Filtering Framework

    NASA Astrophysics Data System (ADS)

    Luo, Xin; You, Zhuhong; Zhou, Mengchu; Li, Shuai; Leung, Hareton; Xia, Yunni; Zhu, Qingsheng

    2015-01-01

    The comprehensive mapping of protein-protein interactions (PPIs) is highly desired for one to gain deep insights into both fundamental cell biology processes and the pathology of diseases. Finely-set small-scale experiments are not only very expensive but also inefficient to identify numerous interactomes despite their high accuracy. High-throughput screening techniques enable efficient identification of PPIs; yet the desire to further extract useful knowledge from these data leads to the problem of binary interactome mapping. Network topology-based approaches prove to be highly efficient in addressing this problem; however, their performance deteriorates significantly on sparse putative PPI networks. Motivated by the success of collaborative filtering (CF)-based approaches to the problem of personalized-recommendation on large, sparse rating matrices, this work aims at implementing a highly efficient CF-based approach to binary interactome mapping. To achieve this, we first propose a CF framework for it. Under this framework, we model the given data into an interactome weight matrix, where the feature-vectors of involved proteins are extracted. With them, we design the rescaled cosine coefficient to model the inter-neighborhood similarity among involved proteins, for taking the mapping process. Experimental results on three large, sparse datasets demonstrate that the proposed approach outperforms several sophisticated topology-based approaches significantly.

  7. Gender Anarchy as Social Justice: An Analytic Reconstruction of the Idea of Epistemic Equality in Jane Roland Martin's "Reclaiming a Conversation"

    ERIC Educational Resources Information Center

    Seals, Greg

    2006-01-01

    Jane Roland Martin's later work, especially as represented in "The Schoolhome: Rethinking Schools for Changing Families," has been attacked as vague, essentialistic, and a formula for the (re)feminization of education. This paper does not attempt to defend Martin against these criticisms because such a defense seems impossible for…

  8. Spherical self-organizing map using efficient indexed geodesic data structure.

    PubMed

    Wu, Yingxin; Takatsuka, Masahiro

    2006-01-01

    The two-dimensional (2D) Self-Organizing Map (SOM) has a well-known "border effect". Several spherical SOMs which use lattices of the tessellated icosahedron have been proposed to solve this problem. However, existing data structures for such SOMs are either not space efficient or are time consuming when searching the neighborhood. We introduce a 2D rectangular grid data structure to store the icosahedron-based geodesic dome. Vertices relationships are maintained by their positions in the data structure rather than by immediate neighbor pointers or an adjacency list. Increasing the number of neurons can be done efficiently because the overhead caused by pointer updates is reduced. Experiments show that the spherical SOM using our data structure, called a GeoSOM, runs with comparable speed to the conventional 2D SOM. The GeoSOM also reduces data distortion due to removal of the boundaries. Furthermore, we developed an interface to project the GeoSOM onto the 2D plane using a cartographic approach, which gives users a global view of the spherical data map. Users can change the center of the 2D data map interactively. In the end, we compare the GeoSOM to the other spherical SOMs by space complexity and time complexity.

  9. "Young ladies are delicate plants": Jane Austen and Greenhouse Romanticism.

    PubMed

    Lynch, Deidre Shauna

    2010-01-01

    By annotating how in Mansfield Park and Northanger Abbey Jane Austen tracks between the novelist's domain and the naturalist's, this essay seeks to unsettle some entrenched assumptions about her relationship to realism—and the ideological work of naturalization that realism is said to sponsor—as well as to romanticism. Austen's era, which we customarily identify as the time of a romantic return to nature, was marked by conflicts over what nature did and did not include—conflicts between, in the parlance of the period, "the botanist" and "the florist," between an account of nature as that which was given and an account of nature as that which (as with the modern nurseryman's new hybrids and luxuriants) might be manufactured. In arranging for her young heroines' stories of growing up to unfold in the "artificial climates" of modern fashionable gardening and amidst "florists' flowers," Austen deliberately works through those conflicts. She inhabits the plot of "natural development" to which she is often linked in elusively oppositional ways.

  10. Optimized efficient liver T1ρ mapping using limited spin lock times

    NASA Astrophysics Data System (ADS)

    Yuan, Jing; Zhao, Feng; Griffith, James F.; Chan, Queenie; Wang, Yi-Xiang J.

    2012-03-01

    T1ρ relaxation has recently been found to be sensitive to liver fibrosis and has potential to be used for early detection of liver fibrosis and grading. Liver T1ρ imaging and accurate mapping are challenging because of the long scan time, respiration motion and high specific absorption rate. Reduction and optimization of spin lock times (TSLs) are an efficient way to reduce scan time and radiofrequency energy deposition of T1ρ imaging, but maintain the near-optimal precision of T1ρ mapping. This work analyzes the precision in T1ρ estimation with limited, in particular two, spin lock times, and explores the feasibility of using two specific operator-selected TSLs for efficient and accurate liver T1ρ mapping. Two optimized TSLs were derived by theoretical analysis and numerical simulations first, and tested experimentally by in vivo rat liver T1ρ imaging at 3 T. The simulation showed that the TSLs of 1 and 50 ms gave optimal T1ρ estimation in a range of 10-100 ms. In the experiment, no significant statistical difference was found between the T1ρ maps generated using the optimized two-TSL combination and the maps generated using the six TSLs of [1, 10, 20, 30, 40, 50] ms according to one-way ANOVA analysis (p = 0.1364 for liver and p = 0.8708 for muscle).

  11. Analysis of improved government geological map information for mineral exploration: Incorporating efficiency, productivity, effectiveness, and risk considerations

    USGS Publications Warehouse

    Bernknopf, R.L.; Wein, A.M.; St-Onge, M. R.; Lucas, S.B.

    2007-01-01

    This bulletin/professional paper focuses on the value of geoscientific information and knowledge, as provided in published government bedrock geological maps, to the mineral exploration sector. An economic model is developed that uses an attribute- ranking approach to convert geological maps into domains of mineral favourability. Information about known deposits in these (or analogous) favourability domains allow the calculation of exploration search statistics that provide input into measures of exploration efficiency, productivity, effectiveness, risk, and cost stemming from the use of the published geological maps. Two case studies, the Flin Flon Belt (Manitoba and Saskatchewan) and the south Baffin Island area (Nunavut), demonstrate that updated, finer resolution maps can be used to identify more exploration campaign options, and campaigns thats are more efficient, more effective, and less risky than old, coarser resolution maps when used as a guide for mineral exploration. The Flin Flon Belt study illustrates that an updated, coarser resolution bedrock map enables improved mineral exploration efficiency, productivity, and effectiveness by locating 60% more targets and supporting an exploration campaign that is 44% more efficient. Refining the map resolution provides an additional 17% reduction in search effort across all favourable domains and a 55% reduction in search effort in the most favourable domain. The south Baffin Island case study projects a 40% increase in expected targets and a 27% reduction in search effort when the updated, finer resolution map is used in lieu of the old, coarser resolution map. On southern Baffin Island, the economic value of the up dated map ranges from CAN$2.28 million to CAN$15.21 million, which can be compared to the CAN$1.86 million that it cost to produce the map (a multiplier effect of up to eight).

  12. Energy-efficient virtual optical network mapping approaches over converged flexible bandwidth optical networks and data centers.

    PubMed

    Chen, Bowen; Zhao, Yongli; Zhang, Jie

    2015-09-21

    In this paper, we develop a virtual link priority mapping (LPM) approach and a virtual node priority mapping (NPM) approach to improve the energy efficiency and to reduce the spectrum usage over the converged flexible bandwidth optical networks and data centers. For comparison, the lower bound of the virtual optical network mapping is used for the benchmark solutions. Simulation results show that the LPM approach achieves the better performance in terms of power consumption, energy efficiency, spectrum usage, and the number of regenerators compared to the NPM approach.

  13. Global map of solar power production efficiency, considering micro climate factors

    NASA Astrophysics Data System (ADS)

    Hassanpour Adeh, E.; Higgins, C. W.

    2017-12-01

    Natural resources degradation and greenhouse gas emissions are creating a global crisis. Renewable energy is the most reliable option to mitigate this environmental dilemma. Abundancy of solar energy makes it highly attractive source of electricity. The existing global spatial maps of available solar energy are created with various models which consider the irradiation, latitude, cloud cover, elevation, shading and aerosols, and neglect the influence of local meteorological conditions. In this research, the influences of microclimatological variables on solar energy productivity were investigated with an in-field study at the Rabbit Hills solar arrays near Oregon State University. The local studies were extended to a global level, where global maps of solar power were produced, taking the micro climate variables into account. These variables included: temperature, relative humidity, wind speed, wind direction, solar radiation. The energy balance approach was used to synthesize the data and compute the efficiencies. The results confirmed that the solar power efficiency can be directly affected by the air temperature and wind speed.

  14. Efficiency and security problems of anonymous key agreement protocol based on chaotic maps

    NASA Astrophysics Data System (ADS)

    Yoon, Eun-Jun

    2012-07-01

    In 2011, Niu-Wang proposed an anonymous key agreement protocol based on chaotic maps in [Niu Y, Wang X. An anonymous key agreement protocol based on chaotic maps. Commun Nonlinear Sci Simulat 2011;16(4):1986-92]. Niu-Wang's protocol not only achieves session key agreement between a server and a user, but also allows the user to anonymously interact with the server. Nevertheless, this paper points out that Niu-Wang's protocol has the following efficiency and security problems: (1) The protocol has computational efficiency problem when a trusted third party decrypts the user sending message. (2) The protocol is vulnerable to Denial of Service (DoS) attack based on illegal message modification by an attacker.

  15. Jane Austen's novels as a guide to social and individual responsibility for high school students.

    PubMed

    Fritzer, P

    1998-01-01

    Jane Austen's novels are particularly appropriate for adolescents. These classics deal with topics of high interest to young people: money, family relationships and obligations, headstrong behavior, and society's rules, all combined in the romantic interplay between the sexes. Indeed, some critics have accused Austen of writing the same book over and over again, in that they deal exhaustively with finding appropriate partners for young people. Yet, each is concerned with personal integrity in different circumstances. Austen grapples with the question of what the individual owes to society and what he or she is obliged to tolerate in the way of strictures on behavior, a question that is especially relevant for adolescents.

  16. The chemical work of Alexander and Jane Marcet.

    PubMed

    Rosenfeld, L

    2001-04-01

    Alexander Marcet was an authority on urinary calculi and their analysis when few medical practitioners appreciated the usefulness of chemistry in the explanation and treatment of disease. In An Essay on The Chemical History and Medical Treatment of Calculous Disorders, he described the discovery of an xanthine stone. He drew line illustrations of simple chemical apparatus useful for bedside analysis. His microtechnique used drops of solution and pinhead pieces of calculi; reagents were acids and alkalies and the blowpipe in conjunction with a small alcohol lamp. He reported the earliest description of a disorder later named "alcaptonuria". Marcet's work and that of a few others, on the chemical composition of urine and calculi, laid the foundations of our present knowledge. Between 1807 and 1820, his lectures to the medical students at Guy's Hospital were illustrated by experiments. Jane Haldimand Marcet wrote the very popular CONVERSATIONS: on Chemistry (16 editions in Great Britain). Her book dominated elementary chemical instruction during the first half of the 19th century. She followed Lavoisier's scheme of classification and explained chemical reactions in terms of affinity, aggregation, gravitation, and repulsion. Her advocacy that experimentation accompany lecture was new. The availability of serious scientific education in the new women's academies set the stage for increasing women's involvement in science. She also published a series of CONVERSATIONS: The topics were Political Economy, Natural Philosophy, and Vegetable Physiology.

  17. Efficient design of nanoplasmonic waveguide devices using the space mapping algorithm.

    PubMed

    Dastmalchi, Pouya; Veronis, Georgios

    2013-12-30

    We show that the space mapping algorithm, originally developed for microwave circuit optimization, can enable the efficient design of nanoplasmonic waveguide devices which satisfy a set of desired specifications. Space mapping utilizes a physics-based coarse model to approximate a fine model accurately describing a device. Here the fine model is a full-wave finite-difference frequency-domain (FDFD) simulation of the device, while the coarse model is based on transmission line theory. We demonstrate that simply optimizing the transmission line model of the device is not enough to obtain a device which satisfies all the required design specifications. On the other hand, when the iterative space mapping algorithm is used, it converges fast to a design which meets all the specifications. In addition, full-wave FDFD simulations of only a few candidate structures are required before the iterative process is terminated. Use of the space mapping algorithm therefore results in large reductions in the required computation time when compared to any direct optimization method of the fine FDFD model.

  18. [Studies of marker screening efficiency and corresponding influencing factors in QTL composite interval mapping].

    PubMed

    Gao, Yong-Ming; Wan, Ping

    2002-06-01

    Screening markers efficiently is the foundation of mapping QTLs by composite interval mapping. Main and interaction markers distinguished, besides using background control for genetic variation, could also be used to construct intervals of two-way searching for mapping QTLs with epistasis, which can save a lot of calculation time. Therefore, the efficiency of marker screening would affect power and precision of QTL mapping. A doubled haploid population with 200 individuals and 5 chromosomes was constructed, with 50 markers evenly distributed at 10 cM space. Among a total of 6 QTLs, one was placed on chromosome I, two linked on chromosome II, and the other three linked on chromosome IV. QTL setting included additive effects and epistatic effects of additive x additive, the corresponding QTL interaction effects were set if data were collected under multiple environments. The heritability was assumed to be 0.5 if no special declaration. The power of marker screening by stepwise regression, forward regression, and three methods for random effect prediction, e.g. best linear unbiased prediction (BLUP), linear unbiased prediction (LUP) and adjusted unbiased prediction (AUP), was studied and compared through 100 Monte Carlo simulations. The results indicated that the marker screening power by stepwise regression at 0.1, 0.05 and 0.01 significant level changed from 2% to 68%, the power changed from 2% to 72% by forward regression. The larger the QTL effects, the higher the marker screening power. While the power of marker screening by three random effect prediction was very low, the maximum was only 13%. That suggested that regression methods were much better than those by using the approaches of random effect prediction to identify efficient markers flanking QTLs, and forward selection method was more simple and efficient. The results of simulation study on heritability showed that heightening of both general heritability and interaction heritability of genotype x

  19. Collection efficiency and acceptance maps of electron detectors for understanding signal detection on modern scanning electron microscopy.

    PubMed

    Agemura, Toshihide; Sekiguchi, Takashi

    2018-02-01

    Collection efficiency and acceptance maps of typical detectors in modern scanning electron microscopes (SEMs) were investigated. Secondary and backscattered electron trajectories from a specimen to through-the-lens and under-the-lens detectors placed on an electron optical axis and an Everhart-Thornley detector mounted on a specimen chamber were simulated three-dimensionally. The acceptance maps were drawn as the relationship between the energy and angle of collected electrons under different working distances. The collection efficiency considering the detector sensitivity was also estimated for the various working distances. These data indicated that the acceptance maps and collection efficiency are keys to understand the detection mechanism and image contrast for each detector in the modern SEMs. Furthermore, the working distance is the dominant parameter because electron trajectories are drastically changed with the working distance.

  20. The Pedagogy and Problems of Jane Andrews's "The Seven Little Sisters Who Live on the Round Ball that Floats in the Air" (1861)

    ERIC Educational Resources Information Center

    Tedesco, Laureen

    2006-01-01

    This essay examines the interplay between Jane Andrews's purpose and her pedagogy in "The Seven Little Sisters Who Live on the Round Ball that Floats in the Air." The book demonstrates the teaching strategies she learned at the First State Normal in Massachusetts, moving from what the child knows to new material, engaging the child in…

  1. Jane Austen's lifelong health problems and final illness: New evidence points to a fatal Hodgkin's disease and excludes the widely accepted Addison's.

    PubMed

    Upfal, A

    2005-06-01

    Jane Austen is typically described as having excellent health until the age of 40 and the onset of a mysterious and fatal illness, initially identified by Sir Zachary Cope in 1964 as Addison's disease. Her biographers, deceived both by Cassandra Austen's destruction of letters containing medical detail, and the cheerful high spirits of the existing letters, have seriously underestimated the extent to which illness affected Austen's life. A medical history reveals that she was particularly susceptible to infection, and suffered unusually severe infective illnesses, as well as a chronic conjunctivitis that impeded her ability to write. There is evidence that Austen was already suffering from an immune deficiency and fatal lymphoma in January 1813, when her second and most popular novel, Pride and Prejudice, was published. Four more novels would follow, written or revised in the shadow of her increasing illness and debility. Whilst it is impossible now to conclusively establish the cause of her death, the existing medical evidence tends to exclude Addison's disease, and suggests there is a high possibility that Jane Austen's fatal illness was Hodgkin's disease, a form of lymphoma.

  2. Efficient Posterior Probability Mapping Using Savage-Dickey Ratios

    PubMed Central

    Penny, William D.; Ridgway, Gerard R.

    2013-01-01

    Statistical Parametric Mapping (SPM) is the dominant paradigm for mass-univariate analysis of neuroimaging data. More recently, a Bayesian approach termed Posterior Probability Mapping (PPM) has been proposed as an alternative. PPM offers two advantages: (i) inferences can be made about effect size thus lending a precise physiological meaning to activated regions, (ii) regions can be declared inactive. This latter facility is most parsimoniously provided by PPMs based on Bayesian model comparisons. To date these comparisons have been implemented by an Independent Model Optimization (IMO) procedure which separately fits null and alternative models. This paper proposes a more computationally efficient procedure based on Savage-Dickey approximations to the Bayes factor, and Taylor-series approximations to the voxel-wise posterior covariance matrices. Simulations show the accuracy of this Savage-Dickey-Taylor (SDT) method to be comparable to that of IMO. Results on fMRI data show excellent agreement between SDT and IMO for second-level models, and reasonable agreement for first-level models. This Savage-Dickey test is a Bayesian analogue of the classical SPM-F and allows users to implement model comparison in a truly interactive manner. PMID:23533640

  3. Improving transmission efficiency of large sequence alignment/map (SAM) files.

    PubMed

    Sakib, Muhammad Nazmus; Tang, Jijun; Zheng, W Jim; Huang, Chin-Tser

    2011-01-01

    Research in bioinformatics primarily involves collection and analysis of a large volume of genomic data. Naturally, it demands efficient storage and transfer of this huge amount of data. In recent years, some research has been done to find efficient compression algorithms to reduce the size of various sequencing data. One way to improve the transmission time of large files is to apply a maximum lossless compression on them. In this paper, we present SAMZIP, a specialized encoding scheme, for sequence alignment data in SAM (Sequence Alignment/Map) format, which improves the compression ratio of existing compression tools available. In order to achieve this, we exploit the prior knowledge of the file format and specifications. Our experimental results show that our encoding scheme improves compression ratio, thereby reducing overall transmission time significantly.

  4. Enhancements to Demilitarization Process Maps Program (ProMap)

    DTIC Science & Technology

    2016-10-14

    map tool, ProMap, was improved by implementing new features, and sharing data with MIDAS and AMDIT databases . Specifically, process efficiency was...improved by 1) providing access to APE information contained in the AMDIT database directly from inside ProMap when constructing a process map, 2...what equipment can be efficiently used to demil a particular munition. Associated with this task was the upgrade of the AMDIT database so that

  5. An Efficient Algorithm for Mapping Imaging Data to 3D Unstructured Grids in Computational Biomechanics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Einstein, Daniel R.; Kuprat, Andrew P.; Jiao, Xiangmin

    2013-01-01

    Geometries for organ scale and multiscale simulations of organ function are now routinely derived from imaging data. However, medical images may also contain spatially heterogeneous information other than geometry that are relevant to such simulations either as initial conditions or in the form of model parameters. In this manuscript, we present an algorithm for the efficient and robust mapping of such data to imaging based unstructured polyhedral grids in parallel. We then illustrate the application of our mapping algorithm to three different mapping problems: 1) the mapping of MRI diffusion tensor data to an unstuctured ventricular grid; 2) the mappingmore » of serial cyro-section histology data to an unstructured mouse brain grid; and 3) the mapping of CT-derived volumetric strain data to an unstructured multiscale lung grid. Execution times and parallel performance are reported for each case.« less

  6. Efficient mapping algorithms for scheduling robot inverse dynamics computation on a multiprocessor system

    NASA Technical Reports Server (NTRS)

    Lee, C. S. G.; Chen, C. L.

    1989-01-01

    Two efficient mapping algorithms for scheduling the robot inverse dynamics computation consisting of m computational modules with precedence relationship to be executed on a multiprocessor system consisting of p identical homogeneous processors with processor and communication costs to achieve minimum computation time are presented. An objective function is defined in terms of the sum of the processor finishing time and the interprocessor communication time. The minimax optimization is performed on the objective function to obtain the best mapping. This mapping problem can be formulated as a combination of the graph partitioning and the scheduling problems; both have been known to be NP-complete. Thus, to speed up the searching for a solution, two heuristic algorithms were proposed to obtain fast but suboptimal mapping solutions. The first algorithm utilizes the level and the communication intensity of the task modules to construct an ordered priority list of ready modules and the module assignment is performed by a weighted bipartite matching algorithm. For a near-optimal mapping solution, the problem can be solved by the heuristic algorithm with simulated annealing. These proposed optimization algorithms can solve various large-scale problems within a reasonable time. Computer simulations were performed to evaluate and verify the performance and the validity of the proposed mapping algorithms. Finally, experiments for computing the inverse dynamics of a six-jointed PUMA-like manipulator based on the Newton-Euler dynamic equations were implemented on an NCUBE/ten hypercube computer to verify the proposed mapping algorithms. Computer simulation and experimental results are compared and discussed.

  7. An optimization method of VON mapping for energy efficiency and routing in elastic optical networks

    NASA Astrophysics Data System (ADS)

    Liu, Huanlin; Xiong, Cuilian; Chen, Yong; Li, Changping; Chen, Derun

    2018-03-01

    To improve resources utilization efficiency, network virtualization in elastic optical networks has been developed by sharing the same physical network for difference users and applications. In the process of virtual nodes mapping, longer paths between physical nodes will consume more spectrum resources and energy. To address the problem, we propose a virtual optical network mapping algorithm called genetic multi-objective optimize virtual optical network mapping algorithm (GM-OVONM-AL), which jointly optimizes the energy consumption and spectrum resources consumption in the process of virtual optical network mapping. Firstly, a vector function is proposed to balance the energy consumption and spectrum resources by optimizing population classification and crowding distance sorting. Then, an adaptive crossover operator based on hierarchical comparison is proposed to improve search ability and convergence speed. In addition, the principle of the survival of the fittest is introduced to select better individual according to the relationship of domination rank. Compared with the spectrum consecutiveness-opaque virtual optical network mapping-algorithm and baseline-opaque virtual optical network mapping algorithm, simulation results show the proposed GM-OVONM-AL can achieve the lowest bandwidth blocking probability and save the energy consumption.

  8. Effective and efficient agricultural drainage pipe mapping with UAS thermal infrared imagery: a case study

    USDA-ARS?s Scientific Manuscript database

    Effective and efficient methods are needed to map agricultural subsurface drainage systems. Visible (VIS), near infrared (NIR), and/or thermal infrared (TIR) imagery obtained by unmanned aircraft systems (UAS) may provide a means for determining drainage pipe locations. Preliminary UAS surveys wit...

  9. Low Cost and Efficient 3d Indoor Mapping Using Multiple Consumer Rgb-D Cameras

    NASA Astrophysics Data System (ADS)

    Chen, C.; Yang, B. S.; Song, S.

    2016-06-01

    Driven by the miniaturization, lightweight of positioning and remote sensing sensors as well as the urgent needs for fusing indoor and outdoor maps for next generation navigation, 3D indoor mapping from mobile scanning is a hot research and application topic. The point clouds with auxiliary data such as colour, infrared images derived from 3D indoor mobile mapping suite can be used in a variety of novel applications, including indoor scene visualization, automated floorplan generation, gaming, reverse engineering, navigation, simulation and etc. State-of-the-art 3D indoor mapping systems equipped with multiple laser scanners product accurate point clouds of building interiors containing billions of points. However, these laser scanner based systems are mostly expensive and not portable. Low cost consumer RGB-D Cameras provides an alternative way to solve the core challenge of indoor mapping that is capturing detailed underlying geometry of the building interiors. Nevertheless, RGB-D Cameras have a very limited field of view resulting in low efficiency in the data collecting stage and incomplete dataset that missing major building structures (e.g. ceilings, walls). Endeavour to collect a complete scene without data blanks using single RGB-D Camera is not technic sound because of the large amount of human labour and position parameters need to be solved. To find an efficient and low cost way to solve the 3D indoor mapping, in this paper, we present an indoor mapping suite prototype that is built upon a novel calibration method which calibrates internal parameters and external parameters of multiple RGB-D Cameras. Three Kinect sensors are mounted on a rig with different view direction to form a large field of view. The calibration procedure is three folds: 1, the internal parameters of the colour and infrared camera inside each Kinect are calibrated using a chess board pattern, respectively; 2, the external parameters between the colour and infrared camera inside each

  10. An efficient chaotic maps-based authentication and key agreement scheme using smartcards for telecare medicine information systems.

    PubMed

    Lee, Tian-Fu

    2013-12-01

    A smartcard-based authentication and key agreement scheme for telecare medicine information systems enables patients, doctors, nurses and health visitors to use smartcards for secure login to medical information systems. Authorized users can then efficiently access remote services provided by the medicine information systems through public networks. Guo and Chang recently improved the efficiency of a smartcard authentication and key agreement scheme by using chaotic maps. Later, Hao et al. reported that the scheme developed by Guo and Chang had two weaknesses: inability to provide anonymity and inefficient double secrets. Therefore, Hao et al. proposed an authentication scheme for telecare medicine information systems that solved these weaknesses and improved performance. However, a limitation in both schemes is their violation of the contributory property of key agreements. This investigation discusses these weaknesses and proposes a new smartcard-based authentication and key agreement scheme that uses chaotic maps for telecare medicine information systems. Compared to conventional schemes, the proposed scheme provides fewer weaknesses, better security, and more efficiency.

  11. The Jane Dent Home: the rise and fall of homes for the aged in low-income communities.

    PubMed

    Reed, Susan C; Davis, Nancy

    2004-11-01

    The Jane Dent Home was established in 1898 (as the Home for Aged and Infirm Colored People) to serve African American elderly barred from admission to most homes for the aged. Sustained by community leadership through difficult times, the Home finally closed in 1975 after growing and persistent racial and economic segregation of Chicago's low-income neighborhoods combined with pressure from state government to ensure fire safety. This history illustrates the decline of not-for-profit homes for the aged while for-profit nursing homes were capturing market share. In Chicago this trend is strongest in low-income communities of color, which may lead to lower quality of care for such communities. Support for indigenous not-for-profit long-term care may promote the goals of health care equity articulated by Healthy People 2010.

  12. Mary Jane Hogue (1883-1962): A pioneer in human brain tissue culture.

    PubMed

    Zottoli, Steven J; Seyfarth, Ernst-August

    2018-05-16

    The ability to maintain human brain explants in tissue culture was a critical step in the use of these cells for the study of central nervous system disorders. Ross G. Harrison (1870-1959) was the first to successfully maintain frog medullary tissue in culture in 1907, but it took another 38 years before successful culture of human brain tissue was accomplished. One of the pioneers in this achievement was Mary Jane Hogue (1883-1962). Hogue was born into a Quaker family in 1883 in West Chester, Pennsylvania, and received her undergraduate degree from Goucher College in Baltimore, Maryland. Research with the developmental biologist Theodor Boveri (1862-1915) in Würzburg, Germany, resulted in her Ph.D. (1909). Hogue transitioned from studying protozoa to the culture of human brain tissue in the 1940s and 1950s, when she was one of the first to culture cells from human fetal, infant, and adult brain explants. We review Hogue's pioneering contributions to the study of human brain cells in culture, her putative identification of progenitor neuroblast and/or glioblast cells, and her use of the cultures to study the cytopathogenic effects of poliovirus. We also put Hogue's work in perspective by discussing how other women pioneers in tissue culture influenced Hogue and her research.

  13. Behavioural Susceptibility Theory: Professor Jane Wardle and the Role of Appetite in Genetic Risk of Obesity.

    PubMed

    Llewellyn, Clare H; Fildes, Alison

    2017-03-01

    There is considerable variability in human body weight, despite the ubiquity of the 'obesogenic' environment. Human body weight has a strong genetic basis and it has been hypothesised that genetic susceptibility to the environment explains variation in human body weight, with differences in appetite being implicated as the mediating mechanism; so-called 'behavioural susceptibility theory' (BST), first described by Professor Jane Wardle. This review summarises the evidence for the role of appetite as a mediator of genetic risk of obesity. Variation in appetitive traits is observable from infancy, drives early weight gain and is highly heritable in infancy and childhood. Obesity-related common genetic variants identified through genome-wide association studies show associations with appetitive traits, and appetite mediates part of the observed association between genetic risk and adiposity. Obesity results from an interaction between genetic susceptibility to overeating and exposure to an 'obesogenic' food environment.

  14. Comparing the efficiency of digital and conventional soil mapping to predict soil types in a semi-arid region in Iran

    NASA Astrophysics Data System (ADS)

    Zeraatpisheh, Mojtaba; Ayoubi, Shamsollah; Jafari, Azam; Finke, Peter

    2017-05-01

    The efficiency of different digital and conventional soil mapping approaches to produce categorical maps of soil types is determined by cost, sample size, accuracy and the selected taxonomic level. The efficiency of digital and conventional soil mapping approaches was examined in the semi-arid region of Borujen, central Iran. This research aimed to (i) compare two digital soil mapping approaches including Multinomial logistic regression and random forest, with the conventional soil mapping approach at four soil taxonomic levels (order, suborder, great group and subgroup levels), (ii) validate the predicted soil maps by the same validation data set to determine the best method for producing the soil maps, and (iii) select the best soil taxonomic level by different approaches at three sample sizes (100, 80, and 60 point observations), in two scenarios with and without a geomorphology map as a spatial covariate. In most predicted maps, using both digital soil mapping approaches, the best results were obtained using the combination of terrain attributes and the geomorphology map, although differences between the scenarios with and without the geomorphology map were not significant. Employing the geomorphology map increased map purity and the Kappa index, and led to a decrease in the 'noisiness' of soil maps. Multinomial logistic regression had better performance at higher taxonomic levels (order and suborder levels); however, random forest showed better performance at lower taxonomic levels (great group and subgroup levels). Multinomial logistic regression was less sensitive than random forest to a decrease in the number of training observations. The conventional soil mapping method produced a map with larger minimum polygon size because of traditional cartographic criteria used to make the geological map 1:100,000 (on which the conventional soil mapping map was largely based). Likewise, conventional soil mapping map had also a larger average polygon size that resulted

  15. Robust and efficient biometrics based password authentication scheme for telecare medicine information systems using extended chaotic maps.

    PubMed

    Lu, Yanrong; Li, Lixiang; Peng, Haipeng; Xie, Dong; Yang, Yixian

    2015-06-01

    The Telecare Medicine Information Systems (TMISs) provide an efficient communicating platform supporting the patients access health-care delivery services via internet or mobile networks. Authentication becomes an essential need when a remote patient logins into the telecare server. Recently, many extended chaotic maps based authentication schemes using smart cards for TMISs have been proposed. Li et al. proposed a secure smart cards based authentication scheme for TMISs using extended chaotic maps based on Lee's and Jiang et al.'s scheme. In this study, we show that Li et al.'s scheme has still some weaknesses such as violation the session key security, vulnerability to user impersonation attack and lack of local verification. To conquer these flaws, we propose a chaotic maps and smart cards based password authentication scheme by applying biometrics technique and hash function operations. Through the informal and formal security analyses, we demonstrate that our scheme is resilient possible known attacks including the attacks found in Li et al.'s scheme. As compared with the previous authentication schemes, the proposed scheme is more secure and efficient and hence more practical for telemedical environments.

  16. "Secrets of the female sex": Jane Sharp, the reproductive female body, and early modern midwifery manuals.

    PubMed

    Hobby, E

    2001-01-01

    Early modern midwifery manuals in Britain were usually the work of men. These books were a significant source of information about the body to the wider reading public: many sold well, and their prefatory materials include injunctions to readers not to make improper use of them. What is particularly interesting about Jane Sharp's Midwives Book (1671) is that it both provides a compendium of current beliefs concerning reproduction, and indicates the author's ironic perception of the misogyny that underpinned accepted ideas about the female reproductive body. This article gives key examples of Sharp's interventions, and also refers to Thomas Bartholin, Bartholinus Anatomy (1688); Richard Bunworth, The Doctresse (1656); Hugh Chamberlen, The Accomplisht Midwife (1673); The Compleat Midwifes Practice (1656); Helkiah Crooke, Microcosmographia (1615); Nicholas Culpeper, A Directory for Midwives (1651); Jacques Guillemeau, Childbirth (1612); Jean Riolan, A Sure Guide (1657); Daniel Sennert, Practical Physick (1664); William Sermon, The Ladies Companion (1671); and Percival Willughby, Observations in Midwifery (c. 1675).

  17. The Stories We Hear, the Stories We Tell What Can the Life of Jane Barker (1652-1732) Tell Us about Women's Leadership in Higher Education in the Twenty-First Century?

    ERIC Educational Resources Information Center

    Wilson, Carol Shiner

    2009-01-01

    Jane Barker--poet, novelist, farm manager, student and practitioner of medical arts--was not allowed to attend university because she was a woman. Yet she was Oxford-educated in the most modern of medical theories of her time. By the end of her life, unmarried by choice, Barker was writing for pay under her own name in an emerging genre--the…

  18. A High-Density Genetic Map Identifies a Novel Major QTL for Boron Efficiency in Oilseed Rape (Brassica napus L.)

    PubMed Central

    Wang, Xiaohua; Zhao, Hua; Shi, Lei; Xu, Fangsen

    2014-01-01

    Low boron (B) seriously limits the growth of oilseed rape (Brassica napus L.), a high B demand species that is sensitive to low B conditions. Significant genotypic variations in response to B deficiency have been observed among B. napus cultivars. To reveal the genetic basis for B efficiency in B. napus, quantitative trait loci (QTLs) for the plant growth traits, B uptake traits and the B efficiency coefficient (BEC) were analyzed using a doubled haploid (DH) population derived from a cross between a B-efficient parent, Qingyou 10, and a B-inefficient parent, Westar 10. A high-density genetic map was constructed based on single nucleotide polymorphisms (SNPs) assayed using Brassica 60 K Infinium BeadChip Array, simple sequence repeats (SSRs) and amplified fragment length polymorphisms (AFLPs). The linkage map covered a total length of 2139.5 cM, with 19 linkage groups (LGs) and an average distance of 1.6 cM between adjacent markers. Based on hydroponic evaluation of six B efficiency traits measured in three separate repeated trials, a total of 52 QTLs were identified, accounting for 6.14–46.27% of the phenotypic variation. A major QTL for BEC, qBEC-A3a, was co-located on A3 with other QTLs for plant growth and B uptake traits under low B stress. Using a subset of substitution lines, qBEC-A3a was validated and narrowed down to the interval between CNU384 and BnGMS436. The results of this study provide a novel major locus located on A3 for B efficiency in B. napus that will be suitable for fine mapping and marker-assisted selection breeding for B efficiency in B. napus. PMID:25375356

  19. "A hint of it, with initials": adultery, textuality and publicity in Jane Austen's Lady Susan.

    PubMed

    Russell, Gillian

    2010-01-01

    In spite of Jane Austen's professed “eye” for an adulteress, comparatively little attention has been paid to adultery and divorce as themes and contexts of her fiction. Her unpublished epistolary novel Lady Susan has a distinctive status in Austen's oeuvre, recognized as being exemplary of her “style” and yet atypical of her later achievement. A neglected context for the novel is the extensive reporting of adultery trials in contemporary print culture and the moral panic concerning adultery in the 1780s and 1790s, focusing initially on the adulteress as the brazen woman of fashion and later as a figure of sentimentalized abjection. A particularly notorious case, that involving Lady Henrietta Grosvenor and George III's brother, the Duke of Cumberland, is directly alluded to in Lady Susan. The textual strategies of adultery trial literature, particularly its emphasis on indirection through the use of detail or “hint”, had a long-term influence on the development of Austen's fiction and her positioning of herself as a professional writer after the 1790s.

  20. Implementation of efficient trajectories for an ultrasonic scanner using chaotic maps

    NASA Astrophysics Data System (ADS)

    Almeda, A.; Baltazar, A.; Treesatayapun, C.; Mijarez, R.

    2012-05-01

    Typical ultrasonic methodology for nondestructive scanning evaluation uses systematic scanning paths. In many cases, this approach is time inefficient and also energy and computational power consuming. Here, a methodology for the scanning of defects using an ultrasonic echo-pulse scanning technique combined with chaotic trajectory generation is proposed. This is implemented in a Cartesian coordinate robotic system developed in our lab. To cover the entire search area, a chaotic function and a proposed mirror mapping were incorporated. To improve detection probability, our proposed scanning methodology is complemented with a probabilistic approach of discontinuity detection. The developed methodology was found to be more efficient than traditional ones used to localize and characterize hidden flaws.

  1. Map reading tools for map libraries.

    USGS Publications Warehouse

    Greenberg, G.L.

    1982-01-01

    Engineers, navigators and military strategists employ a broad array of mechanical devices to facilitate map use. A larger number of map users such as educators, students, tourists, journalists, historians, politicians, economists and librarians are unaware of the available variety of tools which can be used with maps to increase the speed and efficiency of their application and interpretation. This paper identifies map reading tools such as coordinate readers, protractors, dividers, planimeters, and symbol-templets according to a functional classification. Particularly, arrays of tools are suggested for use in determining position, direction, distance, area and form (perimeter-shape-pattern-relief). -from Author

  2. Managing Vocabulary Mapping Services

    PubMed Central

    Che, Chengjian; Monson, Kent; Poon, Kasey B.; Shakib, Shaun C.; Lau, Lee Min

    2005-01-01

    The efficient management and maintenance of large-scale and high-quality vocabulary mapping is an operational challenge. The 3M Health Information Systems (HIS) Healthcare Data Dictionary (HDD) group developed an information management system to provide controlled mapping services, resulting in improved efficiency and quality maintenance. PMID:16779203

  3. Internal quantum efficiency mapping analysis for a >20%-efficiency n-type bifacial solar cell with front-side emitter formed by BBr3 thermal diffusion

    NASA Astrophysics Data System (ADS)

    Simayi, Shalamujiang; Mochizuki, Toshimitsu; Kida, Yasuhiro; Shirasawa, Katsuhiko; Takato, Hidetaka

    2017-10-01

    This paper presents a large-area (239-cm2) high-efficiency n-type bifacial solar cell that is processed using tube-furnace thermal diffusion employing liquid sources BBr3 for the front-side boron emitter and POCl3 for the rear-side phosphorus back surface field (BSF). The SiN x /Al2O3 stack was applied to the front-side boron emitter as a passivation layer. Both the front and rear-side electrodes are obtained using screen-printed contacts with H-patterns. The resulting highest-efficiency solar cell has front- and rear-side efficiencies of 20.3 and 18.7%, respectively, while the corresponding bifaciality is up to 92%. Finally, the passivation quality of the SiN x /Al2O3 stack on the front-side boron emitter and rear-side phosphorus BSF is investigated and visualized by measuring the internal quantum efficiency mapping of the bifacial solar cell.

  4. Topographic mapping

    USGS Publications Warehouse

    ,

    2008-01-01

    The U.S. Geological Survey (USGS) produced its first topographic map in 1879, the same year it was established. Today, more than 100 years and millions of map copies later, topographic mapping is still a central activity for the USGS. The topographic map remains an indispensable tool for government, science, industry, and leisure. Much has changed since early topographers traveled the unsettled West and carefully plotted the first USGS maps by hand. Advances in survey techniques, instrumentation, and design and printing technologies, as well as the use of aerial photography and satellite data, have dramatically improved mapping coverage, accuracy, and efficiency. Yet cartography, the art and science of mapping, may never before have undergone change more profound than today.

  5. A combination of selected mapping and clipping to increase energy efficiency of OFDM systems

    PubMed Central

    Lee, Byung Moo; Rim, You Seung

    2017-01-01

    We propose an energy efficient combination design for OFDM systems based on selected mapping (SLM) and clipping peak-to-average power ratio (PAPR) reduction techniques, and show the related energy efficiency (EE) performance analysis. The combination of two different PAPR reduction techniques can provide a significant benefit in increasing EE, because it can take advantages of both techniques. For the combination, we choose the clipping and SLM techniques, since the former technique is quite simple and effective, and the latter technique does not cause any signal distortion. We provide the structure and the systematic operating method, and show the various analyzes to derive the EE gain based on the combined technique. Our analysis show that the combined technique increases the EE by 69% compared to no PAPR reduction, and by 19.34% compared to only using SLM technique. PMID:29023591

  6. Pressure Mapping and Efficiency Analysis of an EPPLER 857 Hydrokinetic Turbine

    NASA Astrophysics Data System (ADS)

    Clark, Tristan

    A conceptual energy ship is presented to provide renewable energy. The ship, driven by the wind, drags a hydrokinetic turbine through the water. The power generated is used to run electrolysis on board, taking the resultant hydrogen back to shore to be used as an energy source. The basin efficiency (Power/thrust*velocity) of the Hydrokinetic Turbine (HTK) plays a vital role in this process. In order to extract the maximum allowable power from the flow, the blades need to be optimized. The structural analysis of the blade is important, as the blade will undergo high pressure loads from the water. A procedure for analysis of a preliminary Hydrokinetic Turbine blade design is developed. The blade was designed by a non-optimized Blade Element Momentum Theory (BEMT) code. Six simulations were run, with varying mesh resolution, turbulence models, and flow region size. The procedure was developed that provides detailed explanation for the entire process, from geometry and mesh generation to post-processing analysis tools. The efficiency results from the simulations are used to study the mesh resolution, flow region size, and turbulence models. The results are compared to the BEMT model design targets. Static pressure maps are created that can be used for structural analysis of the blades.

  7. Reliable Radiation Hybrid Maps: An Efficient Scalable Clustering-based Approach

    USDA-ARS?s Scientific Manuscript database

    The process of mapping markers from radiation hybrid mapping (RHM) experiments is equivalent to the traveling salesman problem and, thereby, has combinatorial complexity. As an additional problem, experiments typically result in some unreliable markers that reduce the overall quality of the map. We ...

  8. Efficiently mapping structure-property relationships of gas adsorption in porous materials: application to Xe adsorption.

    PubMed

    Kaija, A R; Wilmer, C E

    2017-09-08

    Designing better porous materials for gas storage or separations applications frequently leverages known structure-property relationships. Reliable structure-property relationships, however, only reveal themselves when adsorption data on many porous materials are aggregated and compared. Gathering enough data experimentally is prohibitively time consuming, and even approaches based on large-scale computer simulations face challenges. Brute force computational screening approaches that do not efficiently sample the space of porous materials may be ineffective when the number of possible materials is too large. Here we describe a general and efficient computational method for mapping structure-property spaces of porous materials that can be useful for adsorption related applications. We describe an algorithm that generates random porous "pseudomaterials", for which we calculate structural characteristics (e.g., surface area, pore size and void fraction) and also gas adsorption properties via molecular simulations. Here we chose to focus on void fraction and Xe adsorption at 1 bar, 5 bar, and 10 bar. The algorithm then identifies pseudomaterials with rare combinations of void fraction and Xe adsorption and mutates them to generate new pseudomaterials, thereby selectively adding data only to those parts of the structure-property map that are the least explored. Use of this method can help guide the design of new porous materials for gas storage and separations applications in the future.

  9. Engineering With Nature Geographic Project Mapping Tool (EWN ProMap)

    DTIC Science & Technology

    2015-07-01

    EWN ProMap database provides numerous case studies for infrastructure projects such as breakwaters, river engineering dikes, and seawalls that have...the EWN Project Mapping Tool (EWN ProMap) is to assist users in their search for case study information that can be valuable for developing EWN ideas...Essential elements of EWN include: (1) using science and engineering to produce operational efficiencies supporting sustainable delivery of

  10. A Spatiotemporal Indexing Approach for Efficient Processing of Big Array-Based Climate Data with MapReduce

    NASA Technical Reports Server (NTRS)

    Li, Zhenlong; Hu, Fei; Schnase, John L.; Duffy, Daniel Q.; Lee, Tsengdar; Bowen, Michael K.; Yang, Chaowei

    2016-01-01

    Climate observations and model simulations are producing vast amounts of array-based spatiotemporal data. Efficient processing of these data is essential for assessing global challenges such as climate change, natural disasters, and diseases. This is challenging not only because of the large data volume, but also because of the intrinsic high-dimensional nature of geoscience data. To tackle this challenge, we propose a spatiotemporal indexing approach to efficiently manage and process big climate data with MapReduce in a highly scalable environment. Using this approach, big climate data are directly stored in a Hadoop Distributed File System in its original, native file format. A spatiotemporal index is built to bridge the logical array-based data model and the physical data layout, which enables fast data retrieval when performing spatiotemporal queries. Based on the index, a data-partitioning algorithm is applied to enable MapReduce to achieve high data locality, as well as balancing the workload. The proposed indexing approach is evaluated using the National Aeronautics and Space Administration (NASA) Modern-Era Retrospective Analysis for Research and Applications (MERRA) climate reanalysis dataset. The experimental results show that the index can significantly accelerate querying and processing (10 speedup compared to the baseline test using the same computing cluster), while keeping the index-to-data ratio small (0.0328). The applicability of the indexing approach is demonstrated by a climate anomaly detection deployed on a NASA Hadoop cluster. This approach is also able to support efficient processing of general array-based spatiotemporal data in various geoscience domains without special configuration on a Hadoop cluster.

  11. Efficient search, mapping, and optimization of multi-protein genetic systems in diverse bacteria

    PubMed Central

    Farasat, Iman; Kushwaha, Manish; Collens, Jason; Easterbrook, Michael; Guido, Matthew; Salis, Howard M

    2014-01-01

    Developing predictive models of multi-protein genetic systems to understand and optimize their behavior remains a combinatorial challenge, particularly when measurement throughput is limited. We developed a computational approach to build predictive models and identify optimal sequences and expression levels, while circumventing combinatorial explosion. Maximally informative genetic system variants were first designed by the RBS Library Calculator, an algorithm to design sequences for efficiently searching a multi-protein expression space across a > 10,000-fold range with tailored search parameters and well-predicted translation rates. We validated the algorithm's predictions by characterizing 646 genetic system variants, encoded in plasmids and genomes, expressed in six gram-positive and gram-negative bacterial hosts. We then combined the search algorithm with system-level kinetic modeling, requiring the construction and characterization of 73 variants to build a sequence-expression-activity map (SEAMAP) for a biosynthesis pathway. Using model predictions, we designed and characterized 47 additional pathway variants to navigate its activity space, find optimal expression regions with desired activity response curves, and relieve rate-limiting steps in metabolism. Creating sequence-expression-activity maps accelerates the optimization of many protein systems and allows previous measurements to quantitatively inform future designs. PMID:24952589

  12. Pragmatist Epistemology and Jane Addams: Fundamental Concepts for the Social Paradigm of Occupational Therapy.

    PubMed

    Morrison, Rodolfo

    2016-12-01

    The objective of this manuscript is to contribute to the education of future occupational therapists within the current paradigm of the profession. To this purpose, some of the conceptual foundations of pragmatist epistemology and philosophical contributions of the philosopher Jane Addams are presented. Some pragmatist fundamentals such as the holistic vision of the human being, the paradox of separating "knowing from doing", and giving the usefulness of the activity and knowledge primacy in the process of human development, inspired the profession of Occupational Therapy in the United States at the beginning of the twentieth century. Today, almost 100 years after the founding of the National Society for the Promotion of Occupational Therapy, pragmatism is still relevant to the profession. Specifically, its pertinence is related to the current scenario of the profession, and its powerful development in working scopes related to socio-community fields. This helps identify how we are on the verge of another paradigm, known as Social Paradigm of the Occupation. This new social understanding of the discipline allows us to understand the relevance of professional work in community or social contexts. Future research could address how the pragmatism contributes to the understanding of occupation as a social phenomenon within this new paradigm. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  13. Interjections in the performance of Jane Austen's Pride and prejudice.

    PubMed

    O'Connell, Daniel C; Kowal, Sabine

    2010-08-01

    Three data sets of primary and secondary interjections were compared: (1) the original interjections written into the text of Jane Austen's (1813/1994) novel Pride and prejudice; (2) the interjections read aloud in commercial recordings by six professional readers of the entire text of the novel; (3) the interjections spoken by actresses and actors in the film production (Birtwistle and Langton in Pride and prejudice [TV Mini-series]. London: BBC TV, 1995) whose script, despite modest selectiveness, adheres most closely of all film versions to Austen's original text. Overall, the respective frequencies of occurrence of interjections were 136 < 141 < 398. Of the 136 interjections in Austen's printed text, 96% were attributable to women's roles, particularly Elizabeth Bennet and her mother. The second of these figures (141) is an average across all six readers. Hence, readers added a very modest number of interjections. But the actresses and actors added a large number of interjections. The dramatic oral expressiveness of the film performance is largely carried by and reflected in the actresses' and to a lesser extent in the actors' use of these primary interjections. These findings can well be related to Nübling's (Zeitschrift für Semiotik 26:11-45, 2004, Duden: Die Grammatik (pp 573-640). Mannheim: Dudenverlag, 2005) hypothesis of a spectrum of interjectional expressivity. But Ameka's (J Pragmat 18:101-118, 1992) linguistic hypothesis that pauses will both precede and follow interjections was once again found to be empirically groundless. A large percentage (96%) of the interjections in the film performance served the function of initializing various units of discourse, either after a pause before articulatory phrases, or before a sentence and/or turn. Both the emotional and initiating functions of interjections are characteristic of conceptual and medial orality rather than of conceptual and medial literacy. Accordingly, their usage throws further light on a

  14. Efficient crop type mapping based on remote sensing in the Central Valley, California

    NASA Astrophysics Data System (ADS)

    Zhong, Liheng

    Most agricultural systems in California's Central Valley are purposely flexible and intentionally designed to meet the demands of dynamic markets. Agricultural land use is also impacted by climate change and urban development. As a result, crops change annually and semiannually, which makes estimating agricultural water use difficult, especially given the existing method by which agricultural land use is identified and mapped. A minor portion of agricultural land is surveyed annually for land-use type, and every 5 to 8 years the entire valley is completely evaluated. So far no effort has been made to effectively and efficiently identify specific crop types on an annual basis in this area. The potential of satellite imagery to map agricultural land cover and estimate water usage in the Central Valley is explored. Efforts are made to minimize the cost and reduce the time of production during the mapping process. The land use change analysis shows that a remote sensing based mapping method is the only means to map the frequent change of major crop types. The traditional maximum likelihood classification approach is first utilized to map crop types to test the classification capacity of existing algorithms. High accuracy is achieved with sufficient ground truth data for training, and crop maps of moderate quality can be timely produced to facilitate a near-real-time water use estimate. However, the large set of ground truth data required by this method results in high costs in data collection. It is difficult to reduce the cost because a trained classification algorithm is not transferable between different years or different regions. A phenology based classification (PBC) approach is developed which extracts phenological metrics from annual vegetation index profiles and identifies crop types based on these metrics using decision trees. According to the comparison with traditional maximum likelihood classification, this phenology-based approach shows great advantages

  15. An efficient method for removing point sources from full-sky radio interferometric maps

    NASA Astrophysics Data System (ADS)

    Berger, Philippe; Oppermann, Niels; Pen, Ue-Li; Shaw, J. Richard

    2017-12-01

    A new generation of wide-field radio interferometers designed for 21-cm surveys is being built as drift scan instruments allowing them to observe large fractions of the sky. With large numbers of antennas and frequency channels, the enormous instantaneous data rates of these telescopes require novel, efficient, data management and analysis techniques. The m-mode formalism exploits the periodicity of such data with the sidereal day, combined with the assumption of statistical isotropy of the sky, to achieve large computational savings and render optimal analysis methods computationally tractable. We present an extension to that work that allows us to adopt a more realistic sky model and treat objects such as bright point sources. We develop a linear procedure for deconvolving maps, using a Wiener filter reconstruction technique, which simultaneously allows filtering of these unwanted components. We construct an algorithm, based on the Sherman-Morrison-Woodbury formula, to efficiently invert the data covariance matrix, as required for any optimal signal-to-noise ratio weighting. The performance of our algorithm is demonstrated using simulations of a cylindrical transit telescope.

  16. A Configurational-Bias-Monte-Carlo Back-Mapping Algorithm for Efficient and Rapid Conversion of Coarse-Grained Water Structures Into Atomistic Models.

    PubMed

    Loeffler, Troy David; Chan, Henry; Narayanan, Badri; Cherukara, Mathew J; Gray, Stephen K; Sankaranarayanan, Subramanian K R S

    2018-06-20

    Coarse-grained molecular dynamics (MD) simulations represent a powerful approach to simulate longer time scale and larger length scale phenomena than those accessible to all-atom models. The gain in efficiency, however, comes at the cost of atomistic details. The reverse transformation, also known as back-mapping, of coarse grained beads into their atomistic constituents represents a major challenge. Most existing approaches are limited to specific molecules or specific force-fields and often rely on running a long time atomistic MD of the back-mapped configuration to arrive at an optimal solution. Such approaches are problematic when dealing with systems with high diffusion barriers. Here, we introduce a new extension of the configurational-bias-Monte-Carlo (CBMC) algorithm, which we term the crystalline-configurational-bias-Monte-Carlo (C-CBMC) algortihm, that allows rapid and efficient conversion of a coarse-grained model back into its atomistic representation. Although the method is generic, we use a coarse-grained water model as a representative example and demonstrate the back-mapping or reverse transformation for model systems ranging from the ice-liquid water interface to amorphous and crystalline ice configurations. A series of simulations using the TIP4P/Ice model are performed to compare the new CBMC method to several other standard Monte Carlo and Molecular Dynamics based back-mapping techniques. In all the cases, the C-CBMC algorithm is able to find optimal hydrogen bonded configuration many thousand evaluations/steps sooner than the other methods compared within this paper. For crystalline ice structures such as a hexagonal, cubic, and cubic-hexagonal stacking disorder structures, the C-CBMC was able to find structures that were between 0.05 and 0.1 eV/water molecule lower in energy than the ground state energies predicted by the other methods. Detailed analysis of the atomistic structures show a significantly better global hydrogen positioning when

  17. Analysis of microarray leukemia data using an efficient MapReduce-based K-nearest-neighbor classifier.

    PubMed

    Kumar, Mukesh; Rath, Nitish Kumar; Rath, Santanu Kumar

    2016-04-01

    Microarray-based gene expression profiling has emerged as an efficient technique for classification, prognosis, diagnosis, and treatment of cancer. Frequent changes in the behavior of this disease generates an enormous volume of data. Microarray data satisfies both the veracity and velocity properties of big data, as it keeps changing with time. Therefore, the analysis of microarray datasets in a small amount of time is essential. They often contain a large amount of expression, but only a fraction of it comprises genes that are significantly expressed. The precise identification of genes of interest that are responsible for causing cancer are imperative in microarray data analysis. Most existing schemes employ a two-phase process such as feature selection/extraction followed by classification. In this paper, various statistical methods (tests) based on MapReduce are proposed for selecting relevant features. After feature selection, a MapReduce-based K-nearest neighbor (mrKNN) classifier is also employed to classify microarray data. These algorithms are successfully implemented in a Hadoop framework. A comparative analysis is done on these MapReduce-based models using microarray datasets of various dimensions. From the obtained results, it is observed that these models consume much less execution time than conventional models in processing big data. Copyright © 2016 Elsevier Inc. All rights reserved.

  18. Efficient Swath Mapping Laser Altimetry Demonstration Instrument Incubator Program

    NASA Technical Reports Server (NTRS)

    Yu, Anthony W.; Krainak, Michael A,; Harding, David J.; Abshire, James B.; Sun, Xiaoli; Cavanaugh, John; Valett, Susan

    2010-01-01

    In this paper we will discuss our eighteen-month progress of a three-year Instrument Incubator Program (IIP) funded by NASA Earth Science Technology Office (ESTO) on swath mapping laser altimetry system. This paper will discuss the system approach, enabling technologies and instrument concept for the swath mapping laser altimetry.

  19. Efficient evaluation of sampling quality of molecular dynamics simulations by clustering of dihedral torsion angles and Sammon mapping.

    PubMed

    Frickenhaus, Stephan; Kannan, Srinivasaraghavan; Zacharias, Martin

    2009-02-01

    A direct conformational clustering and mapping approach for peptide conformations based on backbone dihedral angles has been developed and applied to compare conformational sampling of Met-enkephalin using two molecular dynamics (MD) methods. Efficient clustering in dihedrals has been achieved by evaluating all combinations resulting from independent clustering of each dihedral angle distribution, thus resolving all conformational substates. In contrast, Cartesian clustering was unable to accurately distinguish between all substates. Projection of clusters on dihedral principal component (PCA) subspaces did not result in efficient separation of highly populated clusters. However, representation in a nonlinear metric by Sammon mapping was able to separate well the 48 highest populated clusters in just two dimensions. In addition, this approach also allowed us to visualize the transition frequencies between clusters efficiently. Significantly, higher transition frequencies between more distinct conformational substates were found for a recently developed biasing-potential replica exchange MD simulation method allowing faster sampling of possible substates compared to conventional MD simulations. Although the number of theoretically possible clusters grows exponentially with peptide length, in practice, the number of clusters is only limited by the sampling size (typically much smaller), and therefore the method is well suited also for large systems. The approach could be useful to rapidly and accurately evaluate conformational sampling during MD simulations, to compare different sampling strategies and eventually to detect kinetic bottlenecks in folding pathways.

  20. Efficient biometric authenticated key agreements based on extended chaotic maps for telecare medicine information systems.

    PubMed

    Lou, Der-Chyuan; Lee, Tian-Fu; Lin, Tsung-Hung

    2015-05-01

    Authenticated key agreements for telecare medicine information systems provide patients, doctors, nurses and health visitors with accessing medical information systems and getting remote services efficiently and conveniently through an open network. In order to have higher security, many authenticated key agreement schemes appended biometric keys to realize identification except for using passwords and smartcards. Due to too many transmissions and computational costs, these authenticated key agreement schemes are inefficient in communication and computation. This investigation develops two secure and efficient authenticated key agreement schemes for telecare medicine information systems by using biometric key and extended chaotic maps. One scheme is synchronization-based, while the other nonce-based. Compared to related approaches, the proposed schemes not only retain the same security properties with previous schemes, but also provide users with privacy protection and have fewer transmissions and lower computational cost.

  1. MAP - a mapping and analysis program for harvest planning

    Treesearch

    Robert N. Eli; Chris B. LeDoux; Penn A. Peters

    1984-01-01

    The Northeastern Forest Experiment Station and the Department of Civil Engineering at West Virginia University are cooperating in the development of a Mapping and Analysis Program, to be named MAP. The goal of this computer software package is to significantly improve the planning and harvest efficiency of small to moderately sized harvest units located in mountainous...

  2. A Round-Efficient Authenticated Key Agreement Scheme Based on Extended Chaotic Maps for Group Cloud Meeting.

    PubMed

    Lin, Tsung-Hung; Tsung, Chen-Kun; Lee, Tian-Fu; Wang, Zeng-Bo

    2017-12-03

    The security is a critical issue for business purposes. For example, the cloud meeting must consider strong security to maintain the communication privacy. Considering the scenario with cloud meeting, we apply extended chaotic map to present passwordless group authentication key agreement, termed as Passwordless Group Authentication Key Agreement (PL-GAKA). PL-GAKA improves the computation efficiency for the simple group password-based authenticated key agreement (SGPAKE) proposed by Lee et al. in terms of computing the session key. Since the extended chaotic map has equivalent security level to the Diffie-Hellman key exchange scheme applied by SGPAKE, the security of PL-GAKA is not sacrificed when improving the computation efficiency. Moreover, PL-GAKA is a passwordless scheme, so the password maintenance is not necessary. Short-term authentication is considered, hence the communication security is stronger than other protocols by dynamically generating session key in each cloud meeting. In our analysis, we first prove that each meeting member can get the correct information during the meeting. We analyze common security issues for the proposed PL-GAKA in terms of session key security, mutual authentication, perfect forward security, and data integrity. Moreover, we also demonstrate that communicating in PL-GAKA is secure when suffering replay attacks, impersonation attacks, privileged insider attacks, and stolen-verifier attacks. Eventually, an overall comparison is given to show the performance between PL-GAKA, SGPAKE and related solutions.

  3. High-efficient Extraction of Drainage Networks from Digital Elevation Model Data Constrained by Enhanced Flow Enforcement from Known River Map

    NASA Astrophysics Data System (ADS)

    Wu, T.; Li, T.; Li, J.; Wang, G.

    2017-12-01

    Improved drainage network extraction can be achieved by flow enforcement whereby information of known river maps is imposed to the flow-path modeling process. However, the common elevation-based stream burning method can sometimes cause unintended topological errors and misinterpret the overall drainage pattern. We presented an enhanced flow enforcement method to facilitate accurate and efficient process of drainage network extraction. Both the topology of the mapped hydrography and the initial landscape of the DEM are well preserved and fully utilized in the proposed method. An improved stream rasterization is achieved here, yielding continuous, unambiguous and stream-collision-free raster equivalent of stream vectors for flow enforcement. By imposing priority-based enforcement with a complementary flow direction enhancement procedure, the drainage patterns of the mapped hydrography are fully represented in the derived results. The proposed method was tested over the Rogue River Basin, using DEMs with various resolutions. As indicated by the visual and statistical analyses, the proposed method has three major advantages: (1) it significantly reduces the occurrences of topological errors, yielding very accurate watershed partition and channel delineation, (2) it ensures scale-consistent performance at DEMs of various resolutions, and (3) the entire extraction process is well-designed to achieve great computational efficiency.

  4. Peer review of health research funding proposals: A systematic map and systematic review of innovations for effectiveness and efficiency

    PubMed Central

    Frampton, Geoff K.; Pickett, Karen; Wyatt, Jeremy C.

    2018-01-01

    Objective To investigate methods and processes for timely, efficient and good quality peer review of research funding proposals in health. Methods A two-stage evidence synthesis: (1) a systematic map to describe the key characteristics of the evidence base, followed by (2) a systematic review of the studies stakeholders prioritised as relevant from the map on the effectiveness and efficiency of peer review ‘innovations’. Standard processes included literature searching, duplicate inclusion criteria screening, study keyword coding, data extraction, critical appraisal and study synthesis. Results A total of 83 studies from 15 countries were included in the systematic map. The evidence base is diverse, investigating many aspects of the systems for, and processes of, peer review. The systematic review included eight studies from Australia, Canada, and the USA, evaluating a broad range of peer review innovations. These studies showed that simplifying the process by shortening proposal forms, using smaller reviewer panels, or expediting processes can speed up the review process and reduce costs, but this might come at the expense of peer review quality, a key aspect that has not been assessed. Virtual peer review using videoconferencing or teleconferencing appears promising for reducing costs by avoiding the need for reviewers to travel, but again any consequences for quality have not been adequately assessed. Conclusions There is increasing international research activity into the peer review of health research funding. The studies reviewed had methodological limitations and variable generalisability to research funders. Given these limitations it is not currently possible to recommend immediate implementation of these innovations. However, many appear promising based on existing evidence, and could be adapted as necessary by funders and evaluated. Where feasible, experimental evaluation, including randomised controlled trials, should be conducted, evaluating impact

  5. Peer review of health research funding proposals: A systematic map and systematic review of innovations for effectiveness and efficiency.

    PubMed

    Shepherd, Jonathan; Frampton, Geoff K; Pickett, Karen; Wyatt, Jeremy C

    2018-01-01

    To investigate methods and processes for timely, efficient and good quality peer review of research funding proposals in health. A two-stage evidence synthesis: (1) a systematic map to describe the key characteristics of the evidence base, followed by (2) a systematic review of the studies stakeholders prioritised as relevant from the map on the effectiveness and efficiency of peer review 'innovations'. Standard processes included literature searching, duplicate inclusion criteria screening, study keyword coding, data extraction, critical appraisal and study synthesis. A total of 83 studies from 15 countries were included in the systematic map. The evidence base is diverse, investigating many aspects of the systems for, and processes of, peer review. The systematic review included eight studies from Australia, Canada, and the USA, evaluating a broad range of peer review innovations. These studies showed that simplifying the process by shortening proposal forms, using smaller reviewer panels, or expediting processes can speed up the review process and reduce costs, but this might come at the expense of peer review quality, a key aspect that has not been assessed. Virtual peer review using videoconferencing or teleconferencing appears promising for reducing costs by avoiding the need for reviewers to travel, but again any consequences for quality have not been adequately assessed. There is increasing international research activity into the peer review of health research funding. The studies reviewed had methodological limitations and variable generalisability to research funders. Given these limitations it is not currently possible to recommend immediate implementation of these innovations. However, many appear promising based on existing evidence, and could be adapted as necessary by funders and evaluated. Where feasible, experimental evaluation, including randomised controlled trials, should be conducted, evaluating impact on effectiveness, efficiency and quality.

  6. Plane Jane(s).

    ERIC Educational Resources Information Center

    Greenman, Geri

    2001-01-01

    Describes an assignment that was used in an advanced drawing class in which the students created self-portraits, breaking up their images using planes and angles to suggest their bone structure. Explains that the students also had to include three realistic portions in their drawings. (CMK)

  7. Odd complaints and doubtful conditions: norms of hypochondria in Jane Austen and Catherine Belling.

    PubMed

    Nelson, James Lindemann

    2014-06-01

    In her final fragmentary novel Sanditon, Jane Austen develops a theme that pervades her work from her juvenilia onward: illness, and in particular, illness imagined, invented, or self-inflicted. While the "invention of odd complaints" is characteristically a token of folly or weakness throughout her writing, in this last work imagined illness is also both a symbol and a cause of how selves and societies degenerate. In the shifting world of Sanditon, hypochondria is the lubricant for a society bent on turning health into a commodity. As a result, people's rationality and their moral character come under attack. Catherine Belling's recent subtle study, A Condition of Doubt: The Meanings of Hypochondria, unveils hypochondria's discursive and cultural character. Running sharply against the tenor of Austen's treatment, however, she argues in defense of the rationality of hypochondriacs; the notion that the condition may involve morally significant defects is not entertained; any connection to the commercialization of health care is muted. Here, I contrast Austen's morally and epistemically negative rendering of her hypochondriacal characters in Sanditon with Belling's efforts to create a sympathetic understanding of people with hypochondria. I will argue that, despite time gaps and genre differences, joint consideration of these texts can help bioethicists better appreciate how medicine can intensify, pathologize, and exploit anxieties about illness and death, thus adding to the challenges of living well in the face of mortality and morbidity.

  8. IMPROVEMENT OF EFFICIENCY OF CUT AND OVERLAY ASPHALT WORKS BY USING MOBILE MAPPING SYSTEM

    NASA Astrophysics Data System (ADS)

    Yabuki, Nobuyoshi; Nakaniwa, Kazuhide; Kidera, Hiroki; Nishi, Daisuke

    When the cut-and-overlay asphalt work is done for improving road pavement, conventional road surface elevation survey with levels often requires traffic regulation and takes much time and effort. Recently, although new surveying methods using non-prismatic total stations or fixed 3D laser scanners have been proposed in industry, they have not been adopted much due to their high cost. In this research, we propose a new method using Mobile Mapping Systems (MMS) in order to increase the efficiency and to reduce the cost. In this method, small white marks are painted at the intervals of 10m along the road to identify cross sections and to modify the elevations of the white marks with accurate survey data. To verify this proposed method, we executed an experiment and compared this method with the conventional level survey method and the fixed 3D laser scanning method at a road of Osaka University. The result showed that the proposed method had a similar accuracy with other methods and it was more efficient.

  9. A Round-Efficient Authenticated Key Agreement Scheme Based on Extended Chaotic Maps for Group Cloud Meeting

    PubMed Central

    Lee, Tian-Fu; Wang, Zeng-Bo

    2017-01-01

    The security is a critical issue for business purposes. For example, the cloud meeting must consider strong security to maintain the communication privacy. Considering the scenario with cloud meeting, we apply extended chaotic map to present passwordless group authentication key agreement, termed as Passwordless Group Authentication Key Agreement (PL-GAKA). PL-GAKA improves the computation efficiency for the simple group password-based authenticated key agreement (SGPAKE) proposed by Lee et al. in terms of computing the session key. Since the extended chaotic map has equivalent security level to the Diffie–Hellman key exchange scheme applied by SGPAKE, the security of PL-GAKA is not sacrificed when improving the computation efficiency. Moreover, PL-GAKA is a passwordless scheme, so the password maintenance is not necessary. Short-term authentication is considered, hence the communication security is stronger than other protocols by dynamically generating session key in each cloud meeting. In our analysis, we first prove that each meeting member can get the correct information during the meeting. We analyze common security issues for the proposed PL-GAKA in terms of session key security, mutual authentication, perfect forward security, and data integrity. Moreover, we also demonstrate that communicating in PL-GAKA is secure when suffering replay attacks, impersonation attacks, privileged insider attacks, and stolen-verifier attacks. Eventually, an overall comparison is given to show the performance between PL-GAKA, SGPAKE and related solutions. PMID:29207509

  10. Efficient robust reconstruction of dynamic PET activity maps with radioisotope decay constraints.

    PubMed

    Gao, Fei; Liu, Huafeng; Shi, Pengcheng

    2010-01-01

    Dynamic PET imaging performs sequence of data acquisition in order to provide visualization and quantification of physiological changes in specific tissues and organs. The reconstruction of activity maps is generally the first step in dynamic PET. State space Hinfinity approaches have been proved to be a robust method for PET image reconstruction where, however, temporal constraints are not considered during the reconstruction process. In addition, the state space strategies for PET image reconstruction have been computationally prohibitive for practical usage because of the need for matrix inversion. In this paper, we present a minimax formulation of the dynamic PET imaging problem where a radioisotope decay model is employed as physics-based temporal constraints on the photon counts. Furthermore, a robust steady state Hinfinity filter is developed to significantly improve the computational efficiency with minimal loss of accuracy. Experiments are conducted on Monte Carlo simulated image sequences for quantitative analysis and validation.

  11. An Efficient Strategy Combining SSR Markers- and Advanced QTL-seq-driven QTL Mapping Unravels Candidate Genes Regulating Grain Weight in Rice

    PubMed Central

    Daware, Anurag; Das, Sweta; Srivastava, Rishi; Badoni, Saurabh; Singh, Ashok K.; Agarwal, Pinky; Parida, Swarup K.; Tyagi, Akhilesh K.

    2016-01-01

    Development and use of genome-wide informative simple sequence repeat (SSR) markers and novel integrated genomic strategies are vital to drive genomics-assisted breeding applications and for efficient dissection of quantitative trait loci (QTLs) underlying complex traits in rice. The present study developed 6244 genome-wide informative SSR markers exhibiting in silico fragment length polymorphism based on repeat-unit variations among genomic sequences of 11 indica, japonica, aus, and wild rice accessions. These markers were mapped on diverse coding and non-coding sequence components of known cloned/candidate genes annotated from 12 chromosomes and revealed a much higher amplification (97%) and polymorphic potential (88%) along with wider genetic/functional diversity level (16–74% with a mean 53%) especially among accessions belonging to indica cultivar group, suggesting their utility in large-scale genomics-assisted breeding applications in rice. A high-density 3791 SSR markers-anchored genetic linkage map (IR 64 × Sonasal) spanning 2060 cM total map-length with an average inter-marker distance of 0.54 cM was generated. This reference genetic map identified six major genomic regions harboring robust QTLs (31% combined phenotypic variation explained with a 5.7–8.7 LOD) governing grain weight on six rice chromosomes. One strong grain weight major QTL region (OsqGW5.1) was narrowed-down by integrating traditional QTL mapping with high-resolution QTL region-specific integrated SSR and single nucleotide polymorphism markers-based QTL-seq analysis and differential expression profiling. This led us to delineate two natural allelic variants in two known cis-regulatory elements (RAV1AAT and CARGCW8GAT) of glycosyl hydrolase and serine carboxypeptidase genes exhibiting pronounced seed-specific differential regulation in low (Sonasal) and high (IR 64) grain weight mapping parental accessions. Our genome-wide SSR marker resource (polymorphic within/between diverse

  12. Generalized Smooth Transition Map Between Tent and Logistic Maps

    NASA Astrophysics Data System (ADS)

    Sayed, Wafaa S.; Fahmy, Hossam A. H.; Rezk, Ahmed A.; Radwan, Ahmed G.

    There is a continuous demand on novel chaotic generators to be employed in various modeling and pseudo-random number generation applications. This paper proposes a new chaotic map which is a general form for one-dimensional discrete-time maps employing the power function with the tent and logistic maps as special cases. The proposed map uses extra parameters to provide responses that fit multiple applications for which conventional maps were not enough. The proposed generalization covers also maps whose iterative relations are not based on polynomials, i.e. with fractional powers. We introduce a framework for analyzing the proposed map mathematically and predicting its behavior for various combinations of its parameters. In addition, we present and explain the transition map which results in intermediate responses as the parameters vary from their values corresponding to tent map to those corresponding to logistic map case. We study the properties of the proposed map including graph of the map equation, general bifurcation diagram and its key-points, output sequences, and maximum Lyapunov exponent. We present further explorations such as effects of scaling, system response with respect to the new parameters, and operating ranges other than transition region. Finally, a stream cipher system based on the generalized transition map validates its utility for image encryption applications. The system allows the construction of more efficient encryption keys which enhances its sensitivity and other cryptographic properties.

  13. Functionally Relevant Microsatellite Markers From Chickpea Transcription Factor Genes for Efficient Genotyping Applications and Trait Association Mapping

    PubMed Central

    Kujur, Alice; Bajaj, Deepak; Saxena, Maneesha S.; Tripathi, Shailesh; Upadhyaya, Hari D.; Gowda, C.L.L.; Singh, Sube; Jain, Mukesh; Tyagi, Akhilesh K.; Parida, Swarup K.

    2013-01-01

    We developed 1108 transcription factor gene-derived microsatellite (TFGMS) and 161 transcription factor functional domain-associated microsatellite (TFFDMS) markers from 707 TFs of chickpea. The robust amplification efficiency (96.5%) and high intra-specific polymorphic potential (34%) detected by markers suggest their immense utilities in efficient large-scale genotyping applications, including construction of both physical and functional transcript maps and understanding population structure. Candidate gene-based association analysis revealed strong genetic association of TFFDMS markers with three major seed and pod traits. Further, TFGMS markers in the 5′ untranslated regions of TF genes showing differential expression during seed development had higher trait association potential. The significance of TFFDMS markers was demonstrated by correlating their allelic variation with amino acid sequence expansion/contraction in the functional domain and alteration of secondary protein structure encoded by genes. The seed weight-associated markers were validated through traditional bi-parental genetic mapping. The determination of gene-specific linkage disequilibrium (LD) patterns in desi and kabuli based on single nucleotide polymorphism-microsatellite marker haplotypes revealed extended LD decay, enhanced LD resolution and trait association potential of genes. The evolutionary history of a strong seed-size/weight-associated TF based on natural variation and haplotype sharing among desi, kabuli and wild unravelled useful information having implication for seed-size trait evolution during chickpea domestication. PMID:23633531

  14. Efficient mapping of transgene integration sites and local structural changes in Cre transgenic mice using targeted locus amplification

    PubMed Central

    Cain-Hom, Carol; Splinter, Erik; van Min, Max; Simonis, Marieke; van de Heijning, Monique; Martinez, Maria; Asghari, Vida

    2017-01-01

    Abstract Cre/LoxP technology is widely used in the field of mouse genetics for spatial and/or temporal regulation of gene function. For Cre lines generated via pronuclear microinjection of a Cre transgene construct, the integration site is random and in most cases not known. Integration of a transgene can disrupt an endogenous gene, potentially interfering with interpretation of the phenotype. In addition, knowledge of where the transgene is integrated is important for planning of crosses between animals carrying a conditional allele and a given Cre allele in case the alleles are on the same chromosome. We have used targeted locus amplification (TLA) to efficiently map the transgene location in seven previously published Cre and CreERT2 transgenic lines. In all lines, transgene insertion was associated with structural changes of variable complexity, illustrating the importance of testing for rearrangements around the integration site. In all seven lines the exact integration site and breakpoint sequences were identified. Our methods, data and genotyping assays can be used as a resource for the mouse community and our results illustrate the power of the TLA method to not only efficiently map the integration site of any transgene, but also provide additional information regarding the transgene integration events. PMID:28053125

  15. Computer-assisted map projection research

    USGS Publications Warehouse

    Snyder, John Parr

    1985-01-01

    Computers have opened up areas of map projection research which were previously too complicated to utilize, for example, using a least-squares fit to a very large number of points. One application has been in the efficient transfer of data between maps on different projections. While the transfer of moderate amounts of data is satisfactorily accomplished using the analytical map projection formulas, polynomials are more efficient for massive transfers. Suitable coefficients for the polynomials may be determined more easily for general cases using least squares instead of Taylor series. A second area of research is in the determination of a map projection fitting an unlabeled map, so that accurate data transfer can take place. The computer can test one projection after another, and include iteration where required. A third area is in the use of least squares to fit a map projection with optimum parameters to the region being mapped, so that distortion is minimized. This can be accomplished for standard conformal, equalarea, or other types of projections. Even less distortion can result if complex transformations of conformal projections are utilized. This bulletin describes several recent applications of these principles, as well as historical usage and background.

  16. Evaluation of carrier collection probability in bifacial interdigitated-back-contact crystalline silicon solar cells by the internal quantum efficiency mapping method

    NASA Astrophysics Data System (ADS)

    Tachibana, Tomihisa; Tanahashi, Katsuto; Mochizuki, Toshimitsu; Shirasawa, Katsuhiko; Takato, Hidetaka

    2018-04-01

    Bifacial interdigitated-back-contact (IBC) silicon solar cells with a high bifaciality of 0.91 were fabricated. Screen printing and firing technology were used to reduce the production cost. For the first time, the relationship between the rear side structure and carrier collection probability was evaluated using internal quantum efficiency (IQE) mapping. The measurement results showed that the screen-printed electrode and back surface field (BSF) area led to low IQE. The low carrier collection probability by BSF area can be explained by electrical shading effects. Thus, it is clear that the IQE mapping system is useful to evaluate the IBC cell.

  17. Assessing map accuracy in a remotely sensed, ecoregion-scale cover map

    USGS Publications Warehouse

    Edwards, T.C.; Moisen, Gretchen G.; Cutler, D.R.

    1998-01-01

    Landscape- and ecoregion-based conservation efforts increasingly use a spatial component to organize data for analysis and interpretation. A challenge particular to remotely sensed cover maps generated from these efforts is how best to assess the accuracy of the cover maps, especially when they can exceed 1000 s/km2 in size. Here we develop and describe a methodological approach for assessing the accuracy of large-area cover maps, using as a test case the 21.9 million ha cover map developed for Utah Gap Analysis. As part of our design process, we first reviewed the effect of intracluster correlation and a simple cost function on the relative efficiency of cluster sample designs to simple random designs. Our design ultimately combined clustered and subsampled field data stratified by ecological modeling unit and accessibility (hereafter a mixed design). We next outline estimation formulas for simple map accuracy measures under our mixed design and report results for eight major cover types and the three ecoregions mapped as part of the Utah Gap Analysis. Overall accuracy of the map was 83.2% (SE=1.4). Within ecoregions, accuracy ranged from 78.9% to 85.0%. Accuracy by cover type varied, ranging from a low of 50.4% for barren to a high of 90.6% for man modified. In addition, we examined gains in efficiency of our mixed design compared with a simple random sample approach. In regard to precision, our mixed design was more precise than a simple random design, given fixed sample costs. We close with a discussion of the logistical constraints facing attempts to assess the accuracy of large-area, remotely sensed cover maps.

  18. Mapping crustal heterogeneity using Lg propagation efficiency throughout the Middle East, Mediterranean, Southern Europe and Northern Africa

    USGS Publications Warehouse

    McNamara, D.E.; Walter, W.R.

    2001-01-01

    In this paper we describe a technique for mapping the lateral variation of Lg characteristics such as Lg blockage, efficient Lg propagation, and regions of very high attenuation in the Middle East, North Africa, Europe and the Mediterranean regions. Lg is used in a variety of seismological applications from magnitude estimation to identification of nuclear explosions for monitoring compliance with the Comprehensive Nuclear-Test-Ban Treaty (CTBT). These applications can give significantly biased results if the Lg phase is reduced or blocked by discontinuous structure or thin crust. Mapping these structures using quantitative techniques for determining Lg amplitude attenuation can break down when the phase is below background noise. In such cases Lg blockage and inefficient propagation zones are often mapped out by hand. With our approach, we attempt to visually simplify this information by imaging crustal structure anomalies that significantly diminish the amplitude of Lg. The visualization of such anomalies is achieved by defining a grid of cells that covers the entire region of interest. We trace Lg rays for each event/ station pair, which is simply the great circle path, and attribute to each cell a value equal to the maximum value of the Lg/P-coda amplitude ratio for all paths traversing that particular cell. The resulting map, from this empirical approach, is easily interpreted in terms of crustal structure and can successfully image small blockage features often missed by analysis of raypaths alone. This map can then be used to screen out events with blocked Lg prior to performing Q tomography, and to avoid using Lg-based methods of event identification for the CTBT in regions where they cannot work. For this study we applied our technique to one of the most tectonically complex regions on the earth. Nearly 9000 earthquake/station raypaths, traversing the vast region comprised of the Middle East, Mediterranean, Southern Europe and Northern Africa, have been

  19. Genome contact map explorer: a platform for the comparison, interactive visualization and analysis of genome contact maps

    PubMed Central

    Kumar, Rajendra; Sobhy, Haitham

    2017-01-01

    Abstract Hi-C experiments generate data in form of large genome contact maps (Hi-C maps). These show that chromosomes are arranged in a hierarchy of three-dimensional compartments. But to understand how these compartments form and by how much they affect genetic processes such as gene regulation, biologists and bioinformaticians need efficient tools to visualize and analyze Hi-C data. However, this is technically challenging because these maps are big. In this paper, we remedied this problem, partly by implementing an efficient file format and developed the genome contact map explorer platform. Apart from tools to process Hi-C data, such as normalization methods and a programmable interface, we made a graphical interface that let users browse, scroll and zoom Hi-C maps to visually search for patterns in the Hi-C data. In the software, it is also possible to browse several maps simultaneously and plot related genomic data. The software is openly accessible to the scientific community. PMID:28973466

  20. Time-efficient high-resolution whole-brain three-dimensional macromolecular proton fraction mapping

    PubMed Central

    Yarnykh, Vasily L.

    2015-01-01

    Purpose Macromolecular proton fraction (MPF) mapping is a quantitative MRI method that reconstructs parametric maps of a relative amount of macromolecular protons causing the magnetization transfer (MT) effect and provides a biomarker of myelination in neural tissues. This study aimed to develop a high-resolution whole-brain MPF mapping technique utilizing a minimal possible number of source images for scan time reduction. Methods The described technique is based on replacement of an actually acquired reference image without MT saturation by a synthetic one reconstructed from R1 and proton density maps, thus requiring only three source images. This approach enabled whole-brain three-dimensional MPF mapping with isotropic 1.25×1.25×1.25 mm3 voxel size and scan time of 20 minutes. The synthetic reference method was validated against standard MPF mapping with acquired reference images based on data from 8 healthy subjects. Results Mean MPF values in segmented white and gray matter appeared in close agreement with no significant bias and small within-subject coefficients of variation (<2%). High-resolution MPF maps demonstrated sharp white-gray matter contrast and clear visualization of anatomical details including gray matter structures with high iron content. Conclusions Synthetic reference method improves resolution of MPF mapping and combines accurate MPF measurements with unique neuroanatomical contrast features. PMID:26102097

  1. BowMapCL: Burrows-Wheeler Mapping on Multiple Heterogeneous Accelerators.

    PubMed

    Nogueira, David; Tomas, Pedro; Roma, Nuno

    2016-01-01

    The computational demand of exact-search procedures has pressed the exploitation of parallel processing accelerators to reduce the execution time of many applications. However, this often imposes strict restrictions in terms of the problem size and implementation efforts, mainly due to their possibly distinct architectures. To circumvent this limitation, a new exact-search alignment tool (BowMapCL) based on the Burrows-Wheeler Transform and FM-Index is presented. Contrasting to other alternatives, BowMapCL is based on a unified implementation using OpenCL, allowing the exploitation of multiple and possibly different devices (e.g., NVIDIA, AMD/ATI, and Intel GPUs/APUs). Furthermore, to efficiently exploit such heterogeneous architectures, BowMapCL incorporates several techniques to promote its performance and scalability, including multiple buffering, work-queue task-distribution, and dynamic load-balancing, together with index partitioning, bit-encoding, and sampling. When compared with state-of-the-art tools, the attained results showed that BowMapCL (using a single GPU) is 2 × to 7.5 × faster than mainstream multi-threaded CPU BWT-based aligners, like Bowtie, BWA, and SOAP2; and up to 4 × faster than the best performing state-of-the-art GPU implementations (namely, SOAP3 and HPG-BWT). When multiple and completely distinct devices are considered, BowMapCL efficiently scales the offered throughput, ensuring a convenient load-balance of the involved processing in the several distinct devices.

  2. Efficient mapping of transgene integration sites and local structural changes in Cre transgenic mice using targeted locus amplification.

    PubMed

    Cain-Hom, Carol; Splinter, Erik; van Min, Max; Simonis, Marieke; van de Heijning, Monique; Martinez, Maria; Asghari, Vida; Cox, J Colin; Warming, Søren

    2017-05-05

    Cre/LoxP technology is widely used in the field of mouse genetics for spatial and/or temporal regulation of gene function. For Cre lines generated via pronuclear microinjection of a Cre transgene construct, the integration site is random and in most cases not known. Integration of a transgene can disrupt an endogenous gene, potentially interfering with interpretation of the phenotype. In addition, knowledge of where the transgene is integrated is important for planning of crosses between animals carrying a conditional allele and a given Cre allele in case the alleles are on the same chromosome. We have used targeted locus amplification (TLA) to efficiently map the transgene location in seven previously published Cre and CreERT2 transgenic lines. In all lines, transgene insertion was associated with structural changes of variable complexity, illustrating the importance of testing for rearrangements around the integration site. In all seven lines the exact integration site and breakpoint sequences were identified. Our methods, data and genotyping assays can be used as a resource for the mouse community and our results illustrate the power of the TLA method to not only efficiently map the integration site of any transgene, but also provide additional information regarding the transgene integration events. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  3. An efficient depth map preprocessing method based on structure-aided domain transform smoothing for 3D view generation

    PubMed Central

    Ma, Liyan; Qiu, Bo; Cui, Mingyue; Ding, Jianwei

    2017-01-01

    Depth image-based rendering (DIBR), which is used to render virtual views with a color image and the corresponding depth map, is one of the key techniques in the 2D to 3D conversion process. Due to the absence of knowledge about the 3D structure of a scene and its corresponding texture, DIBR in the 2D to 3D conversion process, inevitably leads to holes in the resulting 3D image as a result of newly-exposed areas. In this paper, we proposed a structure-aided depth map preprocessing framework in the transformed domain, which is inspired by recently proposed domain transform for its low complexity and high efficiency. Firstly, our framework integrates hybrid constraints including scene structure, edge consistency and visual saliency information in the transformed domain to improve the performance of depth map preprocess in an implicit way. Then, adaptive smooth localization is cooperated and realized in the proposed framework to further reduce over-smoothness and enhance optimization in the non-hole regions. Different from the other similar methods, the proposed method can simultaneously achieve the effects of hole filling, edge correction and local smoothing for typical depth maps in a united framework. Thanks to these advantages, it can yield visually satisfactory results with less computational complexity for high quality 2D to 3D conversion. Numerical experimental results demonstrate the excellent performances of the proposed method. PMID:28407027

  4. Image processing for optical mapping.

    PubMed

    Ravindran, Prabu; Gupta, Aditya

    2015-01-01

    Optical Mapping is an established single-molecule, whole-genome analysis system, which has been used to gain a comprehensive understanding of genomic structure and to study structural variation of complex genomes. A critical component of Optical Mapping system is the image processing module, which extracts single molecule restriction maps from image datasets of immobilized, restriction digested and fluorescently stained large DNA molecules. In this review, we describe robust and efficient image processing techniques to process these massive datasets and extract accurate restriction maps in the presence of noise, ambiguity and confounding artifacts. We also highlight a few applications of the Optical Mapping system.

  5. Updating flood maps efficiently using existing hydraulic models, very-high-accuracy elevation data, and a geographic information system; a pilot study on the Nisqually River, Washington

    USGS Publications Warehouse

    Jones, Joseph L.; Haluska, Tana L.; Kresch, David L.

    2001-01-01

    A method of updating flood inundation maps at a fraction of the expense of using traditional methods was piloted in Washington State as part of the U.S. Geological Survey Urban Geologic and Hydrologic Hazards Initiative. Large savings in expense may be achieved by building upon previous Flood Insurance Studies and automating the process of flood delineation with a Geographic Information System (GIS); increases in accuracy and detail result from the use of very-high-accuracy elevation data and automated delineation; and the resulting digital data sets contain valuable ancillary information such as flood depth, as well as greatly facilitating map storage and utility. The method consists of creating stage-discharge relations from the archived output of the existing hydraulic model, using these relations to create updated flood stages for recalculated flood discharges, and using a GIS to automate the map generation process. Many of the effective flood maps were created in the late 1970?s and early 1980?s, and suffer from a number of well recognized deficiencies such as out-of-date or inaccurate estimates of discharges for selected recurrence intervals, changes in basin characteristics, and relatively low quality elevation data used for flood delineation. FEMA estimates that 45 percent of effective maps are over 10 years old (FEMA, 1997). Consequently, Congress has mandated the updating and periodic review of existing maps, which have cost the Nation almost 3 billion (1997) dollars. The need to update maps and the cost of doing so were the primary motivations for piloting a more cost-effective and efficient updating method. New technologies such as Geographic Information Systems and LIDAR (Light Detection and Ranging) elevation mapping are key to improving the efficiency of flood map updating, but they also improve the accuracy, detail, and usefulness of the resulting digital flood maps. GISs produce digital maps without manual estimation of inundated areas between

  6. Efficient characterization of phase space mapping in axially symmetric optical systems

    NASA Astrophysics Data System (ADS)

    Barbero, Sergio; Portilla, Javier

    2018-01-01

    Phase space mapping, typically between an object and image plane, characterizes an optical system within a geometrical optics framework. We propose a novel conceptual frame to characterize the phase mapping in axially symmetric optical systems for arbitrary object locations, not restricted to a specific object plane. The idea is based on decomposing the phase mapping into a set of bivariate equations corresponding to different values of the radial coordinate on a specific object surface (most likely the entrance pupil). These equations are then approximated through bivariate Chebyshev interpolation at Chebyshev nodes, which guarantees uniform convergence. Additionally, we propose the use of a new concept (effective object phase space), defined as the set of points of the phase space at the first optical element (typically the entrance pupil) that are effectively mapped onto the image surface. The effective object phase space provides, by means of an inclusion test, a way to avoid tracing rays that do not reach the image surface.

  7. Efficient mapping of agricultural soils using a novel electromagnetic measurement system

    NASA Astrophysics Data System (ADS)

    Trinks, Immo; Pregesbauer, Michael

    2016-04-01

    "Despite all our accomplishments, we owe our existence to a six-inch layer of topsoil and the fact that it rains." - Paul Harvey. Despite the fact, that a farmers most precious good is the soil that he or she cultivates, in most cases actually very little is known about the soils that are being farmed. Agricultural soils are under constant threat through erosion, depletion, pollution and other degrading processes, in particular when considering intensive industrial scale farming. The capability of soils to retain water and soil moisture is of vital importance for their agricultural potential. Detailed knowledge of the physical properties of soils, their types and texture, water content and the depth of the agricultural layer would be of great importance for resource-efficient tillage with sub-area dependent variable depth, and the targeted intelligent application of fertilizers or irrigation. Precision farming, which has seen increasing popularity in the USA as well as Australia, is still in its infancy in Europe. Traditional near-surface geophysical prospection systems for agricultural soil mapping have either been based on earth resistance measurements using electrode-disks that require soil contact, with inherent issues, or electromagnetic induction (EMI) measurements conducted with EMI devices mounted in non-metallic sledges towed several metres behind survey vehicles across the fields. Every farmer passes over the fields several times during each growing season, working the soil and treating the crops. Therefore a novel user-friendly measurement system, the "Topsoil Mapper" (TSM) has been developed, which enables the farmer to simultaneously acquire soil conductivity information and derived soil parameters while anyway passing over the fields using different agricultural implements. The measurement principle of the TSM is electromagnetic induction using a multi-coil array to acquire conductivity information along a vertical profile down to approximately 1.1 m

  8. Automatic and efficient methods applied to the binarization of a subway map

    NASA Astrophysics Data System (ADS)

    Durand, Philippe; Ghorbanzadeh, Dariush; Jaupi, Luan

    2015-12-01

    The purpose of this paper is the study of efficient methods for image binarization. The objective of the work is the metro maps binarization. the goal is to binarize, avoiding noise to disturb the reading of subway stations. Different methods have been tested. By this way, a method given by Otsu gives particularly interesting results. The difficulty of the binarization is the choice of this threshold in order to reconstruct. Image sticky as possible to reality. Vectorization is a step subsequent to that of the binarization. It is to retrieve the coordinates points containing information and to store them in the two matrices X and Y. Subsequently, these matrices can be exported to a file format 'CSV' (Comma Separated Value) enabling us to deal with them in a variety of software including Excel. The algorithm uses quite a time calculation in Matlab because it is composed of two "for" loops nested. But the "for" loops are poorly supported by Matlab, especially in each other. This therefore penalizes the computation time, but seems the only method to do this.

  9. Using qualitative maps to direct reactive robots

    NASA Technical Reports Server (NTRS)

    Bertin, Randolph; Pendleton, Tom

    1992-01-01

    The principal advantage of mobile robots is that they are able to go to specific locations to perform useful tasks rather than have the tasks brought to them. It is important therefore that the robot be used to reach desired locations efficiently and reliably. A mobile robot whose environment extends significantly beyond its sensory horizon must maintain a representation of the environment, a map, in order to attain these efficiency and reliability requirements. We believe that qualitative mapping methods provide useful and robust representation schemes and that such maps may be used to direct the actions of a reactively controlled robot. In this paper we describe our experience in employing qualitative maps to direct, through the selection of desired control strategies, a reactive-behavior based robot. This mapping capability represents the development of one aspect of a successful deliberative/reactive hybrid control architecture.

  10. Accurate Mobile Urban Mapping via Digital Map-Based SLAM †

    PubMed Central

    Roh, Hyunchul; Jeong, Jinyong; Cho, Younggun; Kim, Ayoung

    2016-01-01

    This paper presents accurate urban map generation using digital map-based Simultaneous Localization and Mapping (SLAM). Throughout this work, our main objective is generating a 3D and lane map aiming for sub-meter accuracy. In conventional mapping approaches, achieving extremely high accuracy was performed by either (i) exploiting costly airborne sensors or (ii) surveying with a static mapping system in a stationary platform. Mobile scanning systems recently have gathered popularity but are mostly limited by the availability of the Global Positioning System (GPS). We focus on the fact that the availability of GPS and urban structures are both sporadic but complementary. By modeling both GPS and digital map data as measurements and integrating them with other sensor measurements, we leverage SLAM for an accurate mobile mapping system. Our proposed algorithm generates an efficient graph SLAM and achieves a framework running in real-time and targeting sub-meter accuracy with a mobile platform. Integrated with the SLAM framework, we implement a motion-adaptive model for the Inverse Perspective Mapping (IPM). Using motion estimation derived from SLAM, the experimental results show that the proposed approaches provide stable bird’s-eye view images, even with significant motion during the drive. Our real-time map generation framework is validated via a long-distance urban test and evaluated at randomly sampled points using Real-Time Kinematic (RTK)-GPS. PMID:27548175

  11. An unusual MAP kinase is required for efficient penetration of the plant surface by Ustilago maydis

    PubMed Central

    Brachmann, Andreas; Schirawski, Jan; Müller, Philip; Kahmann, Regine

    2003-01-01

    In Ustilago maydis, pathogenic development is controlled by a heterodimer of the two homeodomain proteins bW and bE. We have identified by RNA fingerprinting a b-regulated gene, kpp6, which encodes an unusual MAP kinase. Kpp6 is similar to a number of other fungal MAP kinases involved in mating and pathogenicity, but contains an additional N-terminal domain unrelated to other proteins. Transcription of the kpp6 gene yields two transcripts differing in length, but encoding proteins of identical mass. One transcript is upregulated by the bW/bE heterodimer, while the other is induced after pheromone stimulation. kpp6 deletion mutants are attenuated in pathogenicity. kpp6T355A,Y357F mutants carrying a non-activatable allele of kpp6 are more severely compromised in pathogenesis. These strains can still form appressoria, but are defective in the subsequent penetration of the plant cuticle. Kpp6 is expressed during all stages of the sexual life cycle except mature spores. We speculate that Kpp6 may respond to a plant signal and regulate the genes necessary for efficient penetration of plant tissue. PMID:12727886

  12. A note on the efficiencies of sampling strategies in two-stage Bayesian regional fine mapping of a quantitative trait.

    PubMed

    Chen, Zhijian; Craiu, Radu V; Bull, Shelley B

    2014-11-01

    In focused studies designed to follow up associations detected in a genome-wide association study (GWAS), investigators can proceed to fine-map a genomic region by targeted sequencing or dense genotyping of all variants in the region, aiming to identify a functional sequence variant. For the analysis of a quantitative trait, we consider a Bayesian approach to fine-mapping study design that incorporates stratification according to a promising GWAS tag SNP in the same region. Improved cost-efficiency can be achieved when the fine-mapping phase incorporates a two-stage design, with identification of a smaller set of more promising variants in a subsample taken in stage 1, followed by their evaluation in an independent stage 2 subsample. To avoid the potential negative impact of genetic model misspecification on inference we incorporate genetic model selection based on posterior probabilities for each competing model. Our simulation study shows that, compared to simple random sampling that ignores genetic information from GWAS, tag-SNP-based stratified sample allocation methods reduce the number of variants continuing to stage 2 and are more likely to promote the functional sequence variant into confirmation studies. © 2014 WILEY PERIODICALS, INC.

  13. Development of efficient and cost-effective distributed hydrological modeling tool MWEasyDHM based on open-source MapWindow GIS

    NASA Astrophysics Data System (ADS)

    Lei, Xiaohui; Wang, Yuhui; Liao, Weihong; Jiang, Yunzhong; Tian, Yu; Wang, Hao

    2011-09-01

    Many regions are still threatened with frequent floods and water resource shortage problems in China. Consequently, the task of reproducing and predicting the hydrological process in watersheds is hard and unavoidable for reducing the risks of damage and loss. Thus, it is necessary to develop an efficient and cost-effective hydrological tool in China as many areas should be modeled. Currently, developed hydrological tools such as Mike SHE and ArcSWAT (soil and water assessment tool based on ArcGIS) show significant power in improving the precision of hydrological modeling in China by considering spatial variability both in land cover and in soil type. However, adopting developed commercial tools in such a large developing country comes at a high cost. Commercial modeling tools usually contain large numbers of formulas, complicated data formats, and many preprocessing or postprocessing steps that may make it difficult for the user to carry out simulation, thus lowering the efficiency of the modeling process. Besides, commercial hydrological models usually cannot be modified or improved to be suitable for some special hydrological conditions in China. Some other hydrological models are open source, but integrated into commercial GIS systems. Therefore, by integrating hydrological simulation code EasyDHM, a hydrological simulation tool named MWEasyDHM was developed based on open-source MapWindow GIS, the purpose of which is to establish the first open-source GIS-based distributed hydrological model tool in China by integrating modules of preprocessing, model computation, parameter estimation, result display, and analysis. MWEasyDHM provides users with a friendly manipulating MapWindow GIS interface, selectable multifunctional hydrological processing modules, and, more importantly, an efficient and cost-effective hydrological simulation tool. The general construction of MWEasyDHM consists of four major parts: (1) a general GIS module for hydrological analysis, (2) a

  14. ZOOM Lite: next-generation sequencing data mapping and visualization software

    PubMed Central

    Zhang, Zefeng; Lin, Hao; Ma, Bin

    2010-01-01

    High-throughput next-generation sequencing technologies pose increasing demands on the efficiency, accuracy and usability of data analysis software. In this article, we present ZOOM Lite, a software for efficient reads mapping and result visualization. With a kernel capable of mapping tens of millions of Illumina or AB SOLiD sequencing reads efficiently and accurately, and an intuitive graphical user interface, ZOOM Lite integrates reads mapping and result visualization into a easy to use pipeline on desktop PC. The software handles both single-end and paired-end reads, and can output both the unique mapping result or the top N mapping results for each read. Additionally, the software takes a variety of input file formats and outputs to several commonly used result formats. The software is freely available at http://bioinfor.com/zoom/lite/. PMID:20530531

  15. Evaluating point count efficiency relative to territory mapping in cropland birds

    Treesearch

    Andre Cyr; Denis Lepage; Kathryn Freemark

    1995-01-01

    Species richness, composition, and abundance of farmland birds were compared between point counts (50-m, 100-m, and 150-m radius half circles) and territory mapping on three 40-ha plots in Québec, Canada. Point counts of smaller radii tended to have larger density estimates than counts of larger radii. Territory mapping detected 10 species more than 150-m...

  16. Soil Parameter Mapping and Ad Hoc Power Analysis to Increase Blocking Efficiency Prior to Establishing a Long-Term Field Experiment

    PubMed Central

    Collins, Doug; Benedict, Chris; Bary, Andy; Cogger, Craig

    2015-01-01

    The spatial heterogeneity of soil and weed populations poses a challenge to researchers. Unlike aboveground variability, below-ground variability is more difficult to discern without a strategic soil sampling pattern. While blocking is commonly used to control environmental variation, this strategy is rarely informed by data about current soil conditions. Fifty georeferenced sites were located in a 0.65 ha area prior to establishing a long-term field experiment. Soil organic matter (OM) and weed seed bank populations were analyzed at each site and the spatial structure was modeled with semivariograms and interpolated with kriging to map the surface. These maps were used to formulate three strategic blocking patterns and the efficiency of each pattern was compared to a completely randomized design and a west to east model not informed by soil variability. Compared to OM, weeds were more variable across the landscape and had a shorter range of autocorrelation, and models to increase blocking efficiency resulted in less increase in power. Weeds and OM were not correlated, so no model examined improved power equally for both parameters. Compared to the west to east blocking pattern, the final blocking pattern chosen resulted in a 7-fold increase in power for OM and a 36% increase in power for weeds. PMID:26247056

  17. Soil Parameter Mapping and Ad Hoc Power Analysis to Increase Blocking Efficiency Prior to Establishing a Long-Term Field Experiment.

    PubMed

    Collins, Doug; Benedict, Chris; Bary, Andy; Cogger, Craig

    2015-01-01

    The spatial heterogeneity of soil and weed populations poses a challenge to researchers. Unlike aboveground variability, below-ground variability is more difficult to discern without a strategic soil sampling pattern. While blocking is commonly used to control environmental variation, this strategy is rarely informed by data about current soil conditions. Fifty georeferenced sites were located in a 0.65 ha area prior to establishing a long-term field experiment. Soil organic matter (OM) and weed seed bank populations were analyzed at each site and the spatial structure was modeled with semivariograms and interpolated with kriging to map the surface. These maps were used to formulate three strategic blocking patterns and the efficiency of each pattern was compared to a completely randomized design and a west to east model not informed by soil variability. Compared to OM, weeds were more variable across the landscape and had a shorter range of autocorrelation, and models to increase blocking efficiency resulted in less increase in power. Weeds and OM were not correlated, so no model examined improved power equally for both parameters. Compared to the west to east blocking pattern, the final blocking pattern chosen resulted in a 7-fold increase in power for OM and a 36% increase in power for weeds.

  18. Improved Mapping of Carbon, Water and Energy Land-Surface Fluxes Using Remotely Sensed Indicators of Canopy Light Use Efficiency

    NASA Astrophysics Data System (ADS)

    Schull, M. A.; Anderson, M. C.; Kustas, W.; Cammalleri, C.; Houborg, R.

    2012-12-01

    A light-use-efficiency (LUE) based model of canopy resistance has been embedded into a thermal-based Two-Source Energy Balance (TSEB) model to facilitate coupled simulations of transpiration and carbon assimilation. The model assumes that deviations of the observed canopy LUE from a nominal stand-level value (LUEn - typically indexed by vegetation class) are due to varying conditions of light, humidity, CO2 concentration and leaf temperature. The deviations are accommodated by adjusting an effective LUE that responds to the varying conditions. The challenge to monitoring fluxes on a larger scale is to capture the physiological responses due to changing conditions. This challenge can be met using remotely sensed leaf chlorophyll (Cab). Since Cab is a vital pigment for absorbing light for use in photosynthesis, it has been recognized as a key parameter for quantifying photosynthetic functioning that are sensitive to these conditions. Recent studies have shown that it is sensitive to changes in LUE, which defines how efficiently a plant can assimilate carbon dioxide (CO2) given the absorbed Photosynthetically Active Radiation (PAR) and is therefore useful for monitoring carbon fluxes. We investigate the feasibility of leaf chlorophyll to capture these variations in LUEn using remotely sensed data. To retrieve Cab from remotely sensed data we use REGFLEC, a physically based tool that translates at-sensor radiances in the green, red and NIR spectral regions from multiple satellite sensors into realistic maps of LAI and Cab. Initial results show that Cab is exponentially correlated to light use efficiency. Incorporating nominal light use efficiency estimated from Cab is shown to improve fluxes of carbon, water and energy most notably in times of stressed vegetation. The result illustrates that Cab is sensitive to changes in plant physiology and can capture plant stress needed for improved estimation of fluxes. The observed relationship and initial results demonstrate the

  19. Improving estimates of genetic maps: a meta-analysis-based approach.

    PubMed

    Stewart, William C L

    2007-07-01

    Inaccurate genetic (or linkage) maps can reduce the power to detect linkage, increase type I error, and distort haplotype and relationship inference. To improve the accuracy of existing maps, I propose a meta-analysis-based method that combines independent map estimates into a single estimate of the linkage map. The method uses the variance of each independent map estimate to combine them efficiently, whether the map estimates use the same set of markers or not. As compared with a joint analysis of the pooled genotype data, the proposed method is attractive for three reasons: (1) it has comparable efficiency to the maximum likelihood map estimate when the pooled data are homogeneous; (2) relative to existing map estimation methods, it can have increased efficiency when the pooled data are heterogeneous; and (3) it avoids the practical difficulties of pooling human subjects data. On the basis of simulated data modeled after two real data sets, the proposed method can reduce the sampling variation of linkage maps commonly used in whole-genome linkage scans. Furthermore, when the independent map estimates are also maximum likelihood estimates, the proposed method performs as well as or better than when they are estimated by the program CRIMAP. Since variance estimates of maps may not always be available, I demonstrate the feasibility of three different variance estimators. Overall, the method should prove useful to investigators who need map positions for markers not contained in publicly available maps, and to those who wish to minimize the negative effects of inaccurate maps. Copyright 2007 Wiley-Liss, Inc.

  20. An Improvement of Robust and Efficient Biometrics Based Password Authentication Scheme for Telecare Medicine Information Systems Using Extended Chaotic Maps.

    PubMed

    Moon, Jongho; Choi, Younsung; Kim, Jiye; Won, Dongho

    2016-03-01

    Recently, numerous extended chaotic map-based password authentication schemes that employ smart card technology were proposed for Telecare Medical Information Systems (TMISs). In 2015, Lu et al. used Li et al.'s scheme as a basis to propose a password authentication scheme for TMISs that is based on biometrics and smart card technology and employs extended chaotic maps. Lu et al. demonstrated that Li et al.'s scheme comprises some weaknesses such as those regarding a violation of the session-key security, a vulnerability to the user impersonation attack, and a lack of local verification. In this paper, however, we show that Lu et al.'s scheme is still insecure with respect to issues such as a violation of the session-key security, and that it is vulnerable to both the outsider attack and the impersonation attack. To overcome these drawbacks, we retain the useful properties of Lu et al.'s scheme to propose a new password authentication scheme that is based on smart card technology and requires the use of chaotic maps. Then, we show that our proposed scheme is more secure and efficient and supports security properties.

  1. Mapping the Campus.

    ERIC Educational Resources Information Center

    Stigner, Kenneth J.

    2001-01-01

    Discusses how aerial photography and photogrammetry technology can help schools create visual records of their campus, land, and properties. Addresses efficiency and cost effectiveness of this method. Discusses how to develop the digital photogrammetry method for mapping from aerial photos. (GR)

  2. National Seabed Mapping Programmes Collaborate to Advance Marine Geomorphological Mapping in Adjoining European Seas

    NASA Astrophysics Data System (ADS)

    Monteys, X.; Guinan, J.; Green, S.; Gafeira, J.; Dove, D.; Baeten, N. J.; Thorsnes, T.

    2017-12-01

    Marine geomorphological mapping is an effective means of characterising and understanding the seabed and its features with direct relevance to; offshore infrastructure placement, benthic habitat mapping, conservation & policy, marine spatial planning, fisheries management and pure research. Advancements in acoustic survey techniques and data processing methods resulting in the availability of high-resolution marine datasets e.g. multibeam echosounder bathymetry and shallow seismic mean that geological interpretations can be greatly improved by combining with geomorphological maps. Since December 2015, representatives from the national seabed mapping programmes of Norway (MAREANO), Ireland (INFOMAR) and the United Kingdom (MAREMAP) have collaborated and established the MIM geomorphology working group) with the common aim of advancing best practice for geological mapping in their adjoining sea areas in north-west Europe. A recently developed two-part classification system for Seabed Geomorphology (`Morphology' and Geomorphology') has been established as a result of an initiative led by the British Geological Survey (BGS) with contributions from the MIM group (Dove et al. 2016). To support the scheme, existing BGS GIS tools (SIGMA) have been adapted to apply this two-part classification system and here we present on the tools effectiveness in mapping geomorphological features, along with progress in harmonising the classification and feature nomenclature. Recognising that manual mapping of seabed features can be time-consuming and subjective, semi-automated approaches for mapping seabed features and improving mapping efficiency is being developed using Arc-GIS based tools. These methods recognise, spatially delineate and morphologically describe seabed features such as pockmarks (Gafeira et al., 2012) and cold-water coral mounds. Such tools utilise multibeam echosounder data or any other bathymetric dataset (e.g. 3D seismic, Geldof et al., 2014) that can produce a

  3. Derivation of the dipole map

    NASA Astrophysics Data System (ADS)

    Ali, Halima; Punjabi, Alkesh; Boozer, Allen

    2004-09-01

    In our method of maps [Punjabi et al., Phy. Rev. Lett. 69, 3322 (1992), and Punjabi et al., J. Plasma Phys. 52, 91 (1994)], symplectic maps are used to calculate the trajectories of magnetic field lines in divertor tokamaks. Effects of the magnetic perturbations are calculated using the low MN map [Ali et al., Phys. Plasmas 11, 1908 (2004)] and the dipole map [Punjabi et al., Phys. Plasmas 10, 3992 (2003)]. The dipole map is used to calculate the effects of externally located current carrying coils on the trajectories of the field lines, the stochastic layer, the magnetic footprint, and the heat load distribution on the collector plates in divertor tokamaks [Punjabi et al., Phys. Plasmas 10, 3992 (2003)]. Symplectic maps are general, efficient, and preserve and respect the Hamiltonian nature of the dynamics. In this brief communication, a rigorous mathematical derivation of the dipole map is given.

  4. Research on the Application of Rapid Surveying and Mapping for Large Scare Topographic Map by Uav Aerial Photography System

    NASA Astrophysics Data System (ADS)

    Gao, Z.; Song, Y.; Li, C.; Zeng, F.; Wang, F.

    2017-08-01

    Rapid acquisition and processing method of large scale topographic map data, which relies on the Unmanned Aerial Vehicle (UAV) low-altitude aerial photogrammetry system, is studied in this paper, elaborating the main work flow. Key technologies of UAV photograph mapping is also studied, developing a rapid mapping system based on electronic plate mapping system, thus changing the traditional mapping mode and greatly improving the efficiency of the mapping. Production test and achievement precision evaluation of Digital Orth photo Map (DOM), Digital Line Graphic (DLG) and other digital production were carried out combined with the city basic topographic map update project, which provides a new techniques for large scale rapid surveying and has obvious technical advantage and good application prospect.

  5. Revision of Primary Series Maps

    USGS Publications Warehouse

    ,

    2000-01-01

    In 1992, the U.S. Geological Survey (USGS) completed a 50-year effort to provide primary series map coverage of the United States. Many of these maps now need to be updated to reflect the construction of new roads and highways and other changes that have taken place over time. The USGS has formulated a graphic revision plan to help keep the primary series maps current. Primary series maps include 1:20,000-scale quadrangles of Puerto Rico, 1:24,000- or 1:25,000-scale quadrangles of the conterminous United States, Hawaii, and U.S. Territories, and 1:63,360-scale quadrangles of Alaska. The revision of primary series maps from new collection sources is accomplished using a variety of processes. The raster revision process combines the scanned content of paper maps with raster updating technologies. The vector revision process involves the automated plotting of updated vector files. Traditional processes use analog stereoplotters and manual scribing instruments on specially coated map separates. The ability to select from or combine these processes increases the efficiency of the National Mapping Division map revision program.

  6. Drawing Road Networks with Mental Maps.

    PubMed

    Lin, Shih-Syun; Lin, Chao-Hung; Hu, Yan-Jhang; Lee, Tong-Yee

    2014-09-01

    Tourist and destination maps are thematic maps designed to represent specific themes in maps. The road network topologies in these maps are generally more important than the geometric accuracy of roads. A road network warping method is proposed to facilitate map generation and improve theme representation in maps. The basic idea is deforming a road network to meet a user-specified mental map while an optimization process is performed to propagate distortions originating from road network warping. To generate a map, the proposed method includes algorithms for estimating road significance and for deforming a road network according to various geometric and aesthetic constraints. The proposed method can produce an iconic mark of a theme from a road network and meet a user-specified mental map. Therefore, the resulting map can serve as a tourist or destination map that not only provides visual aids for route planning and navigation tasks, but also visually emphasizes the presentation of a theme in a map for the purpose of advertising. In the experiments, the demonstrations of map generations show that our method enables map generation systems to generate deformed tourist and destination maps efficiently.

  7. From Open Geographical Data to Tangible Maps: Improving the Accessibility of Maps for Visually Impaired People

    NASA Astrophysics Data System (ADS)

    Ducasse, J.; Macé, M.; Jouffrais, C.

    2015-08-01

    Visual maps must be transcribed into (interactive) raised-line maps to be accessible for visually impaired people. However, these tactile maps suffer from several shortcomings: they are long and expensive to produce, they cannot display a large amount of information, and they are not dynamically modifiable. A number of methods have been developed to automate the production of raised-line maps, but there is not yet any tactile map editor on the market. Tangible interactions proved to be an efficient way to help a visually impaired user manipulate spatial representations. Contrary to raised-line maps, tangible maps can be autonomously constructed and edited. In this paper, we present the scenarios and the main expected contributions of the AccessiMap project, which is based on the availability of many sources of open spatial data: 1/ facilitating the production of interactive tactile maps with the development of an open-source web-based editor; 2/ investigating the use of tangible interfaces for the autonomous construction and exploration of a map by a visually impaired user.

  8. An efficient and secure partial image encryption for wireless multimedia sensor networks using discrete wavelet transform, chaotic maps and substitution box

    NASA Astrophysics Data System (ADS)

    Khan, Muazzam A.; Ahmad, Jawad; Javaid, Qaisar; Saqib, Nazar A.

    2017-03-01

    Wireless Sensor Networks (WSN) is widely deployed in monitoring of some physical activity and/or environmental conditions. Data gathered from WSN is transmitted via network to a central location for further processing. Numerous applications of WSN can be found in smart homes, intelligent buildings, health care, energy efficient smart grids and industrial control systems. In recent years, computer scientists has focused towards findings more applications of WSN in multimedia technologies, i.e. audio, video and digital images. Due to bulky nature of multimedia data, WSN process a large volume of multimedia data which significantly increases computational complexity and hence reduces battery time. With respect to battery life constraints, image compression in addition with secure transmission over a wide ranged sensor network is an emerging and challenging task in Wireless Multimedia Sensor Networks. Due to the open nature of the Internet, transmission of data must be secure through a process known as encryption. As a result, there is an intensive demand for such schemes that is energy efficient as well as highly secure since decades. In this paper, discrete wavelet-based partial image encryption scheme using hashing algorithm, chaotic maps and Hussain's S-Box is reported. The plaintext image is compressed via discrete wavelet transform and then the image is shuffled column-wise and row wise-wise via Piece-wise Linear Chaotic Map (PWLCM) and Nonlinear Chaotic Algorithm, respectively. To get higher security, initial conditions for PWLCM are made dependent on hash function. The permuted image is bitwise XORed with random matrix generated from Intertwining Logistic map. To enhance the security further, final ciphertext is obtained after substituting all elements with Hussain's substitution box. Experimental and statistical results confirm the strength of the anticipated scheme.

  9. Efficient receptive field tiling in primate V1

    PubMed Central

    Nauhaus, Ian; Nielsen, Kristina J.; Callaway, Edward M.

    2017-01-01

    The primary visual cortex (V1) encodes a diverse set of visual features, including orientation, ocular dominance (OD) and spatial frequency (SF), whose joint organization must be precisely structured to optimize coverage within the retinotopic map. Prior experiments have only identified efficient coverage based on orthogonal maps. Here, we used two-photon calcium imaging to reveal an alternative arrangement for OD and SF maps in macaque V1; their gradients run parallel but with unique spatial periods, whereby low SF regions coincide with monocular regions. Next, we mapped receptive fields and find surprisingly precise micro-retinotopy that yields a smaller point-image and requires more efficient inter-map geometry, thus underscoring the significance of map relationships. While smooth retinotopy is constraining, studies suggest that it improves both wiring economy and the V1 population code read downstream. Altogether, these data indicate that connectivity within V1 is finely tuned and precise at the level of individual neurons. PMID:27499086

  10. Execution models for mapping programs onto distributed memory parallel computers

    NASA Technical Reports Server (NTRS)

    Sussman, Alan

    1992-01-01

    The problem of exploiting the parallelism available in a program to efficiently employ the resources of the target machine is addressed. The problem is discussed in the context of building a mapping compiler for a distributed memory parallel machine. The paper describes using execution models to drive the process of mapping a program in the most efficient way onto a particular machine. Through analysis of the execution models for several mapping techniques for one class of programs, we show that the selection of the best technique for a particular program instance can make a significant difference in performance. On the other hand, the results of benchmarks from an implementation of a mapping compiler show that our execution models are accurate enough to select the best mapping technique for a given program.

  11. Inference of personality projected onto fictional characters having an author's first name.

    PubMed

    Barry, H

    2001-12-01

    Jane Austen projected some of her personality characteristics onto her fictional namesakes Jane Bennet in the novel Pride and Prejudice and Jane Fairfax in the novel Emma. Wishful fantasy seems satisfied by two attributes of both Janes. They are very beautiful, and they marry rich men they love. A feeling of inferiority was expressed by two attributes of both Janes, depicted as deficient in social communication and subordinate to the heroine of the novel.

  12. Mapping nutrient resorption efficiencies of subarctic cryptogams and seed plants onto the Tree of Life

    PubMed Central

    Lang, Simone I; Aerts, Rien; van Logtestijn, Richard S P; Schweikert, Wenka; Klahn, Thorsten; Quested, Helen M; van Hal, Jurgen R; Cornelissen, Johannes H C

    2014-01-01

    Nutrient resorption from senescing photosynthetic organs is a powerful mechanism for conserving nitrogen (N) and phosphorus (P) in infertile environments. Evolution has resulted in enhanced differentiation of conducting tissues to facilitate transport of photosynthate to other plant parts, ultimately leading to phloem. Such tissues may also serve to translocate N and P to other plant parts upon their senescence. Therefore, we hypothesize that nutrient resorption efficiency (RE, % of nutrient pool exported) should correspond with the degree of specialization of these conducting tissues across the autotrophic branches of the Tree of Life. To test this hypothesis, we had to compare members of different plant clades and lichens within a climatic region, to minimize confounding effects of climatic drivers on nutrient resorption. Thus, we compared RE among wide-ranging basal clades from the principally N-limited subarctic region, employing a novel method to correct for mass loss during senescence. Even with the limited numbers of species available for certain clades in this region, we found some consistent patterns. Mosses, lichens, and lycophytes generally showed low REN (<20%), liverworts and conifers intermediate (40%) and monilophytes, eudicots, and monocots high (>70%). REP appeared higher in eudicots and liverworts than in mosses. Within mosses, taxa with more efficient conductance also showed higher REN. The differences in REN among clades broadly matched the degree of specialization of conducting tissues. This novel mapping of a physiological process onto the Tree of Life broadly supports the idea that the evolution of conducting tissues toward specialized phloem has aided land plants to optimize their internal nitrogen recycling. The generality of evolutionary lines in conducting tissues and nutrient resorption efficiency needs to be tested across different floras in different climatic regions with different levels of N versus P availability. PMID:25360262

  13. A Game Map Complexity Measure Based on Hamming Distance

    NASA Astrophysics Data System (ADS)

    Li, Yan; Su, Pan; Li, Wenliang

    With the booming of PC game market, Game AI has attracted more and more researches. The interesting and difficulty of a game are relative with the map used in game scenarios. Besides, the path-finding efficiency in a game is also impacted by the complexity of the used map. In this paper, a novel complexity measure based on Hamming distance, called the Hamming complexity, is introduced. This measure is able to estimate the complexity of binary tileworld. We experimentally demonstrated that Hamming complexity is highly relative with the efficiency of A* algorithm, and therefore it is a useful reference to the designer when developing a game map.

  14. Automatic metro map layout using multicriteria optimization.

    PubMed

    Stott, Jonathan; Rodgers, Peter; Martínez-Ovando, Juan Carlos; Walker, Stephen G

    2011-01-01

    This paper describes an automatic mechanism for drawing metro maps. We apply multicriteria optimization to find effective placement of stations with a good line layout and to label the map unambiguously. A number of metrics are defined, which are used in a weighted sum to find a fitness value for a layout of the map. A hill climbing optimizer is used to reduce the fitness value, and find improved map layouts. To avoid local minima, we apply clustering techniques to the map-the hill climber moves both stations and clusters when finding improved layouts. We show the method applied to a number of metro maps, and describe an empirical study that provides some quantitative evidence that automatically-drawn metro maps can help users to find routes more efficiently than either published maps or undistorted maps. Moreover, we have found that, in these cases, study subjects indicate a preference for automatically-drawn maps over the alternatives. © 2011 IEEE Published by the IEEE Computer Society

  15. Color is processed less efficiently than orientation in change detection but more efficiently in visual search.

    PubMed

    Huang, Liqiang

    2015-05-01

    Basic visual features (e.g., color, orientation) are assumed to be processed in the same general way across different visual tasks. Here, a significant deviation from this assumption was predicted on the basis of the analysis of stimulus spatial structure, as characterized by the Boolean-map notion. If a task requires memorizing the orientations of a set of bars, then the map consisting of those bars can be readily used to hold the overall structure in memory and will thus be especially useful. If the task requires visual search for a target, then the map, which contains only an overall structure, will be of little use. Supporting these predictions, the present study demonstrated that in comparison to stimulus colors, bar orientations were processed more efficiently in change-detection tasks but less efficiently in visual search tasks (Cohen's d = 4.24). In addition to offering support for the role of the Boolean map in conscious access, the present work also throws doubts on the generality of processing visual features. © The Author(s) 2015.

  16. WE-AB-209-10: Optimizing the Delivery of Sequential Fluence Maps for Efficient VMAT Delivery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Craft, D; Balvert, M

    2016-06-15

    Purpose: To develop an optimization model and solution approach for computing MLC leaf trajectories and dose rates for high quality matching of a set of optimized fluence maps to be delivered sequentially around a patient in a VMAT treatment. Methods: We formulate the fluence map matching problem as a nonlinear optimization problem where time is discretized but dose rates and leaf positions are continuous variables. For a given allotted time, which is allocated across the fluence maps based on the complexity of each fluence map, the optimization problem searches for the best leaf trajectories and dose rates such that themore » original fluence maps are closely recreated. Constraints include maximum leaf speed, maximum dose rate, and leaf collision avoidance, as well as the constraint that the ending leaf positions for one map are the starting leaf positions for the next map. The resulting model is non-convex but smooth, and therefore we solve it by local searches from a variety of starting positions. We improve solution time by a custom decomposition approach which allows us to decouple the rows of the fluence maps and solve each leaf pair individually. This decomposition also makes the problem easily parallelized. Results: We demonstrate method on a prostate case and a head-and-neck case and show that one can recreate fluence maps to high degree of fidelity in modest total delivery time (minutes). Conclusion: We present a VMAT sequencing method that reproduces optimal fluence maps by searching over a vast number of possible leaf trajectories. By varying the total allotted time given, this approach is the first of its kind to allow users to produce VMAT solutions that span the range of wide-field coarse VMAT deliveries to narrow-field high-MU sliding window-like approaches.« less

  17. a Mapping Method of Slam Based on Look up Table

    NASA Astrophysics Data System (ADS)

    Wang, Z.; Li, J.; Wang, A.; Wang, J.

    2017-09-01

    In the last years several V-SLAM(Visual Simultaneous Localization and Mapping) approaches have appeared showing impressive reconstructions of the world. However these maps are built with far more than the required information. This limitation comes from the whole process of each key-frame. In this paper we present for the first time a mapping method based on the LOOK UP TABLE(LUT) for visual SLAM that can improve the mapping effectively. As this method relies on extracting features in each cell divided from image, it can get the pose of camera that is more representative of the whole key-frame. The tracking direction of key-frames is obtained by counting the number of parallax directions of feature points. LUT stored all mapping needs the number of cell corresponding to the tracking direction which can reduce the redundant information in the key-frame, and is more efficient to mapping. The result shows that a better map with less noise is build using less than one-third of the time. We believe that the capacity of LUT efficiently building maps makes it a good choice for the community to investigate in the scene reconstruction problems.

  18. Virtual optical network mapping and core allocation in elastic optical networks using multi-core fibers

    NASA Astrophysics Data System (ADS)

    Xuan, Hejun; Wang, Yuping; Xu, Zhanqi; Hao, Shanshan; Wang, Xiaoli

    2017-11-01

    Virtualization technology can greatly improve the efficiency of the networks by allowing the virtual optical networks to share the resources of the physical networks. However, it will face some challenges, such as finding the efficient strategies for virtual nodes mapping, virtual links mapping and spectrum assignment. It is even more complex and challenging when the physical elastic optical networks using multi-core fibers. To tackle these challenges, we establish a constrained optimization model to determine the optimal schemes of optical network mapping, core allocation and spectrum assignment. To solve the model efficiently, tailor-made encoding scheme, crossover and mutation operators are designed. Based on these, an efficient genetic algorithm is proposed to obtain the optimal schemes of the virtual nodes mapping, virtual links mapping, core allocation. The simulation experiments are conducted on three widely used networks, and the experimental results show the effectiveness of the proposed model and algorithm.

  19. Selection of transformation-efficient barley genotypes based on TFA (transformation amenability) haplotype and higher resolution mapping of the TFA loci.

    PubMed

    Hisano, Hiroshi; Meints, Brigid; Moscou, Matthew J; Cistue, Luis; Echávarri, Begoña; Sato, Kazuhiro; Hayes, Patrick M

    2017-04-01

    The genetic substitution of transformation amenability alleles from 'Golden Promise' can facilitate the development of transformation-efficient lines from recalcitrant barley cultivars. Barley (Hordeum vulgare) cv. 'Golden Promise' is one of the most useful and well-studied cultivars for genetic manipulation. In a previous report, we identified several transformation amenability (TFA) loci responsible for Agrobacterium-mediated transformation using the F 2 generation of immature embryos, derived from 'Haruna Nijo' × 'Golden Promise,' as explants. In this report, we describe higher density mapping of these TFA regions with additional SNP markers using the same transgenic plants. To demonstrate the robustness of transformability alleles at the TFA loci, we genotyped 202 doubled haploid progeny from the cross 'Golden Promise' × 'Full Pint.' Based on SNP genotype, we selected lines having 'Golden Promise' alleles at TFA loci and used them for transformation. Of the successfully transformed lines, DH120366 came the closest to achieving a level of transformation efficiency comparable to 'Golden Promise.' The results validate that the genetic substitution of TFA alleles from 'Golden Promise' can facilitate the development of transformation-efficient lines from recalcitrant barley cultivars.

  20. Laser-Based Slam with Efficient Occupancy Likelihood Map Learning for Dynamic Indoor Scenes

    NASA Astrophysics Data System (ADS)

    Li, Li; Yao, Jian; Xie, Renping; Tu, Jinge; Feng, Chen

    2016-06-01

    Location-Based Services (LBS) have attracted growing attention in recent years, especially in indoor environments. The fundamental technique of LBS is the map building for unknown environments, this technique also named as simultaneous localization and mapping (SLAM) in robotic society. In this paper, we propose a novel approach for SLAMin dynamic indoor scenes based on a 2D laser scanner mounted on a mobile Unmanned Ground Vehicle (UGV) with the help of the grid-based occupancy likelihood map. Instead of applying scan matching in two adjacent scans, we propose to match current scan with the occupancy likelihood map learned from all previous scans in multiple scales to avoid the accumulation of matching errors. Due to that the acquisition of the points in a scan is sequential but not simultaneous, there unavoidably exists the scan distortion at different extents. To compensate the scan distortion caused by the motion of the UGV, we propose to integrate a velocity of a laser range finder (LRF) into the scan matching optimization framework. Besides, to reduce the effect of dynamic objects such as walking pedestrians often existed in indoor scenes as much as possible, we propose a new occupancy likelihood map learning strategy by increasing or decreasing the probability of each occupancy grid after each scan matching. Experimental results in several challenged indoor scenes demonstrate that our proposed approach is capable of providing high-precision SLAM results.

  1. BatMis: a fast algorithm for k-mismatch mapping.

    PubMed

    Tennakoon, Chandana; Purbojati, Rikky W; Sung, Wing-Kin

    2012-08-15

    Second-generation sequencing (SGS) generates millions of reads that need to be aligned to a reference genome allowing errors. Although current aligners can efficiently map reads allowing a small number of mismatches, they are not well suited for handling a large number of mismatches. The efficiency of aligners can be improved using various heuristics, but the sensitivity and accuracy of the alignments are sacrificed. In this article, we introduce Basic Alignment tool for Mismatches (BatMis)--an efficient method to align short reads to a reference allowing k mismatches. BatMis is a Burrows-Wheeler transformation based aligner that uses a seed and extend approach, and it is an exact method. Benchmark tests show that BatMis performs better than competing aligners in solving the k-mismatch problem. Furthermore, it can compete favorably even when compared with the heuristic modes of the other aligners. BatMis is a useful alternative for applications where fast k-mismatch mappings, unique mappings or multiple mappings of SGS data are required. BatMis is written in C/C++ and is freely available from http://code.google.com/p/batmis/

  2. Efficient dense blur map estimation for automatic 2D-to-3D conversion

    NASA Astrophysics Data System (ADS)

    Vosters, L. P. J.; de Haan, G.

    2012-03-01

    Focus is an important depth cue for 2D-to-3D conversion of low depth-of-field images and video. However, focus can be only reliably estimated on edges. Therefore, Bea et al. [1] first proposed an optimization based approach to propagate focus to non-edge image portions, for single image focus editing. While their approach produces accurate dense blur maps, the computational complexity and memory requirements for solving the resulting sparse linear system with standard multigrid or (multilevel) preconditioning techniques, are infeasible within the stringent requirements of the consumer electronics and broadcast industry. In this paper we propose fast, efficient, low latency, line scanning based focus propagation, which mitigates the need for complex multigrid or (multilevel) preconditioning techniques. In addition we propose facial blur compensation to compensate for false shading edges that cause incorrect blur estimates in people's faces. In general shading leads to incorrect focus estimates, which may lead to unnatural 3D and visual discomfort. Since visual attention mostly tends to faces, our solution solves the most distracting errors. A subjective assessment by paired comparison on a set of challenging low-depth-of-field images shows that the proposed approach achieves equal 3D image quality as optimization based approaches, and that facial blur compensation results in a significant improvement.

  3. Single-edition quadrangle maps

    USGS Publications Warehouse

    ,

    1998-01-01

    In August 1993, the U.S. Geological Survey's (USGS) National Mapping Division and the U.S. Department of Agriculture's Forest Service signed an Interagency Agreement to begin a single-edition joint mapping program. This agreement established the coordination for producing and maintaining single-edition primary series topographic maps for quadrangles containing National Forest System lands. The joint mapping program saves money by eliminating duplication of effort by the agencies and results in a more frequent revision cycle for quadrangles containing national forests. Maps are revised on the basis of jointly developed standards and contain normal features mapped by the USGS, as well as additional features required for efficient management of National Forest System lands. Single-edition maps look slightly different but meet the content, accuracy, and quality criteria of other USGS products. The Forest Service is responsible for the land management of more than 191 million acres of land throughout the continental United States, Alaska, and Puerto Rico, including 155 national forests and 20 national grasslands. These areas make up the National Forest System lands and comprise more than 10,600 of the 56,000 primary series 7.5-minute quadrangle maps (15-minute in Alaska) covering the United States. The Forest Service has assumed responsibility for maintaining these maps, and the USGS remains responsible for printing and distributing them. Before the agreement, both agencies published similar maps of the same areas. The maps were used for different purposes, but had comparable types of features that were revised at different times. Now, the two products have been combined into one so that the revision cycle is stabilized and only one agency revises the maps, thus increasing the number of current maps available for National Forest System lands. This agreement has improved service to the public by requiring that the agencies share the same maps and that the maps meet a

  4. Subsurface Mapping: A Question of Position and Interpretation

    ERIC Educational Resources Information Center

    Kellie, Andrew C.

    2009-01-01

    This paper discusses the character and challenges inherent in the graphical portrayal of features in subsurface mapping. Subsurface structures are, by their nature, hidden and must be mapped based on drilling and/or geophysical data. Efficient use of graphical techniques is central to effectively communicating the results of expensive exploration…

  5. Transmission and visualization of large geographical maps

    NASA Astrophysics Data System (ADS)

    Zhang, Liqiang; Zhang, Liang; Ren, Yingchao; Guo, Zhifeng

    Transmission and visualization of large geographical maps have become a challenging research issue in GIS applications. This paper presents an efficient and robust way to simplify large geographical maps using frame buffers and Voronoi diagrams. The topological relationships are kept during the simplification by removing the Voronoi diagram's self-overlapped regions. With the simplified vector maps, we establish different levels of detail (LOD) models of these maps. Then we introduce a client/server architecture which integrates our out-of-core algorithm, progressive transmission and rendering scheme based on computer graphics hardware. The architecture allows the viewers to view different regions interactively at different LODs on the network. Experimental results show that our proposed scheme provides an effective way for powerful transmission and manipulation of large maps.

  6. Cadastral Map Assembling Using Generalized Hough Transformation

    NASA Astrophysics Data System (ADS)

    Liu, Fei; Ohyama, Wataru; Wakabayashi, Tetsushi; Kimura, Fumitaka

    There are numerous cadastral maps generated by the past land surveying. The raster digitization of these paper maps is in progress. For effective and efficient use of these maps, we have to assemble the set of maps to make them superimposable on other geographic information in a GIS. The problem can be seen as a complex jigsaw puzzle where the pieces are the cadastral sections extracted from the map. We present an automatic solution to this geographic jigsaw puzzle, based on the generalized Hough transformation that detects the longest common boundary between every piece and its neighbors. The experiments have been conducted using the map of Mie Prefecture, Japan and the French cadastral map. The results of the experiments with the French cadastral maps showed that the proposed method, which consists of a flood filling procedure of internal area and detection and normalization of the north arrow direction, is suitable for assembling the cadastral map. The final goal of the process is to integrate every piece of the puzzle into a national geographic reference frame and database.

  7. Development of predictive mapping techniques for soil survey and salinity mapping

    NASA Astrophysics Data System (ADS)

    Elnaggar, Abdelhamid A.

    Conventional soil maps represent a valuable source of information about soil characteristics, however they are subjective, very expensive, and time-consuming to prepare. Also, they do not include explicit information about the conceptual mental model used in developing them nor information about their accuracy, in addition to the error associated with them. Decision tree analysis (DTA) was successfully used in retrieving the expert knowledge embedded in old soil survey data. This knowledge was efficiently used in developing predictive soil maps for the study areas in Benton and Malheur Counties, Oregon and accessing their consistency. A retrieved soil-landscape model from a reference area in Harney County was extrapolated to develop a preliminary soil map for the neighboring unmapped part of Malheur County. The developed map had a low prediction accuracy and only a few soil map units (SMUs) were predicted with significant accuracy, mostly those shallow SMUs that have either a lithic contact with the bedrock or developed on a duripan. On the other hand, the developed soil map based on field data was predicted with very high accuracy (overall was about 97%). Salt-affected areas of the Malheur County study area are indicated by their high spectral reflectance and they are easily discriminated from the remote sensing data. However, remote sensing data fails to distinguish between the different classes of soil salinity. Using the DTA method, five classes of soil salinity were successfully predicted with an overall accuracy of about 99%. Moreover, the calculated area of salt-affected soil was overestimated when mapped using remote sensing data compared to that predicted by using DTA. Hence, DTA could be a very helpful approach in developing soil survey and soil salinity maps in more objective, effective, less-expensive and quicker ways based on field data.

  8. BOREAS TE-17 Production Efficiency Model Images

    NASA Technical Reports Server (NTRS)

    Hall, Forrest G.; Papagno, Andrea (Editor); Goetz, Scott J.; Goward, Samual N.; Prince, Stephen D.; Czajkowski, Kevin; Dubayah, Ralph O.

    2000-01-01

    A Boreal Ecosystem-Atmospheric Study (BOREAS) version of the Global Production Efficiency Model (http://www.inform.umd.edu/glopem/) was developed by TE-17 (Terrestrial Ecology) to generate maps of gross and net primary production, autotrophic respiration, and light use efficiency for the BOREAS region. This document provides basic information on the model and how the maps were generated. The data generated by the model are stored in binary image-format files. The data files are available on a CD-ROM (see document number 20010000884), or from the Oak Ridge National Laboratory (ORNL) Distributed Active Archive Center (DAAC).

  9. MapReduce in the Cloud: A Use Case Study for Efficient Co-Occurrence Processing of MEDLINE Annotations with MeSH.

    PubMed

    Kreuzthaler, Markus; Miñarro-Giménez, Jose Antonio; Schulz, Stefan

    2016-01-01

    Big data resources are difficult to process without a scaled hardware environment that is specifically adapted to the problem. The emergence of flexible cloud-based virtualization techniques promises solutions to this problem. This paper demonstrates how a billion of lines can be processed in a reasonable amount of time in a cloud-based environment. Our use case addresses the accumulation of concept co-occurrence data in MEDLINE annotation as a series of MapReduce jobs, which can be scaled and executed in the cloud. Besides showing an efficient way solving this problem, we generated an additional resource for the scientific community to be used for advanced text mining approaches.

  10. Mean apparent propagator (MAP) MRI: a novel diffusion imaging method for mapping tissue microstructure.

    PubMed

    Özarslan, Evren; Koay, Cheng Guan; Shepherd, Timothy M; Komlosh, Michal E; İrfanoğlu, M Okan; Pierpaoli, Carlo; Basser, Peter J

    2013-09-01

    Diffusion-weighted magnetic resonance (MR) signals reflect information about underlying tissue microstructure and cytoarchitecture. We propose a quantitative, efficient, and robust mathematical and physical framework for representing diffusion-weighted MR imaging (MRI) data obtained in "q-space," and the corresponding "mean apparent propagator (MAP)" describing molecular displacements in "r-space." We also define and map novel quantitative descriptors of diffusion that can be computed robustly using this MAP-MRI framework. We describe efficient analytical representation of the three-dimensional q-space MR signal in a series expansion of basis functions that accurately describes diffusion in many complex geometries. The lowest order term in this expansion contains a diffusion tensor that characterizes the Gaussian displacement distribution, equivalent to diffusion tensor MRI (DTI). Inclusion of higher order terms enables the reconstruction of the true average propagator whose projection onto the unit "displacement" sphere provides an orientational distribution function (ODF) that contains only the orientational dependence of the diffusion process. The representation characterizes novel features of diffusion anisotropy and the non-Gaussian character of the three-dimensional diffusion process. Other important measures this representation provides include the return-to-the-origin probability (RTOP), and its variants for diffusion in one- and two-dimensions-the return-to-the-plane probability (RTPP), and the return-to-the-axis probability (RTAP), respectively. These zero net displacement probabilities measure the mean compartment (pore) volume and cross-sectional area in distributions of isolated pores irrespective of the pore shape. MAP-MRI represents a new comprehensive framework to model the three-dimensional q-space signal and transform it into diffusion propagators. Experiments on an excised marmoset brain specimen demonstrate that MAP-MRI provides several novel

  11. Cerebral metabolic rate of oxygen (CMRO2 ) mapping with hyperventilation challenge using quantitative susceptibility mapping (QSM).

    PubMed

    Zhang, Jingwei; Zhou, Dong; Nguyen, Thanh D; Spincemaille, Pascal; Gupta, Ajay; Wang, Yi

    2017-05-01

    Our objective was to demonstrate the feasibility of using hyperventilation as an efficient vasoconstrictive challenge and prior knowledge as denoising constraints for cerebral metabolic rate of oxygen (CMRO 2 ) mapping based upon quantitative susceptibility mapping (QSM). Three-dimensional (3D) multi-echo gradient echo and arterial spin labeling imaging were performed to calculate QSM and perfusion maps before and after a hyperventilation challenge in 11 healthy subjects. For comparison, this was repeated using a caffeine challenge. Whole-brain CMRO 2 and oxygen extraction fraction (OEF) maps were computed using constrained optimization. Hyperventilation scans were repeated to measure reproducibility. Regional agreement of CMRO 2 and OEF maps was analyzed within the cortical gray matter (CGM) using t-test and Bland-Altman plots. Hyperventilation challenge eliminates the 30-min waiting time needed for caffeine to exert its vasoconstrictive effects. Mean CMRO 2 (in µmol/100g/min) obtained in CGM using the caffeine and repeated hyperventilation scans were 149 ± 16, 153 ± 19, and 150 ± 20, respectively. This corresponded to an OEF of 33.6 ± 3.4%, 32.3 ± 3.2%, and 34.1 ± 3.8% at baseline state and 39.8 ± 4.8%, 43.6 ± 6.2%, and 42.8 ± 6.8% at challenged state, respectively. Hyperventilation scans produced a good agreement of CMRO 2 and OEF values. Hyperventilation is a feasible, reproducible, and efficient vasoconstrictive challenge for QSM-based quantitative CMRO 2 mapping. Magn Reson Med 77:1762-1773, 2017. © 2016 International Society for Magnetic Resonance in Medicine. © 2016 International Society for Magnetic Resonance in Medicine.

  12. High density genetic linkage map and bin mapping for disease resistance QTLs in peanut

    USDA-ARS?s Scientific Manuscript database

    Mapping and identification of QTLs are important for efficient marker-assisted breeding and for analysis of the molecular mechanisms regulating traits. Diseases, such as early and late leaf spots, Tomato spotted wilt virus (TSWV), cause significant loses to peanut growers. Our goal is to develop a h...

  13. Genome Maps, a new generation genome browser.

    PubMed

    Medina, Ignacio; Salavert, Francisco; Sanchez, Rubén; de Maria, Alejandro; Alonso, Roberto; Escobar, Pablo; Bleda, Marta; Dopazo, Joaquín

    2013-07-01

    Genome browsers have gained importance as more genomes and related genomic information become available. However, the increase of information brought about by new generation sequencing technologies is, at the same time, causing a subtle but continuous decrease in the efficiency of conventional genome browsers. Here, we present Genome Maps, a genome browser that implements an innovative model of data transfer and management. The program uses highly efficient technologies from the new HTML5 standard, such as scalable vector graphics, that optimize workloads at both server and client sides and ensure future scalability. Thus, data management and representation are entirely carried out by the browser, without the need of any Java Applet, Flash or other plug-in technology installation. Relevant biological data on genes, transcripts, exons, regulatory features, single-nucleotide polymorphisms, karyotype and so forth, are imported from web services and are available as tracks. In addition, several DAS servers are already included in Genome Maps. As a novelty, this web-based genome browser allows the local upload of huge genomic data files (e.g. VCF or BAM) that can be dynamically visualized in real time at the client side, thus facilitating the management of medical data affected by privacy restrictions. Finally, Genome Maps can easily be integrated in any web application by including only a few lines of code. Genome Maps is an open source collaborative initiative available in the GitHub repository (https://github.com/compbio-bigdata-viz/genome-maps). Genome Maps is available at: http://www.genomemaps.org.

  14. Genome Maps, a new generation genome browser

    PubMed Central

    Medina, Ignacio; Salavert, Francisco; Sanchez, Rubén; de Maria, Alejandro; Alonso, Roberto; Escobar, Pablo; Bleda, Marta; Dopazo, Joaquín

    2013-01-01

    Genome browsers have gained importance as more genomes and related genomic information become available. However, the increase of information brought about by new generation sequencing technologies is, at the same time, causing a subtle but continuous decrease in the efficiency of conventional genome browsers. Here, we present Genome Maps, a genome browser that implements an innovative model of data transfer and management. The program uses highly efficient technologies from the new HTML5 standard, such as scalable vector graphics, that optimize workloads at both server and client sides and ensure future scalability. Thus, data management and representation are entirely carried out by the browser, without the need of any Java Applet, Flash or other plug-in technology installation. Relevant biological data on genes, transcripts, exons, regulatory features, single-nucleotide polymorphisms, karyotype and so forth, are imported from web services and are available as tracks. In addition, several DAS servers are already included in Genome Maps. As a novelty, this web-based genome browser allows the local upload of huge genomic data files (e.g. VCF or BAM) that can be dynamically visualized in real time at the client side, thus facilitating the management of medical data affected by privacy restrictions. Finally, Genome Maps can easily be integrated in any web application by including only a few lines of code. Genome Maps is an open source collaborative initiative available in the GitHub repository (https://github.com/compbio-bigdata-viz/genome-maps). Genome Maps is available at: http://www.genomemaps.org. PMID:23748955

  15. Handling Data Skew in MapReduce Cluster by Using Partition Tuning

    PubMed

    Gao, Yufei; Zhou, Yanjie; Zhou, Bing; Shi, Lei; Zhang, Jiacai

    2017-01-01

    The healthcare industry has generated large amounts of data, and analyzing these has emerged as an important problem in recent years. The MapReduce programming model has been successfully used for big data analytics. However, data skew invariably occurs in big data analytics and seriously affects efficiency. To overcome the data skew problem in MapReduce, we have in the past proposed a data processing algorithm called Partition Tuning-based Skew Handling (PTSH). In comparison with the one-stage partitioning strategy used in the traditional MapReduce model, PTSH uses a two-stage strategy and the partition tuning method to disperse key-value pairs in virtual partitions and recombines each partition in case of data skew. The robustness and efficiency of the proposed algorithm were tested on a wide variety of simulated datasets and real healthcare datasets. The results showed that PTSH algorithm can handle data skew in MapReduce efficiently and improve the performance of MapReduce jobs in comparison with the native Hadoop, Closer, and locality-aware and fairness-aware key partitioning (LEEN). We also found that the time needed for rule extraction can be reduced significantly by adopting the PTSH algorithm, since it is more suitable for association rule mining (ARM) on healthcare data. © 2017 Yufei Gao et al.

  16. Handling Data Skew in MapReduce Cluster by Using Partition Tuning.

    PubMed

    Gao, Yufei; Zhou, Yanjie; Zhou, Bing; Shi, Lei; Zhang, Jiacai

    2017-01-01

    The healthcare industry has generated large amounts of data, and analyzing these has emerged as an important problem in recent years. The MapReduce programming model has been successfully used for big data analytics. However, data skew invariably occurs in big data analytics and seriously affects efficiency. To overcome the data skew problem in MapReduce, we have in the past proposed a data processing algorithm called Partition Tuning-based Skew Handling (PTSH). In comparison with the one-stage partitioning strategy used in the traditional MapReduce model, PTSH uses a two-stage strategy and the partition tuning method to disperse key-value pairs in virtual partitions and recombines each partition in case of data skew. The robustness and efficiency of the proposed algorithm were tested on a wide variety of simulated datasets and real healthcare datasets. The results showed that PTSH algorithm can handle data skew in MapReduce efficiently and improve the performance of MapReduce jobs in comparison with the native Hadoop, Closer, and locality-aware and fairness-aware key partitioning (LEEN). We also found that the time needed for rule extraction can be reduced significantly by adopting the PTSH algorithm, since it is more suitable for association rule mining (ARM) on healthcare data.

  17. Handling Data Skew in MapReduce Cluster by Using Partition Tuning

    PubMed Central

    Zhou, Yanjie; Zhou, Bing; Shi, Lei

    2017-01-01

    The healthcare industry has generated large amounts of data, and analyzing these has emerged as an important problem in recent years. The MapReduce programming model has been successfully used for big data analytics. However, data skew invariably occurs in big data analytics and seriously affects efficiency. To overcome the data skew problem in MapReduce, we have in the past proposed a data processing algorithm called Partition Tuning-based Skew Handling (PTSH). In comparison with the one-stage partitioning strategy used in the traditional MapReduce model, PTSH uses a two-stage strategy and the partition tuning method to disperse key-value pairs in virtual partitions and recombines each partition in case of data skew. The robustness and efficiency of the proposed algorithm were tested on a wide variety of simulated datasets and real healthcare datasets. The results showed that PTSH algorithm can handle data skew in MapReduce efficiently and improve the performance of MapReduce jobs in comparison with the native Hadoop, Closer, and locality-aware and fairness-aware key partitioning (LEEN). We also found that the time needed for rule extraction can be reduced significantly by adopting the PTSH algorithm, since it is more suitable for association rule mining (ARM) on healthcare data. PMID:29065568

  18. Mapping biological ideas: Concept maps as knowledge integration tools for evolution education

    NASA Astrophysics Data System (ADS)

    Schwendimann, Beat Adrian

    -specific form of concept map, called Knowledge Integration Map (KIM), which aims to help learners connect ideas across levels (for example, genotype and phenotype levels) towards an integrated understanding of evolution. Using a design-based research approach (Brown, 1992; Cobb et al., 2003), three iterative studies were implemented in ethically and economically diverse public high schools classrooms using the web-based inquiry science environment (WISE) (Linn et al., 2003; Linn et al., 2004). Study 1 investigates concept maps as generative assessment tools. Study 1A compares the concept map generation and critique process of biology novices and experts. Findings suggest that concept maps are sensitive to different levels of knowledge integration but require scaffolding and revision. Study 1B investigates the implementation of concept maps as summative assessment tools in a WISE evolution module. Results indicate that concept maps can reveal connections between students' alternative ideas of evolution. Study 2 introduces KIMs as embedded collaborative learning tools. After generating KIMs, student dyads revise KIMs through two different critique activities (comparison against an expert or peer generated KIM). Findings indicate that different critique activities can promote the use of different criteria for critique. Results suggest that the combination of generating and critiquing KIMs can support integrating evolution ideas but can be time-consuming. As time in biology classrooms is limited, study 3 distinguishes the learning effects from either generating or critiquing KIMs as more time efficient embedded learning tools. Findings suggest that critiquing KIMs can be more time efficient than generating KIMs. Using KIMs that include common alternative ideas for critique activities can create genuine opportunities for students to critically reflect on new and existing ideas. Critiquing KIMs can encourage knowledge integration by fostering self-monitoring of students' learning

  19. A Simple Secure Hash Function Scheme Using Multiple Chaotic Maps

    NASA Astrophysics Data System (ADS)

    Ahmad, Musheer; Khurana, Shruti; Singh, Sushmita; AlSharari, Hamed D.

    2017-06-01

    The chaotic maps posses high parameter sensitivity, random-like behavior and one-way computations, which favor the construction of cryptographic hash functions. In this paper, we propose to present a novel hash function scheme which uses multiple chaotic maps to generate efficient variable-sized hash functions. The message is divided into four parts, each part is processed by a different 1D chaotic map unit yielding intermediate hash code. The four codes are concatenated to two blocks, then each block is processed through 2D chaotic map unit separately. The final hash value is generated by combining the two partial hash codes. The simulation analyses such as distribution of hashes, statistical properties of confusion and diffusion, message and key sensitivity, collision resistance and flexibility are performed. The results reveal that the proposed anticipated hash scheme is simple, efficient and holds comparable capabilities when compared with some recent chaos-based hash algorithms.

  20. Multiresolution saliency map based object segmentation

    NASA Astrophysics Data System (ADS)

    Yang, Jian; Wang, Xin; Dai, ZhenYou

    2015-11-01

    Salient objects' detection and segmentation are gaining increasing research interest in recent years. A saliency map can be obtained from different models presented in previous studies. Based on this saliency map, the most salient region (MSR) in an image can be extracted. This MSR, generally a rectangle, can be used as the initial parameters for object segmentation algorithms. However, to our knowledge, all of those saliency maps are represented in a unitary resolution although some models have even introduced multiscale principles in the calculation process. Furthermore, some segmentation methods, such as the well-known GrabCut algorithm, need more iteration time or additional interactions to get more precise results without predefined pixel types. A concept of a multiresolution saliency map is introduced. This saliency map is provided in a multiresolution format, which naturally follows the principle of the human visual mechanism. Moreover, the points in this map can be utilized to initialize parameters for GrabCut segmentation by labeling the feature pixels automatically. Both the computing speed and segmentation precision are evaluated. The results imply that this multiresolution saliency map-based object segmentation method is simple and efficient.

  1. Mapping the entangled ontology of science teachers' lived experience

    NASA Astrophysics Data System (ADS)

    Daugbjerg, Peer S.; de Freitas, Elizabeth; Valero, Paola

    2015-09-01

    In this paper we investigate how the bodily activity of teaching, along with the embodied aspect of lived experience, relates to science teachers' ways of dealing with bodies as living organisms which are both the subject matter as well as the site or vehicle of learning. More precisely, the following questions are pursued: (1) In what ways do primary science teachers refer to the lived and living body in teaching and learning? (2) In what ways do primary science teachers tap into past experiences in which the body figured prominently in order to teach students about living organisms? We draw on the relational ontology and intra-action of Karen Barad (J Women Cult Soc 28(3): 801, 2003) as she argues for a "relational ontology" that sees a relation as a dynamic flowing entanglement of a matter and meaning. We combine this with the materialist phenomenological studies of embodiment by SungWon Hwang and Wolff-Michael Roth (Scientific and mathematical bodies, Sense Publishers, Rotterdam, 2011), as they address how the teachers and students are present in the classroom with/in their "living and lived bodies". Our aim is to use theoretical insights from these two different but complementary approaches to map the embodiment of teachers' experiences and actions. We build our understanding of experience on the work of John Dewey (Experience and education, Simon & Schuster, New York, 1938) and also Jean Clandinin and Michael Connelly (Handbook of qualitative research, Sage Publications, California, 2000), leading us to propose three dimensions: settings, relations and continuity. This means that bodies and settings are mutually entailed in the present relation, and furthermore that the past as well as the present of these bodies and settings—their continuity—is also part of the present relation. We analyse the entanglement of lived experience and embodied teaching using these three proposed dimensions of experience. Analysing interviews and observations of three Danish

  2. Constructing linkage maps in the genomics era with MapDisto 2.0.

    PubMed

    Heffelfinger, Christopher; Fragoso, Christopher A; Lorieux, Mathias

    2017-07-15

    Genotyping by sequencing (GBS) generates datasets that are challenging to handle by current genetic mapping software with graphical interface. Geneticists need new user-friendly computer programs that can analyze GBS data on desktop computers. This requires improvements in computation efficiency, both in terms of speed and use of random-access memory (RAM). MapDisto v.2.0 is a user-friendly computer program for construction of genetic linkage maps. It includes several new major features: (i) handling of very large genotyping datasets like the ones generated by GBS; (ii) direct importation and conversion of Variant Call Format (VCF) files; (iii) detection of linkage, i.e. construction of linkage groups in case of segregation distortion; (iv) data imputation on VCF files using a new approach, called LB-Impute. Features i to iv operate through inclusion of new Java modules that are used transparently by MapDisto; (v) QTL detection via a new R/qtl graphical interface. The program is available free of charge at mapdisto.free.fr. mapdisto@gmail.com. Supplementary data are available at Bioinformatics online. © The Author (2017). Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  3. Construction of Ultradense Linkage Maps with Lep-MAP2: Stickleback F2 Recombinant Crosses as an Example

    PubMed Central

    Rastas, Pasi; Calboli, Federico C. F.; Guo, Baocheng; Shikano, Takahito; Merilä, Juha

    2016-01-01

    High-density linkage maps are important tools for genome biology and evolutionary genetics by quantifying the extent of recombination, linkage disequilibrium, and chromosomal rearrangements across chromosomes, sexes, and populations. They provide one of the best ways to validate and refine de novo genome assemblies, with the power to identify errors in assemblies increasing with marker density. However, assembly of high-density linkage maps is still challenging due to software limitations. We describe Lep-MAP2, a software for ultradense genome-wide linkage map construction. Lep-MAP2 can handle various family structures and can account for achiasmatic meiosis to gain linkage map accuracy. Simulations show that Lep-MAP2 outperforms other available mapping software both in computational efficiency and accuracy. When applied to two large F2-generation recombinant crosses between two nine-spined stickleback (Pungitius pungitius) populations, it produced two high-density (∼6 markers/cM) linkage maps containing 18,691 and 20,054 single nucleotide polymorphisms. The two maps showed a high degree of synteny, but female maps were 1.5–2 times longer than male maps in all linkage groups, suggesting genome-wide recombination suppression in males. Comparison with the genome sequence of the three-spined stickleback (Gasterosteus aculeatus) revealed a high degree of interspecific synteny with a low frequency (<5%) of interchromosomal rearrangements. However, a fairly large (ca. 10 Mb) translocation from autosome to sex chromosome was detected in both maps. These results illustrate the utility and novel features of Lep-MAP2 in assembling high-density linkage maps, and their usefulness in revealing evolutionarily interesting properties of genomes, such as strong genome-wide sex bias in recombination rates. PMID:26668116

  4. An efficient self-organizing map designed by genetic algorithms for the traveling salesman problem.

    PubMed

    Jin, Hui-Dong; Leung, Kwong-Sak; Wong, Man-Leung; Xu, Z B

    2003-01-01

    As a typical combinatorial optimization problem, the traveling salesman problem (TSP) has attracted extensive research interest. In this paper, we develop a self-organizing map (SOM) with a novel learning rule. It is called the integrated SOM (ISOM) since its learning rule integrates the three learning mechanisms in the SOM literature. Within a single learning step, the excited neuron is first dragged toward the input city, then pushed to the convex hull of the TSP, and finally drawn toward the middle point of its two neighboring neurons. A genetic algorithm is successfully specified to determine the elaborate coordination among the three learning mechanisms as well as the suitable parameter setting. The evolved ISOM (eISOM) is examined on three sets of TSP to demonstrate its power and efficiency. The computation complexity of the eISOM is quadratic, which is comparable to other SOM-like neural networks. Moreover, the eISOM can generate more accurate solutions than several typical approaches for TSP including the SOM developed by Budinich, the expanding SOM, the convex elastic net, and the FLEXMAP algorithm. Though its solution accuracy is not yet comparable to some sophisticated heuristics, the eISOM is one of the most accurate neural networks for the TSP.

  5. Clustering of color map pixels: an interactive approach

    NASA Astrophysics Data System (ADS)

    Moon, Yiu Sang; Luk, Franklin T.; Yuen, K. N.; Yeung, Hoi Wo

    2003-12-01

    The demand for digital maps continues to arise as mobile electronic devices become more popular nowadays. Instead of creating the entire map from void, we may convert a scanned paper map into a digital one. Color clustering is the very first step of the conversion process. Currently, most of the existing clustering algorithms are fully automatic. They are fast and efficient but may not work well in map conversion because of the numerous ambiguous issues associated with printed maps. Here we introduce two interactive approaches for color clustering on the map: color clustering with pre-calculated index colors (PCIC) and color clustering with pre-calculated color ranges (PCCR). We also introduce a memory model that could enhance and integrate different image processing techniques for fine-tuning the clustering results. Problems and examples of the algorithms are discussed in the paper.

  6. [Application of mind map in teaching of medical parasitology].

    PubMed

    Zhou, Hong-Chang; Shao, Sheng-Wen; Xu, Bo-Ying

    2012-12-30

    To improve the teaching quality of medical parasitology, mind map, a simple and effective learning method, was introduced. The mind map of each chapter was drawn by teacher and distributed to students before the class. It was helpful for teacher to straighten out the teaching idea, and for students to grasp the important learning points, perfect the class notes and improve learning efficiency. The divergent characteristics of mind map can also help to develop the students' innovation ability.

  7. Systems and methods that generate height map models for efficient three dimensional reconstruction from depth information

    DOEpatents

    Frahm, Jan-Michael; Pollefeys, Marc Andre Leon; Gallup, David Robert

    2015-12-08

    Methods of generating a three dimensional representation of an object in a reference plane from a depth map including distances from a reference point to pixels in an image of the object taken from a reference point. Weights are assigned to respective voxels in a three dimensional grid along rays extending from the reference point through the pixels in the image based on the distances in the depth map from the reference point to the respective pixels, and a height map including an array of height values in the reference plane is formed based on the assigned weights. An n-layer height map may be constructed by generating a probabilistic occupancy grid for the voxels and forming an n-dimensional height map comprising an array of layer height values in the reference plane based on the probabilistic occupancy grid.

  8. Predefined Redundant Dictionary for Effective Depth Maps Representation

    NASA Astrophysics Data System (ADS)

    Sebai, Dorsaf; Chaieb, Faten; Ghorbel, Faouzi

    2016-01-01

    The multi-view video plus depth (MVD) video format consists of two components: texture and depth map, where a combination of these components enables a receiver to generate arbitrary virtual views. However, MVD presents a very voluminous video format that requires a compression process for storage and especially for transmission. Conventional codecs are perfectly efficient for texture images compression but not for intrinsic depth maps properties. Depth images indeed are characterized by areas of smoothly varying grey levels separated by sharp discontinuities at the position of object boundaries. Preserving these characteristics is important to enable high quality view synthesis at the receiver side. In this paper, sparse representation of depth maps is discussed. It is shown that a significant gain in sparsity is achieved when particular mixed dictionaries are used for approximating these types of images with greedy selection strategies. Experiments are conducted to confirm the effectiveness at producing sparse representations, and competitiveness, with respect to candidate state-of-art dictionaries. Finally, the resulting method is shown to be effective for depth maps compression and represents an advantage over the ongoing 3D high efficiency video coding compression standard, particularly at medium and high bitrates.

  9. Applying the metro map to software development management

    NASA Astrophysics Data System (ADS)

    Aguirregoitia, Amaia; Dolado, J. Javier; Presedo, Concepción

    2010-01-01

    This paper presents MetroMap, a new graphical representation model for controlling and managing the software development process. Metromap uses metaphors and visual representation techniques to explore several key indicators in order to support problem detection and resolution. The resulting visualization addresses diverse management tasks, such as tracking of deviations from the plan, analysis of patterns of failure detection and correction, overall assessment of change management policies, and estimation of product quality. The proposed visualization uses a metaphor with a metro map along with various interactive techniques to represent information concerning the software development process and to deal efficiently with multivariate visual queries. Finally, the paper shows the implementation of the tool in JavaFX with data of a real project and the results of testing the tool with the aforementioned data and users attempting several information retrieval tasks. The conclusion shows the results of analyzing user response time and efficiency using the MetroMap visualization system. The utility of the tool was positively evaluated.

  10. Alternative transitions between existing representations in multi-scale maps

    NASA Astrophysics Data System (ADS)

    Dumont, Marion; Touya, Guillaume; Duchêne, Cécile

    2018-05-01

    Map users may have issues to achieve multi-scale navigation tasks, as cartographic objects may have various representations across scales. We assume that adding intermediate representations could be one way to reduce the differences between existing representations, and to ease the transitions across scales. We consider an existing multiscale map on the scale range from 1 : 25k to 1 : 100k scales. Based on hypotheses about intermediate representations design, we build custom multi-scale maps with alternative transitions. We will conduct in a next future a user evaluation to compare the efficiency of these alternative maps for multi-scale navigation. This paper discusses the hypotheses and production process of these alternative maps.

  11. Machine learning-based dual-energy CT parametric mapping

    NASA Astrophysics Data System (ADS)

    Su, Kuan-Hao; Kuo, Jung-Wen; Jordan, David W.; Van Hedent, Steven; Klahr, Paul; Wei, Zhouping; Helo, Rose Al; Liang, Fan; Qian, Pengjiang; Pereira, Gisele C.; Rassouli, Negin; Gilkeson, Robert C.; Traughber, Bryan J.; Cheng, Chee-Wai; Muzic, Raymond F., Jr.

    2018-06-01

    The aim is to develop and evaluate machine learning methods for generating quantitative parametric maps of effective atomic number (Zeff), relative electron density (ρ e), mean excitation energy (I x ), and relative stopping power (RSP) from clinical dual-energy CT data. The maps could be used for material identification and radiation dose calculation. Machine learning methods of historical centroid (HC), random forest (RF), and artificial neural networks (ANN) were used to learn the relationship between dual-energy CT input data and ideal output parametric maps calculated for phantoms from the known compositions of 13 tissue substitutes. After training and model selection steps, the machine learning predictors were used to generate parametric maps from independent phantom and patient input data. Precision and accuracy were evaluated using the ideal maps. This process was repeated for a range of exposure doses, and performance was compared to that of the clinically-used dual-energy, physics-based method which served as the reference. The machine learning methods generated more accurate and precise parametric maps than those obtained using the reference method. Their performance advantage was particularly evident when using data from the lowest exposure, one-fifth of a typical clinical abdomen CT acquisition. The RF method achieved the greatest accuracy. In comparison, the ANN method was only 1% less accurate but had much better computational efficiency than RF, being able to produce parametric maps in 15 s. Machine learning methods outperformed the reference method in terms of accuracy and noise tolerance when generating parametric maps, encouraging further exploration of the techniques. Among the methods we evaluated, ANN is the most suitable for clinical use due to its combination of accuracy, excellent low-noise performance, and computational efficiency.

  12. Machine learning-based dual-energy CT parametric mapping.

    PubMed

    Su, Kuan-Hao; Kuo, Jung-Wen; Jordan, David W; Van Hedent, Steven; Klahr, Paul; Wei, Zhouping; Al Helo, Rose; Liang, Fan; Qian, Pengjiang; Pereira, Gisele C; Rassouli, Negin; Gilkeson, Robert C; Traughber, Bryan J; Cheng, Chee-Wai; Muzic, Raymond F

    2018-06-08

    The aim is to develop and evaluate machine learning methods for generating quantitative parametric maps of effective atomic number (Z eff ), relative electron density (ρ e ), mean excitation energy (I x ), and relative stopping power (RSP) from clinical dual-energy CT data. The maps could be used for material identification and radiation dose calculation. Machine learning methods of historical centroid (HC), random forest (RF), and artificial neural networks (ANN) were used to learn the relationship between dual-energy CT input data and ideal output parametric maps calculated for phantoms from the known compositions of 13 tissue substitutes. After training and model selection steps, the machine learning predictors were used to generate parametric maps from independent phantom and patient input data. Precision and accuracy were evaluated using the ideal maps. This process was repeated for a range of exposure doses, and performance was compared to that of the clinically-used dual-energy, physics-based method which served as the reference. The machine learning methods generated more accurate and precise parametric maps than those obtained using the reference method. Their performance advantage was particularly evident when using data from the lowest exposure, one-fifth of a typical clinical abdomen CT acquisition. The RF method achieved the greatest accuracy. In comparison, the ANN method was only 1% less accurate but had much better computational efficiency than RF, being able to produce parametric maps in 15 s. Machine learning methods outperformed the reference method in terms of accuracy and noise tolerance when generating parametric maps, encouraging further exploration of the techniques. Among the methods we evaluated, ANN is the most suitable for clinical use due to its combination of accuracy, excellent low-noise performance, and computational efficiency.

  13. Value stream mapping of the Pap test processing procedure: a lean approach to improve quality and efficiency.

    PubMed

    Michael, Claire W; Naik, Kalyani; McVicker, Michael

    2013-05-01

    We developed a value stream map (VSM) of the Papanicolaou test procedure to identify opportunities to reduce waste and errors, created a new VSM, and implemented a new process emphasizing Lean tools. Preimplementation data revealed the following: (1) processing time (PT) for 1,140 samples averaged 54 hours; (2) 27 accessioning errors were detected on review of 357 random requisitions (7.6%); (3) 5 of the 20,060 tests had labeling errors that had gone undetected in the processing stage. Four were detected later during specimen processing but 1 reached the reporting stage. Postimplementation data were as follows: (1) PT for 1,355 samples averaged 31 hours; (2) 17 accessioning errors were detected on review of 385 random requisitions (4.4%); and (3) no labeling errors were undetected. Our results demonstrate that implementation of Lean methods, such as first-in first-out processes and minimizing batch size by staff actively participating in the improvement process, allows for higher quality, greater patient safety, and improved efficiency.

  14. Digital Mapping Techniques '07 - Workshop Proceedings

    USGS Publications Warehouse

    Soller, David R.

    2008-01-01

    The Digital Mapping Techniques '07 (DMT'07) workshop was attended by 85 technical experts from 49 agencies, universities, and private companies, including representatives from 27 state geological surveys. This year's meeting, the tenth in the annual series, was hosted by the South Carolina Geological Survey, from May 20-23, 2007, on the University of South Carolina campus in Columbia, South Carolina. Each DMT workshop has been coordinated by the U.S. Geological Survey's National Geologic Map Database Project and the Association of American State Geologists (AASG). As in previous year's meetings, the objective was to foster informal discussion and exchange of technical information, principally in order to develop more efficient methods for digital mapping, cartography, GIS analysis, and information management. At this meeting, oral and poster presentations and special discussion sessions emphasized: 1) methods for creating and publishing map products (here, 'publishing' includes Web-based release); 2) field data capture software and techniques, including the use of LIDAR; 3) digital cartographic techniques; 4) migration of digital maps into ArcGIS Geodatabase format; 5) analytical GIS techniques; and 6) continued development of the National Geologic Map Database.

  15. Ecology and space: A case study in mapping harmful invasive species

    USGS Publications Warehouse

    David T. Barnett,; Jarnevich, Catherine S.; Chong, Geneva W.; Stohlgren, Thomas J.; Sunil Kumar,; Holcombe, Tracy R.; Brunn, Stanley D.; Dodge, Martin

    2017-01-01

    The establishment and invasion of non-native plant species have the ability to alter the composition of native species and functioning of ecological systems with financial costs resulting from mitigation and loss of ecological services. Spatially documenting invasions has applications for management and theory, but the utility of maps is challenged by availability and uncertainty of data, and the reliability of extrapolating mapped data in time and space. The extent and resolution of projections also impact the ability to inform invasive species science and management. Early invasive species maps were coarse-grained representations that underscored the phenomena, but had limited capacity to direct management aside from development of watch lists for priorities for prevention and containment. Integrating mapped data sets with fine-resolution environmental variables in the context of species-distribution models allows a description of species-environment relationships and an understanding of how, why, and where invasions may occur. As with maps, the extent and resolution of models impact the resulting insight. Models of cheatgrass (Bromus tectorum) across a variety of spatial scales and grain result in divergent species-environment relationships. New data can improve models and efficiently direct further inventories. Mapping can target areas of greater model uncertainty or the bounds of modeled distribution to efficiently refine models and maps. This iterative process results in dynamic, living maps capable of describing the ongoing process of species invasions.

  16. Conditional Random Field-Based Offline Map Matching for Indoor Environments

    PubMed Central

    Bataineh, Safaa; Bahillo, Alfonso; Díez, Luis Enrique; Onieva, Enrique; Bataineh, Ikram

    2016-01-01

    In this paper, we present an offline map matching technique designed for indoor localization systems based on conditional random fields (CRF). The proposed algorithm can refine the results of existing indoor localization systems and match them with the map, using loose coupling between the existing localization system and the proposed map matching technique. The purpose of this research is to investigate the efficiency of using the CRF technique in offline map matching problems for different scenarios and parameters. The algorithm was applied to several real and simulated trajectories of different lengths. The results were then refined and matched with the map using the CRF algorithm. PMID:27537892

  17. Conditional Random Field-Based Offline Map Matching for Indoor Environments.

    PubMed

    Bataineh, Safaa; Bahillo, Alfonso; Díez, Luis Enrique; Onieva, Enrique; Bataineh, Ikram

    2016-08-16

    In this paper, we present an offline map matching technique designed for indoor localization systems based on conditional random fields (CRF). The proposed algorithm can refine the results of existing indoor localization systems and match them with the map, using loose coupling between the existing localization system and the proposed map matching technique. The purpose of this research is to investigate the efficiency of using the CRF technique in offline map matching problems for different scenarios and parameters. The algorithm was applied to several real and simulated trajectories of different lengths. The results were then refined and matched with the map using the CRF algorithm.

  18. $n$ -Dimensional Discrete Cat Map Generation Using Laplace Expansions.

    PubMed

    Wu, Yue; Hua, Zhongyun; Zhou, Yicong

    2016-11-01

    Different from existing methods that use matrix multiplications and have high computation complexity, this paper proposes an efficient generation method of n -dimensional ( [Formula: see text]) Cat maps using Laplace expansions. New parameters are also introduced to control the spatial configurations of the [Formula: see text] Cat matrix. Thus, the proposed method provides an efficient way to mix dynamics of all dimensions at one time. To investigate its implementations and applications, we further introduce a fast implementation algorithm of the proposed method with time complexity O(n 4 ) and a pseudorandom number generator using the Cat map generated by the proposed method. The experimental results show that, compared with existing generation methods, the proposed method has a larger parameter space and simpler algorithm complexity, generates [Formula: see text] Cat matrices with a lower inner correlation, and thus yields more random and unpredictable outputs of [Formula: see text] Cat maps.

  19. Rapid Gynogenetic Mapping of Xenopus tropicalis Mutations to Chromosomes

    PubMed Central

    Khokha, Mustafa K.; Krylov, Vladimir; Reilly, Michael J.; Gall, Joseph G.; Bhattacharya, Dipankan; Cheung, Chung Yan J.; Kaufman, Sarah; Lam, Dang Khoa; Macha, Jaroslav; Ngo, Catherine; Prakash, Neha; Schmidt, Philip; Tlapakova, Tereza; Trivedi, Toral; Tumova, Lucie; Abu-Daya, Anita; Geach, Timothy; Vendrell, Elisenda; Ironfield, Holly; Sinzelle, Ludivine; Sater, Amy K.; Wells, Dan E.; Harland, Richard M.; Zimmerman, Lyle B.

    2010-01-01

    Pilot forward genetic screens in Xenopus tropicalis have isolated over 60 recessive mutations (Grammer et al., 2005; Noramly et al., 2005; Goda et al., 2006). Here we present a simple method for mapping mutations to chromosomes using gynogenesis and centromeric markers. When coupled with available genomic resources, gross mapping facilitates evaluation of candidate genes as well as higher resolution linkage studies. Using gynogenesis, we have mapped the genetic locations of the 10 X. tropicalis centromeres, and performed Fluorescence In Situ Hybridization to validate these locations cytologically. We demonstrate the use of this very small set of centromeric markers to map mutations efficiently to specific chromosomes. PMID:19441086

  20. Exploiting Surroundedness for Saliency Detection: A Boolean Map Approach.

    PubMed

    Zhang, Jianming; Sclaroff, Stan

    2016-05-01

    We demonstrate the usefulness of surroundedness for eye fixation prediction by proposing a Boolean Map based Saliency model (BMS). In our formulation, an image is characterized by a set of binary images, which are generated by randomly thresholding the image's feature maps in a whitened feature space. Based on a Gestalt principle of figure-ground segregation, BMS computes a saliency map by discovering surrounded regions via topological analysis of Boolean maps. Furthermore, we draw a connection between BMS and the Minimum Barrier Distance to provide insight into why and how BMS can properly captures the surroundedness cue via Boolean maps. The strength of BMS is verified by its simplicity, efficiency and superior performance compared with 10 state-of-the-art methods on seven eye tracking benchmark datasets.

  1. Digital Geologic Mapping and Integration with the Geoweb: The Death Knell for Exclusively Paper Geologic Maps

    NASA Astrophysics Data System (ADS)

    House, P. K.

    2008-12-01

    The combination of traditional methods of geologic mapping with rapidly developing web-based geospatial applications ('the geoweb') and the various collaborative opportunities of web 2.0 have the potential to change the nature, value, and relevance of geologic maps and related field studies. Parallel advances in basic GPS technology, digital photography, and related integrative applications provide practicing geologic mappers with greatly enhanced methods for collecting, visualizing, interpreting, and disseminating geologic information. Even a cursory application of available tools can make field and office work more enriching and efficient; whereas more advanced and systematic applications provide new avenues for collaboration, outreach, and public education. Moreover, they ensure a much broader audience among an immense number of internet savvy end-users with very specific expectations for geospatial data availability. Perplexingly, the geologic community as a whole is not fully exploring this opportunity despite the inevitable revolution in portends. The slow acceptance follows a broad generational trend wherein seasoned professionals are lagging behind geology students and recent graduates in their grasp of and interest in the capabilities of the geoweb and web 2.0 types of applications. Possible explanations for this include: fear of the unknown, fear of learning curve, lack of interest, lack of academic/professional incentive, and (hopefully not) reluctance toward open collaboration. Although some aspects of the expanding geoweb are cloaked in arcane computer code, others are extremely simple to understand and use. A particularly obvious and simple application to enhance any field study is photo geotagging, the digital documentation of the locations of key outcrops, illustrative vistas, and particularly complicated geologic field relations. Viewing geotagged photos in their appropriate context on a virtual globe with high-resolution imagery can be an

  2. Impact of Schematic Designs on the Cognition of Underground Tube Maps

    NASA Astrophysics Data System (ADS)

    Liu, Zheng; Li, Zhilin

    2016-06-01

    Schematic maps have been popularly employed to represent transport networks, particularly underground tube lines (or metro lines), since its adoption by the Official London Underground in early 1930s. Such maps employ straightened lines along horizontal, vertical and diagonal directions. Recently, some researchers started to argue that the distortion in such a schematization may cause big distortion and some new designs are proposed. This project aims to make a comparative analysis of such a schematic design with a new design proposed by Mark Noad in 2011, which makes use of lines along 30º and 60º directions instead of the 45º direction. Tasks have been designed for evaluating the effect of schematic designs on route planning by travellers. The participant was asked to choose the route s/he would take among two or three possible route options and then read the name of the selected transfer station. Eye-tracking technique has been employed to track the map recognition process. Total travel time is used as criterion for effectiveness; completion time and mental work cost are used for efficiency evaluation. It has been found that (1) the design of map style has significant impact on users' travel decision making, especially map distance and transfer station symbol designs, and (2) the design style of a schematic map will have great impact on the effectiveness and efficiency of map recognition.

  3. A Bike Built for Magnetic Mapping

    NASA Astrophysics Data System (ADS)

    Schattner, U.; Segev, A.; Lyakhovsky, V.

    2017-12-01

    Understanding the magnetic signature of the subsurface geology is crucial for structural, groundwater, earthquake propagation, and mineral studies. The cheapest measuring method is by walking with sensors. This approach yields high-resolution maps, yet its coverage is limited. We invented a new design that records magnetic data while riding a bicycle. The new concept offers an efficient, low-cost method of collecting high-resolution ground magnetic field data over rough terrain where conventional vehicles dare not venture. It improves the efficiency of the traditional method by more than five times. The Bike-magnetic scales up ground magnetism from a localized site survey to regional coverage. By now we covered 3300 square KM (about the size of Rhode Island) across northern Israel, in profile spacing of 1-2 km. Initial Total Magnetic Intensity maps reveal a myriad of new features that were not detected by the low-resolution regional aeromagnetic survey that collected data from 1000 m height.

  4. To the National Map and beyond

    USGS Publications Warehouse

    Kelmelis, J.

    2003-01-01

    Scientific understanding, technology, and social, economic, and environmental conditions have driven a rapidly changing demand for geographic information, both digital and analog. For more than a decade, the U.S. Geological Survey (USGS) has been developing innovative partnerships with other government agencies and private industry to produce and distribute geographic information efficiently; increase activities in remote sensing to ensure ongoing monitoring of the land surface; and develop new understanding of the causes and consequences of land surface change. These activities are now contributing to a more robust set of geographic information called The National Map (TNM). The National Map is designed to provide an up-to-date, seamless, horizontally and vertically integrated set of basic digital geographic data, a frequent monitoring of changes on the land surface, and an understanding of the condition of the Earth's surface and many of the processes that shape it. The USGS has reorganized its National Mapping Program into three programs to address the continuum of scientific activities-describing (mapping), monitoring, understanding, modeling, and predicting. The Cooperative Topographic Mapping Program focuses primarily on the mapping and revision aspects of TNM. The National Map also includes results from the Land Remote Sensing and Geographic Analysis and Monitoring Programs that provide continual updates, new insights, and analytical tools. The National Map is valuable as a framework for current research, management, and operational activities. It also provides a critical framework for the development of distributed, spatially enabled decision support systems.

  5. A case study of tuning MapReduce for efficient Bioinformatics in the cloud

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shi, Lizhen; Wang, Zhong; Yu, Weikuan

    The combination of the Hadoop MapReduce programming model and cloud computing allows biological scientists to analyze next-generation sequencing (NGS) data in a timely and cost-effective manner. Cloud computing platforms remove the burden of IT facility procurement and management from end users and provide ease of access to Hadoop clusters. However, biological scientists are still expected to choose appropriate Hadoop parameters for running their jobs. More importantly, the available Hadoop tuning guidelines are either obsolete or too general to capture the particular characteristics of bioinformatics applications. In this paper, we aim to minimize the cloud computing cost spent on bioinformatics datamore » analysis by optimizing the extracted significant Hadoop parameters. When using MapReduce-based bioinformatics tools in the cloud, the default settings often lead to resource underutilization and wasteful expenses. We choose k-mer counting, a representative application used in a large number of NGS data analysis tools, as our study case. Experimental results show that, with the fine-tuned parameters, we achieve a total of 4× speedup compared with the original performance (using the default settings). Finally, this paper presents an exemplary case for tuning MapReduce-based bioinformatics applications in the cloud, and documents the key parameters that could lead to significant performance benefits.« less

  6. Intermediate quantum maps for quantum computation

    NASA Astrophysics Data System (ADS)

    Giraud, O.; Georgeot, B.

    2005-10-01

    We study quantum maps displaying spectral statistics intermediate between Poisson and Wigner-Dyson. It is shown that they can be simulated on a quantum computer with a small number of gates, and efficiently yield information about fidelity decay or spectral statistics. We study their matrix elements and entanglement production and show that they converge with time to distributions which differ from random matrix predictions. A randomized version of these maps can be implemented even more economically and yields pseudorandom operators with original properties, enabling, for example, one to produce fractal random vectors. These algorithms are within reach of present-day quantum computers.

  7. Accurate atom-mapping computation for biochemical reactions.

    PubMed

    Latendresse, Mario; Malerich, Jeremiah P; Travers, Mike; Karp, Peter D

    2012-11-26

    The complete atom mapping of a chemical reaction is a bijection of the reactant atoms to the product atoms that specifies the terminus of each reactant atom. Atom mapping of biochemical reactions is useful for many applications of systems biology, in particular for metabolic engineering where synthesizing new biochemical pathways has to take into account for the number of carbon atoms from a source compound that are conserved in the synthesis of a target compound. Rapid, accurate computation of the atom mapping(s) of a biochemical reaction remains elusive despite significant work on this topic. In particular, past researchers did not validate the accuracy of mapping algorithms. We introduce a new method for computing atom mappings called the minimum weighted edit-distance (MWED) metric. The metric is based on bond propensity to react and computes biochemically valid atom mappings for a large percentage of biochemical reactions. MWED models can be formulated efficiently as Mixed-Integer Linear Programs (MILPs). We have demonstrated this approach on 7501 reactions of the MetaCyc database for which 87% of the models could be solved in less than 10 s. For 2.1% of the reactions, we found multiple optimal atom mappings. We show that the error rate is 0.9% (22 reactions) by comparing these atom mappings to 2446 atom mappings of the manually curated Kyoto Encyclopedia of Genes and Genomes (KEGG) RPAIR database. To our knowledge, our computational atom-mapping approach is the most accurate and among the fastest published to date. The atom-mapping data will be available in the MetaCyc database later in 2012; the atom-mapping software will be available within the Pathway Tools software later in 2012.

  8. The MAP Autonomous Mission Control System

    NASA Technical Reports Server (NTRS)

    Breed, Juile; Coyle, Steven; Blahut, Kevin; Dent, Carolyn; Shendock, Robert; Rowe, Roger

    2000-01-01

    The Microwave Anisotropy Probe (MAP) mission is the second mission in NASA's Office of Space Science low-cost, Medium-class Explorers (MIDEX) program. The Explorers Program is designed to accomplish frequent, low cost, high quality space science investigations utilizing innovative, streamlined, efficient management, design and operations approaches. The MAP spacecraft will produce an accurate full-sky map of the cosmic microwave background temperature fluctuations with high sensitivity and angular resolution. The MAP spacecraft is planned for launch in early 2001, and will be staffed by only single-shift operations. During the rest of the time the spacecraft must be operated autonomously, with personnel available only on an on-call basis. Four (4) innovations will work cooperatively to enable a significant reduction in operations costs for the MAP spacecraft. First, the use of a common ground system for Spacecraft Integration and Test (I&T) as well as Operations. Second, the use of Finite State Modeling for intelligent autonomy. Third, the integration of a graphical planning engine to drive the autonomous systems without an intermediate manual step. And fourth, the ability for distributed operations via Web and pager access.

  9. "We Teach Being with Books"

    ERIC Educational Resources Information Center

    Stanistreet, Paul

    2008-01-01

    Life is serious and serious literature can help people through it. That's the message of Jane Davis's remarkable Get into Reading project. Jane Davis's Get into Reading project is one of two main strands of work undertaken by The Reader organization, at Liverpool University. In 1997, Jane and two other continuing education teachers, Sarah Coley…

  10. MIDAS: Regionally linear multivariate discriminative statistical mapping.

    PubMed

    Varol, Erdem; Sotiras, Aristeidis; Davatzikos, Christos

    2018-07-01

    Statistical parametric maps formed via voxel-wise mass-univariate tests, such as the general linear model, are commonly used to test hypotheses about regionally specific effects in neuroimaging cross-sectional studies where each subject is represented by a single image. Despite being informative, these techniques remain limited as they ignore multivariate relationships in the data. Most importantly, the commonly employed local Gaussian smoothing, which is important for accounting for registration errors and making the data follow Gaussian distributions, is usually chosen in an ad hoc fashion. Thus, it is often suboptimal for the task of detecting group differences and correlations with non-imaging variables. Information mapping techniques, such as searchlight, which use pattern classifiers to exploit multivariate information and obtain more powerful statistical maps, have become increasingly popular in recent years. However, existing methods may lead to important interpretation errors in practice (i.e., misidentifying a cluster as informative, or failing to detect truly informative voxels), while often being computationally expensive. To address these issues, we introduce a novel efficient multivariate statistical framework for cross-sectional studies, termed MIDAS, seeking highly sensitive and specific voxel-wise brain maps, while leveraging the power of regional discriminant analysis. In MIDAS, locally linear discriminative learning is applied to estimate the pattern that best discriminates between two groups, or predicts a variable of interest. This pattern is equivalent to local filtering by an optimal kernel whose coefficients are the weights of the linear discriminant. By composing information from all neighborhoods that contain a given voxel, MIDAS produces a statistic that collectively reflects the contribution of the voxel to the regional classifiers as well as the discriminative power of the classifiers. Critically, MIDAS efficiently assesses the

  11. The Relation of Knowledge-Text Integration Processes and Reading Comprehension in 7th- to 12th-Grade Students

    ERIC Educational Resources Information Center

    Barnes, Marcia A.; Ahmed, Yusra; Barth, Amy; Francis, David J.

    2015-01-01

    The integration of knowledge during reading was tested in 1,109 secondary school students. Reading times for the second sentence in a pair (Jane's headache went away) were compared in conditions where the first sentence was either causally or temporally related to the first sentence (Jane took an aspirin vs. Jane looked for an aspirin).…

  12. Beebook: light field mapping app

    NASA Astrophysics Data System (ADS)

    De Donatis, Mauro; Di Pietro, Gianfranco; Rinnone, Fabio

    2014-05-01

    In the last decade the mobile systems for field digital mapping were developed (see Wikipedia for "Digital geologic mapping"), also against many skeptic traditional geologists. Until now, hardware was often heavy (tablet PC) and software sometime difficult also for expert GIS users. At present, the advent of light tablet and applications makes things easier, but we are far to find a whole solution for a complex survey like the geological one where you have to manage complexities such information, hypothesis, data, interpretation. Beebook is a new app for Android devices, has been developed for fast ad easy mapping work in the field trying to try to solve this problem. The main features are: • off-line raster management, GeoTIFF ed other raster format using; • on-line map visualisation (Google Maps, OSM, WMS, WFS); • SR management and conversion using PROJ.4; • vector file mash-up (KML and SQLite format); • editing of vector data on the map (lines, points, polygons); • augmented reality using "Mixare" platform; • export of vector data in KML, CSV, SQLite (Spatialite) format; • note: GPS or manual point inserting linked to other application files (pictures, spreadsheet, etc.); • form: creation, edition and filling of customized form; • GPS: status control, tracker and positioning on map; • sharing: synchronization and sharing of data, forms, positioning and other information can be done among users. The input methods are different from digital keyboard to fingers touch, from voice recording to stylus. In particular the most efficient way of inserting information is the stylus (or pen): field geologists are familiar with annotation and sketches. Therefore we suggest the use of devices with stylus. The main point is that Beebook is the first "transparent" mobile GIS for tablet and smartphone deriving from previous experience as traditional mapping and different previous digital mapping software ideation and development (MapIT, BeeGIS, Geopaparazzi

  13. Determination of a Limited Scope Network's Lightning Detection Efficiency

    NASA Technical Reports Server (NTRS)

    Rompala, John T.; Blakeslee, R.

    2008-01-01

    This paper outlines a modeling technique to map lightning detection efficiency variations over a region surveyed by a sparse array of ground based detectors. A reliable flash peak current distribution (PCD) for the region serves as the technique's base. This distribution is recast as an event probability distribution function. The technique then uses the PCD together with information regarding: site signal detection thresholds, type of solution algorithm used, and range attenuation; to formulate the probability that a flash at a specified location will yield a solution. Applying this technique to the full region produces detection efficiency contour maps specific to the parameters employed. These contours facilitate a comparative analysis of each parameter's effect on the network's detection efficiency. In an alternate application, this modeling technique gives an estimate of the number, strength, and distribution of events going undetected. This approach leads to a variety of event density contour maps. This application is also illustrated. The technique's base PCD can be empirical or analytical. A process for formulating an empirical PCD specific to the region and network being studied is presented. A new method for producing an analytical representation of the empirical PCD is also introduced.

  14. Earth mapping - aerial or satellite imagery comparative analysis

    NASA Astrophysics Data System (ADS)

    Fotev, Svetlin; Jordanov, Dimitar; Lukarski, Hristo

    Nowadays, solving the tasks for revision of existing map products and creation of new maps requires making a choice of the land cover image source. The issue of the effectiveness and cost of the usage of aerial mapping systems versus the efficiency and cost of very-high resolution satellite imagery is topical [1, 2, 3, 4]. The price of any remotely sensed image depends on the product (panchromatic or multispectral), resolution, processing level, scale, urgency of task and on whether the needed image is available in the archive or has to be requested. The purpose of the present work is: to make a comparative analysis between the two approaches for mapping the Earth having in mind two parameters: quality and cost. To suggest an approach for selection of the map information sources - airplane-based or spacecraft-based imaging systems with very-high spatial resolution. Two cases are considered: area that equals approximately one satellite scene and area that equals approximately the territory of Bulgaria.

  15. Efficiently modeling neural networks on massively parallel computers

    NASA Technical Reports Server (NTRS)

    Farber, Robert M.

    1993-01-01

    Neural networks are a very useful tool for analyzing and modeling complex real world systems. Applying neural network simulations to real world problems generally involves large amounts of data and massive amounts of computation. To efficiently handle the computational requirements of large problems, we have implemented at Los Alamos a highly efficient neural network compiler for serial computers, vector computers, vector parallel computers, and fine grain SIMD computers such as the CM-2 connection machine. This paper describes the mapping used by the compiler to implement feed-forward backpropagation neural networks for a SIMD (Single Instruction Multiple Data) architecture parallel computer. Thinking Machines Corporation has benchmarked our code at 1.3 billion interconnects per second (approximately 3 gigaflops) on a 64,000 processor CM-2 connection machine (Singer 1990). This mapping is applicable to other SIMD computers and can be implemented on MIMD computers such as the CM-5 connection machine. Our mapping has virtually no communications overhead with the exception of the communications required for a global summation across the processors (which has a sub-linear runtime growth on the order of O(log(number of processors)). We can efficiently model very large neural networks which have many neurons and interconnects and our mapping can extend to arbitrarily large networks (within memory limitations) by merging the memory space of separate processors with fast adjacent processor interprocessor communications. This paper will consider the simulation of only feed forward neural network although this method is extendable to recurrent networks.

  16. A method to correct coordinate distortion in EBSD maps

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, Y.B., E-mail: yubz@dtu.dk; Elbrønd, A.; Lin, F.X.

    2014-10-15

    Drift during electron backscatter diffraction mapping leads to coordinate distortions in resulting orientation maps, which affects, in some cases significantly, the accuracy of analysis. A method, thin plate spline, is introduced and tested to correct such coordinate distortions in the maps after the electron backscatter diffraction measurements. The accuracy of the correction as well as theoretical and practical aspects of using the thin plate spline method is discussed in detail. By comparing with other correction methods, it is shown that the thin plate spline method is most efficient to correct different local distortions in the electron backscatter diffraction maps. -more » Highlights: • A new method is suggested to correct nonlinear spatial distortion in EBSD maps. • The method corrects EBSD maps more precisely than presently available methods. • Errors less than 1–2 pixels are typically obtained. • Direct quantitative analysis of dynamic data are available after this correction.« less

  17. Single-molecule optical genome mapping of a human HapMap and a colorectal cancer cell line.

    PubMed

    Teo, Audrey S M; Verzotto, Davide; Yao, Fei; Nagarajan, Niranjan; Hillmer, Axel M

    2015-01-01

    Next-generation sequencing (NGS) technologies have changed our understanding of the variability of the human genome. However, the identification of genome structural variations based on NGS approaches with read lengths of 35-300 bases remains a challenge. Single-molecule optical mapping technologies allow the analysis of DNA molecules of up to 2 Mb and as such are suitable for the identification of large-scale genome structural variations, and for de novo genome assemblies when combined with short-read NGS data. Here we present optical mapping data for two human genomes: the HapMap cell line GM12878 and the colorectal cancer cell line HCT116. High molecular weight DNA was obtained by embedding GM12878 and HCT116 cells, respectively, in agarose plugs, followed by DNA extraction under mild conditions. Genomic DNA was digested with KpnI and 310,000 and 296,000 DNA molecules (≥ 150 kb and 10 restriction fragments), respectively, were analyzed per cell line using the Argus optical mapping system. Maps were aligned to the human reference by OPTIMA, a new glocal alignment method. Genome coverage of 6.8× and 5.7× was obtained, respectively; 2.9× and 1.7× more than the coverage obtained with previously available software. Optical mapping allows the resolution of large-scale structural variations of the genome, and the scaffold extension of NGS-based de novo assemblies. OPTIMA is an efficient new alignment method; our optical mapping data provide a resource for genome structure analyses of the human HapMap reference cell line GM12878, and the colorectal cancer cell line HCT116.

  18. An image-space parallel convolution filtering algorithm based on shadow map

    NASA Astrophysics Data System (ADS)

    Li, Hua; Yang, Huamin; Zhao, Jianping

    2017-07-01

    Shadow mapping is commonly used in real-time rendering. In this paper, we presented an accurate and efficient method of soft shadows generation from planar area lights. First this method generated a depth map from light's view, and analyzed the depth-discontinuities areas as well as shadow boundaries. Then these areas were described as binary values in the texture map called binary light-visibility map, and a parallel convolution filtering algorithm based on GPU was enforced to smooth out the boundaries with a box filter. Experiments show that our algorithm is an effective shadow map based method that produces perceptually accurate soft shadows in real time with more details of shadow boundaries compared with the previous works.

  19. Utilizing HDF4 File Content Maps for the Cloud

    NASA Technical Reports Server (NTRS)

    Lee, Hyokyung Joe

    2016-01-01

    We demonstrate a prototype study that HDF4 file content map can be used for efficiently organizing data in cloud object storage system to facilitate cloud computing. This approach can be extended to any binary data formats and to any existing big data analytics solution powered by cloud computing because HDF4 file content map project started as long term preservation of NASA data that doesn't require HDF4 APIs to access data.

  20. Building America Top Innovations 2012: Building Science-Based Climate Maps

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    none,

    2013-01-01

    This Building America Top Innovations profile describes the Building America-developed climate zone map, which serves as a consistent framework for energy-efficiency requirements in the national model energy code starting with the 2004 IECC Supplement and the ASHRAE 90.1 2004 edition. The map also provides a critical foundation for climate-specific guidance in the widely disseminated EEBA Builder Guides and Building America Best Practice Guides.

  1. Trajectory Segmentation Map-Matching Approach for Large-Scale, High-Resolution GPS Data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhu, Lei; Holden, Jacob R.; Gonder, Jeffrey D.

    With the development of smartphones and portable GPS devices, large-scale, high-resolution GPS data can be collected. Map matching is a critical step in studying vehicle driving activity and recognizing network traffic conditions from the data. A new trajectory segmentation map-matching algorithm is proposed to deal accurately and efficiently with large-scale, high-resolution GPS trajectory data. The new algorithm separated the GPS trajectory into segments. It found the shortest path for each segment in a scientific manner and ultimately generated a best-matched path for the entire trajectory. The similarity of a trajectory segment and its matched path is described by a similaritymore » score system based on the longest common subsequence. The numerical experiment indicated that the proposed map-matching algorithm was very promising in relation to accuracy and computational efficiency. Large-scale data set applications verified that the proposed method is robust and capable of dealing with real-world, large-scale GPS data in a computationally efficient and accurate manner.« less

  2. Trajectory Segmentation Map-Matching Approach for Large-Scale, High-Resolution GPS Data

    DOE PAGES

    Zhu, Lei; Holden, Jacob R.; Gonder, Jeffrey D.

    2017-01-01

    With the development of smartphones and portable GPS devices, large-scale, high-resolution GPS data can be collected. Map matching is a critical step in studying vehicle driving activity and recognizing network traffic conditions from the data. A new trajectory segmentation map-matching algorithm is proposed to deal accurately and efficiently with large-scale, high-resolution GPS trajectory data. The new algorithm separated the GPS trajectory into segments. It found the shortest path for each segment in a scientific manner and ultimately generated a best-matched path for the entire trajectory. The similarity of a trajectory segment and its matched path is described by a similaritymore » score system based on the longest common subsequence. The numerical experiment indicated that the proposed map-matching algorithm was very promising in relation to accuracy and computational efficiency. Large-scale data set applications verified that the proposed method is robust and capable of dealing with real-world, large-scale GPS data in a computationally efficient and accurate manner.« less

  3. An efficient hole-filling method based on depth map in 3D view generation

    NASA Astrophysics Data System (ADS)

    Liang, Haitao; Su, Xiu; Liu, Yilin; Xu, Huaiyuan; Wang, Yi; Chen, Xiaodong

    2018-01-01

    New virtual view is synthesized through depth image based rendering(DIBR) using a single color image and its associated depth map in 3D view generation. Holes are unavoidably generated in the 2D to 3D conversion process. We propose a hole-filling method based on depth map to address the problem. Firstly, we improve the process of DIBR by proposing a one-to-four (OTF) algorithm. The "z-buffer" algorithm is used to solve overlap problem. Then, based on the classical patch-based algorithm of Criminisi et al., we propose a hole-filling algorithm using the information of depth map to handle the image after DIBR. In order to improve the accuracy of the virtual image, inpainting starts from the background side. In the calculation of the priority, in addition to the confidence term and the data term, we add the depth term. In the search for the most similar patch in the source region, we define the depth similarity to improve the accuracy of searching. Experimental results show that the proposed method can effectively improve the quality of the 3D virtual view subjectively and objectively.

  4. Compression of color-mapped images

    NASA Technical Reports Server (NTRS)

    Hadenfeldt, A. C.; Sayood, Khalid

    1992-01-01

    In a standard image coding scenario, pixel-to-pixel correlation nearly always exists in the data, especially if the image is a natural scene. This correlation is what allows predictive coding schemes (e.g., DPCM) to perform efficient compression. In a color-mapped image, the values stored in the pixel array are no longer directly related to the pixel intensity. Two color indices which are numerically adjacent (close) may point to two very different colors. The correlation still exists, but only via the colormap. This fact can be exploited by sorting the color map to reintroduce the structure. The sorting of colormaps is studied and it is shown how the resulting structure can be used in both lossless and lossy compression of images.

  5. Spatially explicit shallow landslide susceptibility mapping over large areas

    USGS Publications Warehouse

    Bellugi, Dino; Dietrich, William E.; Stock, Jonathan D.; McKean, Jim; Kazian, Brian; Hargrove, Paul

    2011-01-01

    Recent advances in downscaling climate model precipitation predictions now yield spatially explicit patterns of rainfall that could be used to estimate shallow landslide susceptibility over large areas. In California, the United States Geological Survey is exploring community emergency response to the possible effects of a very large simulated storm event and to do so it has generated downscaled precipitation maps for the storm. To predict the corresponding pattern of shallow landslide susceptibility across the state, we have used the model Shalstab (a coupled steady state runoff and infinite slope stability model) which susceptibility spatially explicit estimates of relative potential instability. Such slope stability models that include the effects of subsurface runoff on potentially destabilizing pore pressure evolution require water routing and hence the definition of upslope drainage area to each potential cell. To calculate drainage area efficiently over a large area we developed a parallel framework to scale-up Shalstab and specifically introduce a new efficient parallel drainage area algorithm which produces seamless results. The single seamless shallow landslide susceptibility map for all of California was accomplished in a short run time, and indicates that much larger areas can be efficiently modelled. As landslide maps generally over predict the extent of instability for any given storm. Local empirical data on the fraction of predicted unstable cells that failed for observed rainfall intensity can be used to specify the likely extent of hazard for a given storm. This suggests that campaigns to collect local precipitation data and detailed shallow landslide location maps after major storms could be used to calibrate models and improve their use in hazard assessment for individual storms.

  6. Web Based Rapid Mapping of Disaster Areas using Satellite Images, Web Processing Service, Web Mapping Service, Frequency Based Change Detection Algorithm and J-iView

    NASA Astrophysics Data System (ADS)

    Bandibas, J. C.; Takarada, S.

    2013-12-01

    Timely identification of areas affected by natural disasters is very important for a successful rescue and effective emergency relief efforts. This research focuses on the development of a cost effective and efficient system of identifying areas affected by natural disasters, and the efficient distribution of the information. The developed system is composed of 3 modules which are the Web Processing Service (WPS), Web Map Service (WMS) and the user interface provided by J-iView (fig. 1). WPS is an online system that provides computation, storage and data access services. In this study, the WPS module provides online access of the software implementing the developed frequency based change detection algorithm for the identification of areas affected by natural disasters. It also sends requests to WMS servers to get the remotely sensed data to be used in the computation. WMS is a standard protocol that provides a simple HTTP interface for requesting geo-registered map images from one or more geospatial databases. In this research, the WMS component provides remote access of the satellite images which are used as inputs for land cover change detection. The user interface in this system is provided by J-iView, which is an online mapping system developed at the Geological Survey of Japan (GSJ). The 3 modules are seamlessly integrated into a single package using J-iView, which could rapidly generate a map of disaster areas that is instantaneously viewable online. The developed system was tested using ASTER images covering the areas damaged by the March 11, 2011 tsunami in northeastern Japan. The developed system efficiently generated a map showing areas devastated by the tsunami. Based on the initial results of the study, the developed system proved to be a useful tool for emergency workers to quickly identify areas affected by natural disasters.

  7. Introduction of the trapezoidal thermodynamic technique method for measuring and mapping the efficiency of waste-to-energy plants: A potential replacement to the R1 formula.

    PubMed

    Vakalis, Stergios; Moustakas, Konstantinos; Loizidou, Maria

    2018-06-01

    Waste-to-energy plants have the peculiarity of being considered both as energy production and as waste destruction facilities and this distinction is important for legislative reasons. The efficiency of waste-to-energy plants must be objective and consistent, independently if the focus is the production of energy, the destruction of waste or the recovery/upgrade of materials. With the introduction of polygeneration technologies, like gasification, the production of energy and the recovery/upgrade of materials, are interconnected. The existing methodology for assessing the efficiency of waste-to-energy plants is the R1 formula, which does not take into consideration the full spectrum of the operations that take place in waste-to-energy plants. This study introduces a novel methodology for assessing the efficiency of waste-to-energy plants and is defined as the 3T method, which stands for 'trapezoidal thermodynamic technique'. The 3T method is an integrated approach for assessing the efficiency of waste-to-energy plants, which takes into consideration not only the production of energy but also the quality of the products. The value that is returned from the 3T method can be placed in a tertiary diagram and the global efficiency map of waste-to-energy plants can be produced. The application of the 3T method showed that the waste-to-energy plants with high combined heat and power efficiency and high recovery of materials are favoured and these outcomes are in accordance with the cascade principle and with the high cogeneration standards that are set by the EU Energy Efficiency Directive.

  8. Utilizing Hierarchical Clustering to improve Efficiency of Self-Organizing Feature Map to Identify Hydrological Homogeneous Regions

    NASA Astrophysics Data System (ADS)

    Farsadnia, Farhad; Ghahreman, Bijan

    2016-04-01

    Hydrologic homogeneous group identification is considered both fundamental and applied research in hydrology. Clustering methods are among conventional methods to assess the hydrological homogeneous regions. Recently, Self-Organizing feature Map (SOM) method has been applied in some studies. However, the main problem of this method is the interpretation on the output map of this approach. Therefore, SOM is used as input to other clustering algorithms. The aim of this study is to apply a two-level Self-Organizing feature map and Ward hierarchical clustering method to determine the hydrologic homogenous regions in North and Razavi Khorasan provinces. At first by principal component analysis, we reduced SOM input matrix dimension, then the SOM was used to form a two-dimensional features map. To determine homogeneous regions for flood frequency analysis, SOM output nodes were used as input into the Ward method. Generally, the regions identified by the clustering algorithms are not statistically homogeneous. Consequently, they have to be adjusted to improve their homogeneity. After adjustment of the homogeneity regions by L-moment tests, five hydrologic homogeneous regions were identified. Finally, adjusted regions were created by a two-level SOM and then the best regional distribution function and associated parameters were selected by the L-moment approach. The results showed that the combination of self-organizing maps and Ward hierarchical clustering by principal components as input is more effective than the hierarchical method, by principal components or standardized inputs to achieve hydrologic homogeneous regions.

  9. A Tailored Systems Engineering Framework for Science and Technology Projects

    DTIC Science & Technology

    2009-03-01

    PROJECTS THESIS Stephen M. Behm Major, USAF J . Bradford Pitzer Major, USAF Jane F. White Civilian, US AFIT/GSE/ENV/09-M02 DEPARTMENT OF THE...Engineering Stephen M. Behm, Major, USAF J . Bradford Pitzer, Major, USAF Jane F. White, Civilian, US March 2009 APPROVED FOR PUBLIC...Stephen M. Behm, Major, USAF J . Bradford Pitzer, Major, USAF Jane F. White, Civilian, US Approved: iv AFIT/GSE/ENV

  10. Hierarchical tone mapping for high dynamic range image visualization

    NASA Astrophysics Data System (ADS)

    Qiu, Guoping; Duan, Jiang

    2005-07-01

    In this paper, we present a computationally efficient, practically easy to use tone mapping techniques for the visualization of high dynamic range (HDR) images in low dynamic range (LDR) reproduction devices. The new method, termed hierarchical nonlinear linear (HNL) tone-mapping operator maps the pixels in two hierarchical steps. The first step allocates appropriate numbers of LDR display levels to different HDR intensity intervals according to the pixel densities of the intervals. The second step linearly maps the HDR intensity intervals to theirs allocated LDR display levels. In the developed HNL scheme, the assignment of LDR display levels to HDR intensity intervals is controlled by a very simple and flexible formula with a single adjustable parameter. We also show that our new operators can be used for the effective enhancement of ordinary images.

  11. Entanglement verification with detection efficiency mismatch

    NASA Astrophysics Data System (ADS)

    Zhang, Yanbao; Lütkenhaus, Norbert

    Entanglement is a necessary condition for secure quantum key distribution (QKD). When there is an efficiency mismatch between various detectors used in the QKD system, it is still an open problem how to verify entanglement. Here we present a method to address this problem, given that the detection efficiency mismatch is characterized and known. The method works without assuming an upper bound on the number of photons going to each threshold detector. Our results suggest that the efficiency mismatch affects the ability to verify entanglement: the larger the efficiency mismatch is, the smaller the set of entangled states that can be verified becomes. When there is no mismatch, our method can verify entanglement even if the method based on squashing maps [PRL 101, 093601 (2008)] fails.

  12. Robust, Efficient Depth Reconstruction With Hierarchical Confidence-Based Matching.

    PubMed

    Sun, Li; Chen, Ke; Song, Mingli; Tao, Dacheng; Chen, Gang; Chen, Chun

    2017-07-01

    In recent years, taking photos and capturing videos with mobile devices have become increasingly popular. Emerging applications based on the depth reconstruction technique have been developed, such as Google lens blur. However, depth reconstruction is difficult due to occlusions, non-diffuse surfaces, repetitive patterns, and textureless surfaces, and it has become more difficult due to the unstable image quality and uncontrolled scene condition in the mobile setting. In this paper, we present a novel hierarchical framework with multi-view confidence-based matching for robust, efficient depth reconstruction in uncontrolled scenes. Particularly, the proposed framework combines local cost aggregation with global cost optimization in a complementary manner that increases efficiency and accuracy. A depth map is efficiently obtained in a coarse-to-fine manner by using an image pyramid. Moreover, confidence maps are computed to robustly fuse multi-view matching cues, and to constrain the stereo matching on a finer scale. The proposed framework has been evaluated with challenging indoor and outdoor scenes, and has achieved robust and efficient depth reconstruction.

  13. AlignerBoost: A Generalized Software Toolkit for Boosting Next-Gen Sequencing Mapping Accuracy Using a Bayesian-Based Mapping Quality Framework.

    PubMed

    Zheng, Qi; Grice, Elizabeth A

    2016-10-01

    Accurate mapping of next-generation sequencing (NGS) reads to reference genomes is crucial for almost all NGS applications and downstream analyses. Various repetitive elements in human and other higher eukaryotic genomes contribute in large part to ambiguously (non-uniquely) mapped reads. Most available NGS aligners attempt to address this by either removing all non-uniquely mapping reads, or reporting one random or "best" hit based on simple heuristics. Accurate estimation of the mapping quality of NGS reads is therefore critical albeit completely lacking at present. Here we developed a generalized software toolkit "AlignerBoost", which utilizes a Bayesian-based framework to accurately estimate mapping quality of ambiguously mapped NGS reads. We tested AlignerBoost with both simulated and real DNA-seq and RNA-seq datasets at various thresholds. In most cases, but especially for reads falling within repetitive regions, AlignerBoost dramatically increases the mapping precision of modern NGS aligners without significantly compromising the sensitivity even without mapping quality filters. When using higher mapping quality cutoffs, AlignerBoost achieves a much lower false mapping rate while exhibiting comparable or higher sensitivity compared to the aligner default modes, therefore significantly boosting the detection power of NGS aligners even using extreme thresholds. AlignerBoost is also SNP-aware, and higher quality alignments can be achieved if provided with known SNPs. AlignerBoost's algorithm is computationally efficient, and can process one million alignments within 30 seconds on a typical desktop computer. AlignerBoost is implemented as a uniform Java application and is freely available at https://github.com/Grice-Lab/AlignerBoost.

  14. AlignerBoost: A Generalized Software Toolkit for Boosting Next-Gen Sequencing Mapping Accuracy Using a Bayesian-Based Mapping Quality Framework

    PubMed Central

    Zheng, Qi; Grice, Elizabeth A.

    2016-01-01

    Accurate mapping of next-generation sequencing (NGS) reads to reference genomes is crucial for almost all NGS applications and downstream analyses. Various repetitive elements in human and other higher eukaryotic genomes contribute in large part to ambiguously (non-uniquely) mapped reads. Most available NGS aligners attempt to address this by either removing all non-uniquely mapping reads, or reporting one random or "best" hit based on simple heuristics. Accurate estimation of the mapping quality of NGS reads is therefore critical albeit completely lacking at present. Here we developed a generalized software toolkit "AlignerBoost", which utilizes a Bayesian-based framework to accurately estimate mapping quality of ambiguously mapped NGS reads. We tested AlignerBoost with both simulated and real DNA-seq and RNA-seq datasets at various thresholds. In most cases, but especially for reads falling within repetitive regions, AlignerBoost dramatically increases the mapping precision of modern NGS aligners without significantly compromising the sensitivity even without mapping quality filters. When using higher mapping quality cutoffs, AlignerBoost achieves a much lower false mapping rate while exhibiting comparable or higher sensitivity compared to the aligner default modes, therefore significantly boosting the detection power of NGS aligners even using extreme thresholds. AlignerBoost is also SNP-aware, and higher quality alignments can be achieved if provided with known SNPs. AlignerBoost’s algorithm is computationally efficient, and can process one million alignments within 30 seconds on a typical desktop computer. AlignerBoost is implemented as a uniform Java application and is freely available at https://github.com/Grice-Lab/AlignerBoost. PMID:27706155

  15. Fast and robust generation of feature maps for region-based visual attention.

    PubMed

    Aziz, Muhammad Zaheer; Mertsching, Bärbel

    2008-05-01

    Visual attention is one of the important phenomena in biological vision which can be followed to achieve more efficiency, intelligence, and robustness in artificial vision systems. This paper investigates a region-based approach that performs pixel clustering prior to the processes of attention in contrast to late clustering as done by contemporary methods. The foundation steps of feature map construction for the region-based attention model are proposed here. The color contrast map is generated based upon the extended findings from the color theory, the symmetry map is constructed using a novel scanning-based method, and a new algorithm is proposed to compute a size contrast map as a formal feature channel. Eccentricity and orientation are computed using the moments of obtained regions and then saliency is evaluated using the rarity criteria. The efficient design of the proposed algorithms allows incorporating five feature channels while maintaining a processing rate of multiple frames per second. Another salient advantage over the existing techniques is the reusability of the salient regions in the high-level machine vision procedures due to preservation of their shapes and precise locations. The results indicate that the proposed model has the potential to efficiently integrate the phenomenon of attention into the main stream of machine vision and systems with restricted computing resources such as mobile robots can benefit from its advantages.

  16. Real-Time Large-Scale Dense Mapping with Surfels

    PubMed Central

    Fu, Xingyin; Zhu, Feng; Wu, Qingxiao; Sun, Yunlei; Lu, Rongrong; Yang, Ruigang

    2018-01-01

    Real-time dense mapping systems have been developed since the birth of consumer RGB-D cameras. Currently, there are two commonly used models in dense mapping systems: truncated signed distance function (TSDF) and surfel. The state-of-the-art dense mapping systems usually work fine with small-sized regions. The generated dense surface may be unsatisfactory around the loop closures when the system tracking drift grows large. In addition, the efficiency of the system with surfel model slows down when the number of the model points in the map becomes large. In this paper, we propose to use two maps in the dense mapping system. The RGB-D images are integrated into a local surfel map. The old surfels that reconstructed in former times and far away from the camera frustum are moved from the local map to the global map. The updated surfels in the local map when every frame arrives are kept bounded. Therefore, in our system, the scene that can be reconstructed is very large, and the frame rate of our system remains high. We detect loop closures and optimize the pose graph to distribute system tracking drift. The positions and normals of the surfels in the map are also corrected using an embedded deformation graph so that they are consistent with the updated poses. In order to deal with large surface deformations, we propose a new method for constructing constraints with system trajectories and loop closure keyframes. The proposed new method stabilizes large-scale surface deformation. Experimental results show that our novel system behaves better than the prior state-of-the-art dense mapping systems. PMID:29747450

  17. Empty tracks optimization based on Z-Map model

    NASA Astrophysics Data System (ADS)

    Liu, Le; Yan, Guangrong; Wang, Zaijun; Zang, Genao

    2017-12-01

    For parts with many features, there are more empty tracks during machining. If these tracks are not optimized, the machining efficiency will be seriously affected. In this paper, the characteristics of the empty tracks are studied in detail. Combining with the existing optimization algorithm, a new tracks optimization method based on Z-Map model is proposed. In this method, the tool tracks are divided into the unit processing section, and then the Z-Map model simulation technique is used to analyze the order constraint between the unit segments. The empty stroke optimization problem is transformed into the TSP with sequential constraints, and then through the genetic algorithm solves the established TSP problem. This kind of optimization method can not only optimize the simple structural parts, but also optimize the complex structural parts, so as to effectively plan the empty tracks and greatly improve the processing efficiency.

  18. Evaluation of color mapping algorithms in different color spaces

    NASA Astrophysics Data System (ADS)

    Bronner, Timothée.-Florian; Boitard, Ronan; Pourazad, Mahsa T.; Nasiopoulos, Panos; Ebrahimi, Touradj

    2016-09-01

    The color gamut supported by current commercial displays is only a subset of the full spectrum of colors visible by the human eye. In High-Definition (HD) television technology, the scope of the supported colors covers 35.9% of the full visible gamut. For comparison, Ultra High-Definition (UHD) television, which is currently being deployed on the market, extends this range to 75.8%. However, when reproducing content with a wider color gamut than that of a television, typically UHD content on HD television, some original color information may lie outside the reproduction capabilities of the television. Efficient gamut mapping techniques are required in order to fit the colors of any source content into the gamut of a given display. The goal of gamut mapping is to minimize the distortion, in terms of perceptual quality, when converting video from one color gamut to another. It is assumed that the efficiency of gamut mapping depends on the color space in which it is computed. In this article, we evaluate 14 gamut mapping techniques, 12 combinations of two projection methods across six color spaces as well as R'G'B' Clipping and wrong gamut interpretation. Objective results, using the CIEDE2000 metric, show that the R'G'B' Clipping is slightly outperformed by only one combination of color space and projection method. However, analysis of images shows that R'G'B' Clipping can result in loss of contrast in highly saturated images, greatly impairing the quality of the mapped image.

  19. Advancing precision cosmology with 21 cm intensity mapping

    NASA Astrophysics Data System (ADS)

    Masui, Kiyoshi Wesley

    In this thesis we make progress toward establishing the observational method of 21 cm intensity mapping as a sensitive and efficient method for mapping the large-scale structure of the Universe. In Part I we undertake theoretical studies to better understand the potential of intensity mapping. This includes forecasting the ability of intensity mapping experiments to constrain alternative explanations to dark energy for the Universe's accelerated expansion. We also considered how 21 cm observations of the neutral gas in the early Universe (after recombination but before reionization) could be used to detect primordial gravity waves, thus providing a window into cosmological inflation. Finally we showed that scientifically interesting measurements could in principle be performed using intensity mapping in the near term, using existing telescopes in pilot surveys or prototypes for larger dedicated surveys. Part II describes observational efforts to perform some of the first measurements using 21 cm intensity mapping. We develop a general data analysis pipeline for analyzing intensity mapping data from single dish radio telescopes. We then apply the pipeline to observations using the Green Bank Telescope. By cross-correlating the intensity mapping survey with a traditional galaxy redshift survey we put a lower bound on the amplitude of the 21 cm signal. The auto-correlation provides an upper bound on the signal amplitude and we thus constrain the signal from both above and below. This pilot survey represents a pioneering effort in establishing 21 cm intensity mapping as a probe of the Universe.

  20. GIS-based realization of international standards for digital geological mapping - developments in planetary mapping

    NASA Astrophysics Data System (ADS)

    Nass, Andrea; van Gasselt, Stephan; Jaumann, Ralf

    2010-05-01

    data model in the next step. As second task, we designed a concept for description of the digital mapping result. Therefore, we are creating a metadata template based on existing standards for individual needs in planetary sciences. This template is subdivided in (meta) data about the general map content (e.g. on which data the mapping result based on) and in metadata for each individual mapping element/layer comprising information like minimum mapping scale, interpretation hints, etc. The assignment of such a metadata description in combination with the usage of a predefined mapping schema facilitates the efficient and traceable storage of data information on a network server and enables a subsequent representation, e.g. as a mapserver data structure. Acknowledgement: This work is partly supported by DLR and the Helmholtz Alliance "Planetary Evolution and Life".

  1. Projection Mapping User Interface for Disabled People

    PubMed Central

    Simutis, Rimvydas; Maskeliūnas, Rytis

    2018-01-01

    Difficulty in communicating is one of the key challenges for people suffering from severe motor and speech disabilities. Often such person can communicate and interact with the environment only using assistive technologies. This paper presents a multifunctional user interface designed to improve communication efficiency and person independence. The main component of this interface is a projection mapping technique used to highlight objects in the environment. Projection mapping makes it possible to create a natural augmented reality information presentation method. The user interface combines a depth sensor and a projector to create camera-projector system. We provide a detailed description of camera-projector system calibration procedure. The described system performs tabletop object detection and automatic projection mapping. Multiple user input modalities have been integrated into the multifunctional user interface. Such system can be adapted to the needs of people with various disabilities. PMID:29686827

  2. Projection Mapping User Interface for Disabled People.

    PubMed

    Gelšvartas, Julius; Simutis, Rimvydas; Maskeliūnas, Rytis

    2018-01-01

    Difficulty in communicating is one of the key challenges for people suffering from severe motor and speech disabilities. Often such person can communicate and interact with the environment only using assistive technologies. This paper presents a multifunctional user interface designed to improve communication efficiency and person independence. The main component of this interface is a projection mapping technique used to highlight objects in the environment. Projection mapping makes it possible to create a natural augmented reality information presentation method. The user interface combines a depth sensor and a projector to create camera-projector system. We provide a detailed description of camera-projector system calibration procedure. The described system performs tabletop object detection and automatic projection mapping. Multiple user input modalities have been integrated into the multifunctional user interface. Such system can be adapted to the needs of people with various disabilities.

  3. An authenticated image encryption scheme based on chaotic maps and memory cellular automata

    NASA Astrophysics Data System (ADS)

    Bakhshandeh, Atieh; Eslami, Ziba

    2013-06-01

    This paper introduces a new image encryption scheme based on chaotic maps, cellular automata and permutation-diffusion architecture. In the permutation phase, a piecewise linear chaotic map is utilized to confuse the plain-image and in the diffusion phase, we employ the Logistic map as well as a reversible memory cellular automata to obtain an efficient and secure cryptosystem. The proposed method admits advantages such as highly secure diffusion mechanism, computational efficiency and ease of implementation. A novel property of the proposed scheme is its authentication ability which can detect whether the image is tampered during the transmission or not. This is particularly important in applications where image data or part of it contains highly sensitive information. Results of various analyses manifest high security of this new method and its capability for practical image encryption.

  4. FASTdoop: a versatile and efficient library for the input of FASTA and FASTQ files for MapReduce Hadoop bioinformatics applications.

    PubMed

    Ferraro Petrillo, Umberto; Roscigno, Gianluca; Cattaneo, Giuseppe; Giancarlo, Raffaele

    2017-05-15

    MapReduce Hadoop bioinformatics applications require the availability of special-purpose routines to manage the input of sequence files. Unfortunately, the Hadoop framework does not provide any built-in support for the most popular sequence file formats like FASTA or BAM. Moreover, the development of these routines is not easy, both because of the diversity of these formats and the need for managing efficiently sequence datasets that may count up to billions of characters. We present FASTdoop, a generic Hadoop library for the management of FASTA and FASTQ files. We show that, with respect to analogous input management routines that have appeared in the Literature, it offers versatility and efficiency. That is, it can handle collections of reads, with or without quality scores, as well as long genomic sequences while the existing routines concentrate mainly on NGS sequence data. Moreover, in the domain where a comparison is possible, the routines proposed here are faster than the available ones. In conclusion, FASTdoop is a much needed addition to Hadoop-BAM. The software and the datasets are available at http://www.di.unisa.it/FASTdoop/ . umberto.ferraro@uniroma1.it. Supplementary data are available at Bioinformatics online. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  5. Geologic map of Chickasaw National Recreation Area, Murray County, Oklahoma

    USGS Publications Warehouse

    Blome, Charles D.; Lidke, David J.; Wahl, Ronald R.; Golab, James A.

    2013-01-01

    This 1:24,000-scale geologic map is a compilation of previous geologic maps and new geologic mapping of areas in and around Chickasaw National Recreation Area. The geologic map includes revisions of numerous unit contacts and faults and a number of previously “undifferentiated” rock units were subdivided in some areas. Numerous circular-shaped hills in and around Chickasaw National Recreation Area are probably the result of karst-related collapse and may represent the erosional remnants of large, exhumed sinkholes. Geospatial registration of existing, smaller scale (1:72,000- and 1:100,000-scale) geologic maps of the area and construction of an accurate Geographic Information System (GIS) database preceded 2 years of fieldwork wherein previously mapped geology (unit contacts and faults) was verified and new geologic mapping was carried out. The geologic map of Chickasaw National Recreation Area and this pamphlet include information pertaining to how the geologic units and structural features in the map area relate to the formation of the northern Arbuckle Mountains and its Arbuckle-Simpson aquifer. The development of an accurate geospatial GIS database and the use of a handheld computer in the field greatly increased both the accuracy and efficiency in producing the 1:24,000-scale geologic map.

  6. A high-resolution physically-based global flood hazard map

    NASA Astrophysics Data System (ADS)

    Kaheil, Y.; Begnudelli, L.; McCollum, J.

    2016-12-01

    We present the results from a physically-based global flood hazard model. The model uses a physically-based hydrologic model to simulate river discharges, and 2D hydrodynamic model to simulate inundation. The model is set up such that it allows the application of large-scale flood hazard through efficient use of parallel computing. For hydrology, we use the Hillslope River Routing (HRR) model. HRR accounts for surface hydrology using Green-Ampt parameterization. The model is calibrated against observed discharge data from the Global Runoff Data Centre (GRDC) network, among other publicly-available datasets. The parallel-computing framework takes advantage of the river network structure to minimize cross-processor messages, and thus significantly increases computational efficiency. For inundation, we implemented a computationally-efficient 2D finite-volume model with wetting/drying. The approach consists of simulating flood along the river network by forcing the hydraulic model with the streamflow hydrographs simulated by HRR, and scaled up to certain return levels, e.g. 100 years. The model is distributed such that each available processor takes the next simulation. Given an approximate criterion, the simulations are ordered from most-demanding to least-demanding to ensure that all processors finalize almost simultaneously. Upon completing all simulations, the maximum envelope of flood depth is taken to generate the final map. The model is applied globally, with selected results shown from different continents and regions. The maps shown depict flood depth and extent at different return periods. These maps, which are currently available at 3 arc-sec resolution ( 90m) can be made available at higher resolutions where high resolution DEMs are available. The maps can be utilized by flood risk managers at the national, regional, and even local levels to further understand their flood risk exposure, exercise certain measures of mitigation, and/or transfer the residual

  7. Jane's Law

    THOMAS, 113th Congress

    Rep. Cohen, Steve [D-TN-9

    2013-02-05

    House - 02/28/2013 Referred to the Subcommittee on Crime, Terrorism, Homeland Security, And Investigations. (All Actions) Tracker: This bill has the status IntroducedHere are the steps for Status of Legislation:

  8. LPmerge: an R package for merging genetic maps by linear programming.

    PubMed

    Endelman, Jeffrey B; Plomion, Christophe

    2014-06-01

    Consensus genetic maps constructed from multiple populations are an important resource for both basic and applied research, including genome-wide association analysis, genome sequence assembly and studies of evolution. The LPmerge software uses linear programming to efficiently minimize the mean absolute error between the consensus map and the linkage maps from each population. This minimization is performed subject to linear inequality constraints that ensure the ordering of the markers in the linkage maps is preserved. When marker order is inconsistent between linkage maps, a minimum set of ordinal constraints is deleted to resolve the conflicts. LPmerge is on CRAN at http://cran.r-project.org/web/packages/LPmerge. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  9. Multiplexed single-mode wavelength-to-time mapping of multimode light

    PubMed Central

    Chandrasekharan, Harikumar K; Izdebski, Frauke; Gris-Sánchez, Itandehui; Krstajić, Nikola; Walker, Richard; Bridle, Helen L.; Dalgarno, Paul A.; MacPherson, William N.; Henderson, Robert K.; Birks, Tim A.; Thomson, Robert R.

    2017-01-01

    When an optical pulse propagates along an optical fibre, different wavelengths travel at different group velocities. As a result, wavelength information is converted into arrival-time information, a process known as wavelength-to-time mapping. This phenomenon is most cleanly observed using a single-mode fibre transmission line, where spatial mode dispersion is not present, but the use of such fibres restricts possible applications. Here we demonstrate that photonic lanterns based on tapered single-mode multicore fibres provide an efficient way to couple multimode light to an array of single-photon avalanche detectors, each of which has its own time-to-digital converter for time-correlated single-photon counting. Exploiting this capability, we demonstrate the multiplexed single-mode wavelength-to-time mapping of multimode light using a multicore fibre photonic lantern with 121 single-mode cores, coupled to 121 detectors on a 32 × 32 detector array. This work paves the way to efficient multimode wavelength-to-time mapping systems with the spectral performance of single-mode systems. PMID:28120822

  10. Semiautomatic mapping of permafrost in the Yukon Flats, Alaska

    NASA Astrophysics Data System (ADS)

    Gulbrandsen, Mats Lundh; Minsley, Burke J.; Ball, Lyndsay B.; Hansen, Thomas Mejer

    2016-12-01

    Thawing of permafrost due to global warming can have major impacts on hydrogeological processes, climate feedback, arctic ecology, and local environments. To understand these effects and processes, it is crucial to know the distribution of permafrost. In this study we exploit the fact that airborne electromagnetic (AEM) data are sensitive to the distribution of permafrost and demonstrate how the distribution of permafrost in the Yukon Flats, Alaska, is mapped in an efficient (semiautomatic) way, using a combination of supervised and unsupervised (machine) learning algorithms, i.e., Smart Interpretation and K-means clustering. Clustering is used to sort unfrozen and frozen regions, and Smart Interpretation is used to predict the depth of permafrost based on expert interpretations. This workflow allows, for the first time, a quantitative and objective approach to efficiently map permafrost based on large amounts of AEM data.

  11. Semiautomatic mapping of permafrost in the Yukon Flats, Alaska

    USGS Publications Warehouse

    Gulbrandsen, Mats Lundh; Minsley, Burke J.; Ball, Lyndsay B.; Hansen, Thomas Mejer

    2016-01-01

    Thawing of permafrost due to global warming can have major impacts on hydrogeological processes, climate feedback, arctic ecology, and local environments. To understand these effects and processes, it is crucial to know the distribution of permafrost. In this study we exploit the fact that airborne electromagnetic (AEM) data are sensitive to the distribution of permafrost and demonstrate how the distribution of permafrost in the Yukon Flats, Alaska, is mapped in an efficient (semiautomatic) way, using a combination of supervised and unsupervised (machine) learning algorithms, i.e., Smart Interpretation and K-means clustering. Clustering is used to sort unfrozen and frozen regions, and Smart Interpretation is used to predict the depth of permafrost based on expert interpretations. This workflow allows, for the first time, a quantitative and objective approach to efficiently map permafrost based on large amounts of AEM data.

  12. Google Maps offers a new way to evaluate claudication.

    PubMed

    Khambati, Husain; Boles, Kim; Jetty, Prasad

    2017-05-01

    Accurate determination of walking capacity is important for the clinical diagnosis and management plan for patients with peripheral arterial disease. The current "gold standard" of measurement is walking distance on a treadmill. However, treadmill testing is not always reflective of the patient's natural walking conditions, and it may not be fully accessible in every vascular clinic. The objective of this study was to determine whether Google Maps, the readily available GPS-based mapping tool, offers an accurate and accessible method of evaluating walking distances in vascular claudication patients. Patients presenting to the outpatient vascular surgery clinic between November 2013 and April 2014 at the Ottawa Hospital with vasculogenic calf, buttock, and thigh claudication symptoms were identified and prospectively enrolled in our study. Onset of claudication symptoms and maximal walking distance (MWD) were evaluated using four tools: history; Walking Impairment Questionnaire (WIQ), a validated claudication survey; Google Maps distance calculator (patients were asked to report their daily walking routes on the Google Maps-based tool runningmap.com, and walking distances were calculated accordingly); and treadmill testing for onset of symptoms and MWD, recorded in a double-blinded fashion. Fifteen patients were recruited for the study. Determination of walking distances using Google Maps proved to be more accurate than by both clinical history and WIQ, correlating highly with the gold standard of treadmill testing for both claudication onset (r = .805; P < .001) and MWD (r = .928; P < .0001). In addition, distances were generally under-reported on history and WIQ. The Google Maps tool was also efficient, with reporting times averaging below 4 minutes. For vascular claudicants with no other walking limitations, Google Maps is a promising new tool that combines the objective strengths of the treadmill test and incorporates real-world walking environments. It

  13. A fruit quality gene map of Prunus

    PubMed Central

    2009-01-01

    Background Prunus fruit development, growth, ripening, and senescence includes major biochemical and sensory changes in texture, color, and flavor. The genetic dissection of these complex processes has important applications in crop improvement, to facilitate maximizing and maintaining stone fruit quality from production and processing through to marketing and consumption. Here we present an integrated fruit quality gene map of Prunus containing 133 genes putatively involved in the determination of fruit texture, pigmentation, flavor, and chilling injury resistance. Results A genetic linkage map of 211 markers was constructed for an intraspecific peach (Prunus persica) progeny population, Pop-DG, derived from a canning peach cultivar 'Dr. Davis' and a fresh market cultivar 'Georgia Belle'. The Pop-DG map covered 818 cM of the peach genome and included three morphological markers, 11 ripening candidate genes, 13 cold-responsive genes, 21 novel EST-SSRs from the ChillPeach database, 58 previously reported SSRs, 40 RAFs, 23 SRAPs, 14 IMAs, and 28 accessory markers from candidate gene amplification. The Pop-DG map was co-linear with the Prunus reference T × E map, with 39 SSR markers in common to align the maps. A further 158 markers were bin-mapped to the reference map: 59 ripening candidate genes, 50 cold-responsive genes, and 50 novel EST-SSRs from ChillPeach, with deduced locations in Pop-DG via comparative mapping. Several candidate genes and EST-SSRs co-located with previously reported major trait loci and quantitative trait loci for chilling injury symptoms in Pop-DG. Conclusion The candidate gene approach combined with bin-mapping and availability of a community-recognized reference genetic map provides an efficient means of locating genes of interest in a target genome. We highlight the co-localization of fruit quality candidate genes with previously reported fruit quality QTLs. The fruit quality gene map developed here is a valuable tool for dissecting the

  14. Semi-automatic mapping of geological Structures using UAV-based photogrammetric data: An image analysis approach

    NASA Astrophysics Data System (ADS)

    Vasuki, Yathunanthan; Holden, Eun-Jung; Kovesi, Peter; Micklethwaite, Steven

    2014-08-01

    Recent advances in data acquisition technologies, such as Unmanned Aerial Vehicles (UAVs), have led to a growing interest in capturing high-resolution rock surface images. However, due to the large volumes of data that can be captured in a short flight, efficient analysis of this data brings new challenges, especially the time it takes to digitise maps and extract orientation data. We outline a semi-automated method that allows efficient mapping of geological faults using photogrammetric data of rock surfaces, which was generated from aerial photographs collected by a UAV. Our method harnesses advanced automated image analysis techniques and human data interaction to rapidly map structures and then calculate their dip and dip directions. Geological structures (faults, joints and fractures) are first detected from the primary photographic dataset and the equivalent three dimensional (3D) structures are then identified within a 3D surface model generated by structure from motion (SfM). From this information the location, dip and dip direction of the geological structures are calculated. A structure map generated by our semi-automated method obtained a recall rate of 79.8% when compared against a fault map produced using expert manual digitising and interpretation methods. The semi-automated structure map was produced in 10 min whereas the manual method took approximately 7 h. In addition, the dip and dip direction calculation, using our automated method, shows a mean±standard error of 1.9°±2.2° and 4.4°±2.6° respectively with field measurements. This shows the potential of using our semi-automated method for accurate and efficient mapping of geological structures, particularly from remote, inaccessible or hazardous sites.

  15. Glimpse: Sparsity based weak lensing mass-mapping tool

    NASA Astrophysics Data System (ADS)

    Lanusse, F.; Starck, J.-L.; Leonard, A.; Pires, S.

    2018-02-01

    Glimpse, also known as Glimpse2D, is a weak lensing mass-mapping tool that relies on a robust sparsity-based regularization scheme to recover high resolution convergence from either gravitational shear alone or from a combination of shear and flexion. Including flexion allows the supplementation of the shear on small scales in order to increase the sensitivity to substructures and the overall resolution of the convergence map. To preserve all available small scale information, Glimpse avoids any binning of the irregularly sampled input shear and flexion fields and treats the mass-mapping problem as a general ill-posed inverse problem, regularized using a multi-scale wavelet sparsity prior. The resulting algorithm incorporates redshift, reduced shear, and reduced flexion measurements for individual galaxies and is made highly efficient by the use of fast Fourier estimators.

  16. An Efficient Method of Sharing Mass Spatio-Temporal Trajectory Data Based on Cloudera Impala for Traffic Distribution Mapping in an Urban City

    PubMed Central

    Zhou, Lianjie; Chen, Nengcheng; Yuan, Sai; Chen, Zeqiang

    2016-01-01

    The efficient sharing of spatio-temporal trajectory data is important to understand traffic congestion in mass data. However, the data volumes of bus networks in urban cities are growing rapidly, reaching daily volumes of one hundred million datapoints. Accessing and retrieving mass spatio-temporal trajectory data in any field is hard and inefficient due to limited computational capabilities and incomplete data organization mechanisms. Therefore, we propose an optimized and efficient spatio-temporal trajectory data retrieval method based on the Cloudera Impala query engine, called ESTRI, to enhance the efficiency of mass data sharing. As an excellent query tool for mass data, Impala can be applied for mass spatio-temporal trajectory data sharing. In ESTRI we extend the spatio-temporal trajectory data retrieval function of Impala and design a suitable data partitioning method. In our experiments, the Taiyuan BeiDou (BD) bus network is selected, containing 2300 buses with BD positioning sensors, producing 20 million records every day, resulting in two difficulties as described in the Introduction section. In addition, ESTRI and MongoDB are applied in experiments. The experiments show that ESTRI achieves the most efficient data retrieval compared to retrieval using MongoDB for data volumes of fifty million, one hundred million, one hundred and fifty million, and two hundred million. The performance of ESTRI is approximately seven times higher than that of MongoDB. The experiments show that ESTRI is an effective method for retrieving mass spatio-temporal trajectory data. Finally, bus distribution mapping in Taiyuan city is achieved, describing the buses density in different regions at different times throughout the day, which can be applied in future studies of transport, such as traffic scheduling, traffic planning and traffic behavior management in intelligent public transportation systems. PMID:27801869

  17. An Efficient Method of Sharing Mass Spatio-Temporal Trajectory Data Based on Cloudera Impala for Traffic Distribution Mapping in an Urban City.

    PubMed

    Zhou, Lianjie; Chen, Nengcheng; Yuan, Sai; Chen, Zeqiang

    2016-10-29

    The efficient sharing of spatio-temporal trajectory data is important to understand traffic congestion in mass data. However, the data volumes of bus networks in urban cities are growing rapidly, reaching daily volumes of one hundred million datapoints. Accessing and retrieving mass spatio-temporal trajectory data in any field is hard and inefficient due to limited computational capabilities and incomplete data organization mechanisms. Therefore, we propose an optimized and efficient spatio-temporal trajectory data retrieval method based on the Cloudera Impala query engine, called ESTRI, to enhance the efficiency of mass data sharing. As an excellent query tool for mass data, Impala can be applied for mass spatio-temporal trajectory data sharing. In ESTRI we extend the spatio-temporal trajectory data retrieval function of Impala and design a suitable data partitioning method. In our experiments, the Taiyuan BeiDou (BD) bus network is selected, containing 2300 buses with BD positioning sensors, producing 20 million records every day, resulting in two difficulties as described in the Introduction section. In addition, ESTRI and MongoDB are applied in experiments. The experiments show that ESTRI achieves the most efficient data retrieval compared to retrieval using MongoDB for data volumes of fifty million, one hundred million, one hundred and fifty million, and two hundred million. The performance of ESTRI is approximately seven times higher than that of MongoDB. The experiments show that ESTRI is an effective method for retrieving mass spatio-temporal trajectory data. Finally, bus distribution mapping in Taiyuan city is achieved, describing the buses density in different regions at different times throughout the day, which can be applied in future studies of transport, such as traffic scheduling, traffic planning and traffic behavior management in intelligent public transportation systems.

  18. A Neurocomputational Account of Taxonomic Responding and Fast Mapping in Early Word Learning

    ERIC Educational Resources Information Center

    Mayor, Julien; Plunkett, Kim

    2010-01-01

    We present a neurocomputational model with self-organizing maps that accounts for the emergence of taxonomic responding and fast mapping in early word learning, as well as a rapid increase in the rate of acquisition of words observed in late infancy. The quality and efficiency of generalization of word-object associations is directly related to…

  19. Digital Mapping Techniques '09-Workshop Proceedings, Morgantown, West Virginia, May 10-13, 2009

    USGS Publications Warehouse

    Soller, David R.

    2011-01-01

    As in the previous years' meetings, the objective was to foster informal discussion and exchange of technical information, principally in order to develop more efficient methods for digital mapping, cartography, GIS analysis, and information management. At this meeting, oral and poster presentations and special discussion sessions emphasized (1) methods for creating and publishing map products (here, "publishing" includes Web-based release); (2) field data capture software and techniques, including the use of LiDAR; (3) digital cartographic techniques; (4) migration of digital maps into ArcGIS Geodatabase format; (5) analytical GIS techniques; and (6) continued development of the National Geologic Map Database.

  20. Sensor-Motor Maps for Describing Linear Reflex Composition in Hopping.

    PubMed

    Schumacher, Christian; Seyfarth, André

    2017-01-01

    In human and animal motor control several sensory organs contribute to a network of sensory pathways modulating the motion depending on the task and the phase of execution to generate daily motor tasks such as locomotion. To better understand the individual and joint contribution of reflex pathways in locomotor tasks, we developed a neuromuscular model that describes hopping movements. In this model, we consider the influence of proprioceptive length (LFB), velocity (VFB) and force feedback (FFB) pathways of a leg extensor muscle on hopping stability, performance and efficiency (metabolic effort). Therefore, we explore the space describing the blending of the monosynaptic reflex pathway gains. We call this reflex parameter space a sensor-motor map . The sensor-motor maps are used to visualize the functional contribution of sensory pathways in multisensory integration. We further evaluate the robustness of these sensor-motor maps to changes in tendon elasticity, body mass, segment length and ground compliance. The model predicted that different reflex pathway compositions selectively optimize specific hopping characteristics (e.g., performance and efficiency). Both FFB and LFB were pathways that enable hopping. FFB resulted in the largest hopping heights, LFB enhanced hopping efficiency and VFB had the ability to disable hopping. For the tested case, the topology of the sensor-motor maps as well as the location of functionally optimal compositions were invariant to changes in system designs (tendon elasticity, body mass, segment length) or environmental parameters (ground compliance). Our results indicate that different feedback pathway compositions may serve different functional roles. The topology of the sensor-motor map was predicted to be robust against changes in the mechanical system design indicating that the reflex system can use different morphological designs, which does not apply for most robotic systems (for which the control often follows a specific

  1. System Considerations and Challendes in 3d Mapping and Modeling Using Low-Cost Uav Systems

    NASA Astrophysics Data System (ADS)

    Lari, Z.; El-Sheimy, N.

    2015-08-01

    In the last few years, low-cost UAV systems have been acknowledged as an affordable technology for geospatial data acquisition that can meet the needs of a variety of traditional and non-traditional mapping applications. In spite of its proven potential, UAV-based mapping is still lacking in terms of what is needed for it to become an acceptable mapping tool. In other words, a well-designed system architecture that considers payload restrictions as well as the specifications of the utilized direct geo-referencing component and the imaging systems in light of the required mapping accuracy and intended application is still required. Moreover, efficient data processing workflows, which are capable of delivering the mapping products with the specified quality while considering the synergistic characteristics of the sensors onboard, the wide range of potential users who might lack deep knowledge in mapping activities, and time constraints of emerging applications, are still needed to be adopted. Therefore, the introduced challenges by having low-cost imaging and georeferencing sensors onboard UAVs with limited payload capability, the necessity of efficient data processing techniques for delivering required products for intended applications, and the diversity of potential users with insufficient mapping-related expertise needs to be fully investigated and addressed by UAV-based mapping research efforts. This paper addresses these challenges and reviews system considerations, adaptive processing techniques, and quality assurance/quality control procedures for achievement of accurate mapping products from these systems.

  2. Mapping QTLs for water-use efficiency reveals the potential candidate genes involved in regulating the trait in apple under drought stress.

    PubMed

    Wang, Haibo; Zhao, Shuang; Mao, Ke; Dong, Qinglong; Liang, Bowen; Li, Chao; Wei, Zhiwei; Li, Mingjun; Ma, Fengwang

    2018-06-26

    Improvement of water-use efficiency (WUE) can effectively reduce production losses caused by drought stress. A better understanding of the genetic determination of WUE in crops under drought stress has great potential value for developing cultivars adapted to arid regions. To identify the genetic loci associated with WUE and reveal genes responsible for the trait in apple, we aim to map the quantitative trait loci (QTLs) for carbon isotope composition, the proxy for WUE, applying two contrasting irrigating regimes over the two-year experiment and search for the candidate genes encompassed in the mapped QTLs. We constructed a high-density genetic linkage map with 10,172 markers of apple, using single nucleotide polymorphism (SNP) markers obtained through restriction site-associated DNA sequencing (RADseq) and a final segregating population of 350 seedlings from the cross of Honeycrisp and Qinguan. In total, 33 QTLs were identified for carbon isotope composition in apple under both well-watered and drought-stressed conditions. Three QTLs were stable over 2 years under drought stress on linkage groups LG8, LG15 and LG16, as validated by Kompetitive Allele-Specific PCR (KASP) assays. In those validated QTLs, 258 genes were screened according to their Gene Ontology functional annotations. Among them, 28 genes were identified, which exhibited significant responses to drought stress in 'Honeycrisp' and/or 'Qinguan'. These genes are involved in signaling, photosynthesis, response to stresses, carbohydrate metabolism, protein metabolism and modification, hormone metabolism and transport, transport, respiration, transcriptional regulation, and development regulation. They, especially those for photoprotection and relevant signal transduction, are potential candidate genes connected with WUE regulation in drought-stressed apple. We detected three stable QTLs for carbon isotope composition in apple under drought stress over 2 years, and validated them by KASP assay. Twenty

  3. Efficient feature extraction from wide-area motion imagery by MapReduce in Hadoop

    NASA Astrophysics Data System (ADS)

    Cheng, Erkang; Ma, Liya; Blaisse, Adam; Blasch, Erik; Sheaff, Carolyn; Chen, Genshe; Wu, Jie; Ling, Haibin

    2014-06-01

    Wide-Area Motion Imagery (WAMI) feature extraction is important for applications such as target tracking, traffic management and accident discovery. With the increasing amount of WAMI collections and feature extraction from the data, a scalable framework is needed to handle the large amount of information. Cloud computing is one of the approaches recently applied in large scale or big data. In this paper, MapReduce in Hadoop is investigated for large scale feature extraction tasks for WAMI. Specifically, a large dataset of WAMI images is divided into several splits. Each split has a small subset of WAMI images. The feature extractions of WAMI images in each split are distributed to slave nodes in the Hadoop system. Feature extraction of each image is performed individually in the assigned slave node. Finally, the feature extraction results are sent to the Hadoop File System (HDFS) to aggregate the feature information over the collected imagery. Experiments of feature extraction with and without MapReduce are conducted to illustrate the effectiveness of our proposed Cloud-Enabled WAMI Exploitation (CAWE) approach.

  4. Effect of Co-segregating Markers on High-Density Genetic Maps and Prediction of Map Expansion Using Machine Learning Algorithms.

    PubMed

    N'Diaye, Amidou; Haile, Jemanesh K; Fowler, D Brian; Ammar, Karim; Pozniak, Curtis J

    2017-01-01

    unavoidable. Therefore, we suggest developers improve linkage mapping algorithms for efficient analysis of high-throughput data. This study outlines a practical strategy to estimate the IF due to the proportion of co-segregating markers and outlines a method to scale the length of the map accordingly.

  5. Effect of Co-segregating Markers on High-Density Genetic Maps and Prediction of Map Expansion Using Machine Learning Algorithms

    PubMed Central

    N’Diaye, Amidou; Haile, Jemanesh K.; Fowler, D. Brian; Ammar, Karim; Pozniak, Curtis J.

    2017-01-01

    expansion unavoidable. Therefore, we suggest developers improve linkage mapping algorithms for efficient analysis of high-throughput data. This study outlines a practical strategy to estimate the IF due to the proportion of co-segregating markers and outlines a method to scale the length of the map accordingly. PMID:28878789

  6. An efficient approach to the travelling salesman problem using self-organizing maps.

    PubMed

    Vieira, Frederico Carvalho; Dória Neto, Adrião Duarte; Costa, José Alfredo Ferreira

    2003-04-01

    This paper presents an approach to the well-known Travelling Salesman Problem (TSP) using Self-Organizing Maps (SOM). The SOM algorithm has interesting topological information about its neurons configuration on cartesian space, which can be used to solve optimization problems. Aspects of initialization, parameters adaptation, and complexity analysis of the proposed SOM based algorithm are discussed. The results show an average deviation of 3.7% from the optimal tour length for a set of 12 TSP instances.

  7. Forest type mapping of the Interior West

    Treesearch

    Bonnie Ruefenacht; Gretchen G. Moisen; Jock A. Blackard

    2004-01-01

    This paper develops techniques for the mapping of forest types in Arizona, New Mexico, and Wyoming. The methods involve regression-tree modeling using a variety of remote sensing and GIS layers along with Forest Inventory Analysis (FIA) point data. Regression-tree modeling is a fast and efficient technique of estimating variables for large data sets with high accuracy...

  8. Automated land-use mapping from spacecraft data. [Oakland County, Michigan

    NASA Technical Reports Server (NTRS)

    Chase, P. E. (Principal Investigator); Rogers, R. H.; Reed, L. E.

    1974-01-01

    The author has identified the following significant results. In response to the need for a faster, more economical means of producing land use maps, this study evaluated the suitability of using ERTS-1 computer compatible tape (CCT) data as a basis for automatic mapping. Significant findings are: (1) automatic classification accuracy greater than 90% is achieved on categories of deep and shallow water, tended grass, rangeland, extractive (bare earth), urban, forest land, and nonforested wet lands; (2) computer-generated printouts by target class provide a quantitative measure of land use; and (3) the generation of map overlays showing land use from ERTS-1 CCTs offers a significant breakthrough in the rate at which land use maps are generated. Rather than uncorrected classified imagery or computer line printer outputs, the processing results in geometrically-corrected computer-driven pen drawing of land categories, drawn on a transparent material at a scale specified by the operator. These map overlays are economically produced and provide an efficient means of rapidly updating maps showing land use.

  9. 78 FR 40960 - Drawbridge Operation Regulation; Lake Washington Ship Canal at Seattle, WA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-07-09

    ..., WA. The deviation is necessary to accommodate the ``See Jane Run Women's Half Marathon.'' This... necessary to accommodate the ``See Jane Run Women's Half Marathon''. This deviation allows the bridges to...

  10. Mapping proteins to disease terminologies: from UniProt to MeSH

    PubMed Central

    Mottaz, Anaïs; Yip, Yum L; Ruch, Patrick; Veuthey, Anne-Lise

    2008-01-01

    Background Although the UniProt KnowledgeBase is not a medical-oriented database, it contains information on more than 2,000 human proteins involved in pathologies. However, these annotations are not standardized, which impairs the interoperability between biological and clinical resources. In order to make these data easily accessible to clinical researchers, we have developed a procedure to link diseases described in the UniProtKB/Swiss-Prot entries to the MeSH disease terminology. Results We mapped disease names extracted either from the UniProtKB/Swiss-Prot entry comment lines or from the corresponding OMIM entry to the MeSH. Different methods were assessed on a benchmark set of 200 disease names manually mapped to MeSH terms. The performance of the retained procedure in term of precision and recall was 86% and 64% respectively. Using the same procedure, more than 3,000 disease names in Swiss-Prot were mapped to MeSH with comparable efficiency. Conclusions This study is a first attempt to link proteins in UniProtKB to the medical resources. The indexing we provided will help clinicians and researchers navigate from diseases to genes and from genes to diseases in an efficient way. The mapping is available at: . PMID:18460185

  11. Map Downloads | USDA Plant Hardiness Zone Map

    Science.gov Websites

    formats. National, regional, and state maps are available under the View Maps section. Print Quality Maps dpi Graphic TIF 222 MB US Map 300 dpi Adobe Photoshop PS 25 MB *Print quality maps are very large | Non-Discrimination Statement | Information Quality | USA.gov | Whitehouse.gov

  12. Site Map | USDA Plant Hardiness Zone Map

    Science.gov Websites

    Acknowledgments & Citation Copyright Map & Data Downloads Map Downloads Geography (GIS) Downloads Multi ; Citation Copyright Map & Data Downloads Map Downloads Geography (GIS) Downloads Multi-ZIP Code Finder

  13. destiny: diffusion maps for large-scale single-cell data in R.

    PubMed

    Angerer, Philipp; Haghverdi, Laleh; Büttner, Maren; Theis, Fabian J; Marr, Carsten; Buettner, Florian

    2016-04-15

    : Diffusion maps are a spectral method for non-linear dimension reduction and have recently been adapted for the visualization of single-cell expression data. Here we present destiny, an efficient R implementation of the diffusion map algorithm. Our package includes a single-cell specific noise model allowing for missing and censored values. In contrast to previous implementations, we further present an efficient nearest-neighbour approximation that allows for the processing of hundreds of thousands of cells and a functionality for projecting new data on existing diffusion maps. We exemplarily apply destiny to a recent time-resolved mass cytometry dataset of cellular reprogramming. destiny is an open-source R/Bioconductor package "bioconductor.org/packages/destiny" also available at www.helmholtz-muenchen.de/icb/destiny A detailed vignette describing functions and workflows is provided with the package. carsten.marr@helmholtz-muenchen.de or f.buettner@helmholtz-muenchen.de Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  14. New conformal mapping for adaptive resolving of the complex singularities of Stokes wave

    PubMed Central

    Dyachenko, Sergey A.; A. Silantyev, Denis

    2017-01-01

    A new highly efficient method is developed for computation of travelling periodic waves (Stokes waves) on the free surface of deep water. A convergence of numerical approximation is determined by the complex singularities above the free surface for the analytical continuation of the travelling wave into the complex plane. An auxiliary conformal mapping is introduced which moves singularities away from the free surface thus dramatically speeding up numerical convergence by adapting the numerical grid for resolving singularities while being consistent with the fluid dynamics. The efficiency of that conformal mapping is demonstrated for the Stokes wave approaching the limiting Stokes wave (the wave of the greatest height) which significantly expands the family of numerically accessible solutions. It allows us to provide a detailed study of the oscillatory approach of these solutions to the limiting wave. Generalizations of the conformal mapping to resolve multiple singularities are also introduced. PMID:28690418

  15. New conformal mapping for adaptive resolving of the complex singularities of Stokes wave.

    PubMed

    Lushnikov, Pavel M; Dyachenko, Sergey A; A Silantyev, Denis

    2017-06-01

    A new highly efficient method is developed for computation of travelling periodic waves (Stokes waves) on the free surface of deep water. A convergence of numerical approximation is determined by the complex singularities above the free surface for the analytical continuation of the travelling wave into the complex plane. An auxiliary conformal mapping is introduced which moves singularities away from the free surface thus dramatically speeding up numerical convergence by adapting the numerical grid for resolving singularities while being consistent with the fluid dynamics. The efficiency of that conformal mapping is demonstrated for the Stokes wave approaching the limiting Stokes wave (the wave of the greatest height) which significantly expands the family of numerically accessible solutions. It allows us to provide a detailed study of the oscillatory approach of these solutions to the limiting wave. Generalizations of the conformal mapping to resolve multiple singularities are also introduced.

  16. Predicting successful tactile mapping of virtual objects.

    PubMed

    Brayda, Luca; Campus, Claudio; Gori, Monica

    2013-01-01

    Improving spatial ability of blind and visually impaired people is the main target of orientation and mobility (O&M) programs. In this study, we use a minimalistic mouse-shaped haptic device to show a new approach aimed at evaluating devices providing tactile representations of virtual objects. We consider psychophysical, behavioral, and subjective parameters to clarify under which circumstances mental representations of spaces (cognitive maps) can be efficiently constructed with touch by blindfolded sighted subjects. We study two complementary processes that determine map construction: low-level perception (in a passive stimulation task) and high-level information integration (in an active exploration task). We show that jointly considering a behavioral measure of information acquisition and a subjective measure of cognitive load can give an accurate prediction and a practical interpretation of mapping performance. Our simple TActile MOuse (TAMO) uses haptics to assess spatial ability: this may help individuals who are blind or visually impaired to be better evaluated by O&M practitioners or to evaluate their own performance.

  17. Demonstration of Hadoop-GIS: A Spatial Data Warehousing System Over MapReduce

    PubMed Central

    Aji, Ablimit; Sun, Xiling; Vo, Hoang; Liu, Qioaling; Lee, Rubao; Zhang, Xiaodong; Saltz, Joel; Wang, Fusheng

    2016-01-01

    The proliferation of GPS-enabled devices, and the rapid improvement of scientific instruments have resulted in massive amounts of spatial data in the last decade. Support of high performance spatial queries on large volumes data has become increasingly important in numerous fields, which requires a scalable and efficient spatial data warehousing solution as existing approaches exhibit scalability limitations and efficiency bottlenecks for large scale spatial applications. In this demonstration, we present Hadoop-GIS – a scalable and high performance spatial query system over MapReduce. Hadoop-GIS provides an efficient spatial query engine to process spatial queries, data and space based partitioning, and query pipelines that parallelize queries implicitly on MapReduce. Hadoop-GIS also provides an expressive, SQL-like spatial query language for workload specification. We will demonstrate how spatial queries are expressed in spatially extended SQL queries, and submitted through a command line/web interface for execution. Parallel to our system demonstration, we explain the system architecture and details on how queries are translated to MapReduce operators, optimized, and executed on Hadoop. In addition, we will showcase how the system can be used to support two representative real world use cases: large scale pathology analytical imaging, and geo-spatial data warehousing. PMID:27617325

  18. Demonstration of Hadoop-GIS: A Spatial Data Warehousing System Over MapReduce.

    PubMed

    Aji, Ablimit; Sun, Xiling; Vo, Hoang; Liu, Qioaling; Lee, Rubao; Zhang, Xiaodong; Saltz, Joel; Wang, Fusheng

    2013-11-01

    The proliferation of GPS-enabled devices, and the rapid improvement of scientific instruments have resulted in massive amounts of spatial data in the last decade. Support of high performance spatial queries on large volumes data has become increasingly important in numerous fields, which requires a scalable and efficient spatial data warehousing solution as existing approaches exhibit scalability limitations and efficiency bottlenecks for large scale spatial applications. In this demonstration, we present Hadoop-GIS - a scalable and high performance spatial query system over MapReduce. Hadoop-GIS provides an efficient spatial query engine to process spatial queries, data and space based partitioning, and query pipelines that parallelize queries implicitly on MapReduce. Hadoop-GIS also provides an expressive, SQL-like spatial query language for workload specification. We will demonstrate how spatial queries are expressed in spatially extended SQL queries, and submitted through a command line/web interface for execution. Parallel to our system demonstration, we explain the system architecture and details on how queries are translated to MapReduce operators, optimized, and executed on Hadoop. In addition, we will showcase how the system can be used to support two representative real world use cases: large scale pathology analytical imaging, and geo-spatial data warehousing.

  19. Large-scale mapping of mutations affecting zebrafish development.

    PubMed

    Geisler, Robert; Rauch, Gerd-Jörg; Geiger-Rudolph, Silke; Albrecht, Andrea; van Bebber, Frauke; Berger, Andrea; Busch-Nentwich, Elisabeth; Dahm, Ralf; Dekens, Marcus P S; Dooley, Christopher; Elli, Alexandra F; Gehring, Ines; Geiger, Horst; Geisler, Maria; Glaser, Stefanie; Holley, Scott; Huber, Matthias; Kerr, Andy; Kirn, Anette; Knirsch, Martina; Konantz, Martina; Küchler, Axel M; Maderspacher, Florian; Neuhauss, Stephan C; Nicolson, Teresa; Ober, Elke A; Praeg, Elke; Ray, Russell; Rentzsch, Brit; Rick, Jens M; Rief, Eva; Schauerte, Heike E; Schepp, Carsten P; Schönberger, Ulrike; Schonthaler, Helia B; Seiler, Christoph; Sidi, Samuel; Söllner, Christian; Wehner, Anja; Weiler, Christian; Nüsslein-Volhard, Christiane

    2007-01-09

    Large-scale mutagenesis screens in the zebrafish employing the mutagen ENU have isolated several hundred mutant loci that represent putative developmental control genes. In order to realize the potential of such screens, systematic genetic mapping of the mutations is necessary. Here we report on a large-scale effort to map the mutations generated in mutagenesis screening at the Max Planck Institute for Developmental Biology by genome scanning with microsatellite markers. We have selected a set of microsatellite markers and developed methods and scoring criteria suitable for efficient, high-throughput genome scanning. We have used these methods to successfully obtain a rough map position for 319 mutant loci from the Tübingen I mutagenesis screen and subsequent screening of the mutant collection. For 277 of these the corresponding gene is not yet identified. Mapping was successful for 80 % of the tested loci. By comparing 21 mutation and gene positions of cloned mutations we have validated the correctness of our linkage group assignments and estimated the standard error of our map positions to be approximately 6 cM. By obtaining rough map positions for over 300 zebrafish loci with developmental phenotypes, we have generated a dataset that will be useful not only for cloning of the affected genes, but also to suggest allelism of mutations with similar phenotypes that will be identified in future screens. Furthermore this work validates the usefulness of our methodology for rapid, systematic and inexpensive microsatellite mapping of zebrafish mutations.

  20. Linkage map of the honey bee, Apis mellifera, based on RAPD markers

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hunt, G.J.; Page, R.E. Jr.

    A linkage map was constructed for the honey bee based on the segregation of 365 random amplified polymorphic DNA (RAPD) markers in haploid male progeny of a single female bee. The X locus for sex determination and genes for black body color and malate dehydrogenase were mapped to separate linkage groups. RAPD markers were very efficient for mapping, with an average of about 2.8 loci mapped for each 10-nucleotide primer that was used in polymerase chain reactions. The mean interval size between markers on the map was 9.1 cM. The map covered 3110 cM of linked markers on 26 linkagemore » groups. We estimate the total genome size to be {approximately}3450 cM. The size of the map indicated a very high recombination rate for the honey bee. The relationship of physical to genetic distance was estimated at 52 kb/cM, suggesting that map-based cloning of genes will be feasible for this species. 71 refs., 6 figs., 1 tab.« less

  1. Experiences from using Autonomous Underwater Vehicles and Synthetic Aperture Sonar for Sediment and Habitat Mapping

    NASA Astrophysics Data System (ADS)

    Thorsnes, T.; Bjarnadóttir, L. R.

    2017-12-01

    Emerging platforms and tools like autonomous underwater vehicles and synthetic aperture sonars provide interesting opportunities for making seabed mapping more efficient and precise. Sediment grain-size maps are an important product in their own right and a key input for habitat and biotope maps. National and regional mapping programmes are tasked with mapping large areas, and survey efficiency, data quality, and resulting map confidence are important considerations when selecting the mapping strategy. Since 2005, c. 175,000 square kilometres of the Norwegian continental shelf and continental slope has been mapped with respect to sediments, habitats and biodiversity, and pollution under the MAREANO programme (www.mareano.no). At present the sediment mapping is based on a combination of ship-borne multibeam bathymetry and backscatter, visual documentation using a towed video platform, and grab sampling. We have now tested a new approach, using an Autonomous Underwater Vehicle (AUV) as the survey platform for the collection of acoustic data (Synthetic Aperture Sonar (SAS), EM2040 bathymetry and backscatter) and visual data (still images using a TFish colour photo system). This pilot project was conducted together the Norwegian Hydrographic Service, the Institute of Marine Research (biology observations) and the Norwegian Defence Research Establishment (operation of ship and AUV). The test site reported here is the Vesterdjupet area, offshore Lofoten, northern Norway. The water depth is between 170 and 300 metres, with sediments ranging from gravel, cobbles and boulders to sandy mud. A cold-water coral reef, associated with bioclastic sediments was also present in the study area. The presentation will give an overview of the main findings and experiences gained from this pilot project with a focus on geological mapping and will also discuss the relevance of AUV-based mapping to large-area mapping programmes like MAREANO.

  2. Fault-Tolerant and Elastic Streaming MapReduce with Decentralized Coordination

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kumbhare, Alok; Frincu, Marc; Simmhan, Yogesh

    2015-06-29

    The MapReduce programming model, due to its simplicity and scalability, has become an essential tool for processing large data volumes in distributed environments. Recent Stream Processing Systems (SPS) extend this model to provide low-latency analysis of high-velocity continuous data streams. However, integrating MapReduce with streaming poses challenges: first, the runtime variations in data characteristics such as data-rates and key-distribution cause resource overload, that inturn leads to fluctuations in the Quality of the Service (QoS); and second, the stateful reducers, whose state depends on the complete tuple history, necessitates efficient fault-recovery mechanisms to maintain the desired QoS in the presence ofmore » resource failures. We propose an integrated streaming MapReduce architecture leveraging the concept of consistent hashing to support runtime elasticity along with locality-aware data and state replication to provide efficient load-balancing with low-overhead fault-tolerance and parallel fault-recovery from multiple simultaneous failures. Our evaluation on a private cloud shows up to 2:8 improvement in peak throughput compared to Apache Storm SPS, and a low recovery latency of 700 -1500 ms from multiple failures.« less

  3. Genetic Map of Mango: A Tool for Mango Breeding

    PubMed Central

    Kuhn, David N.; Bally, Ian S. E.; Dillon, Natalie L.; Innes, David; Groh, Amy M.; Rahaman, Jordon; Ophir, Ron; Cohen, Yuval; Sherman, Amir

    2017-01-01

    Mango (Mangifera indica) is an economically and nutritionally important tropical/subtropical tree fruit crop. Most of the current commercial cultivars are selections rather than the products of breeding programs. To improve the efficiency of mango breeding, molecular markers have been used to create a consensus genetic map that identifies all 20 linkage groups in seven mapping populations. Polyembryony is an important mango trait, used for clonal propagation of cultivars and rootstocks. In polyembryonic mango cultivars, in addition to a zygotic embryo, several apomictic embryos develop from maternal tissue surrounding the fertilized egg cell. This trait has been associated with linkage group 8 in our consensus genetic map and has been validated in two of the seven mapping populations. In addition, we have observed a significant association between trait and single nucleotide polymorphism (SNP) markers for the vegetative trait of branch habit and the fruit traits of bloom, ground skin color, blush intensity, beak shape, and pulp color. PMID:28473837

  4. Detection And Mapping (DAM) package. Volume 4B: Software System Manual, part 2

    NASA Technical Reports Server (NTRS)

    Schlosser, E. H.

    1980-01-01

    Computer programs, graphic devices, and an integrated set of manual procedures designed for efficient production of precisely registered and formatted maps from digital data are presented. The software can be used on any Univac 1100 series computer. The software includes pre-defined spectral limits for use in classifying and mapping surface water for LANDSAT-1, LANDSAT-2, and LANDSAT-3.

  5. Molecular mapping of chromosomes 17 and X

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Barker, D.F.

    1991-01-15

    Progress toward the construction of high density genetic maps of chromosomes 17 and X has been made by isolating and characterizing a relatively large set of polymorphic probes for each chromosome and using these probes to construct genetic maps. We have mapped the same polymorphic probes against a series of chromosome breakpoints on X and 17. The probes could be assigned to over 30 physical intervals on the X chromosome and 7 intervals on 17. In many cases, this process resulted in improved characterization of the relative locations of the breakpoints with respect to each other and the definition ofmore » new physical intervals. The strategy for isolation of the polymorphic clones utilized chromosome specific libraries of 1--15 kb segments from each of the two chromosomes. From these libraries, clones were screened for those detecting restriction fragment length polymorphisms. The markers were further characterized, the chromosomal assignments confirmed and in most cases segments of the original probes were subcloned into plasmids to produce probes with improved signal to noise ratios for use in the genetic marker studies. The linkage studies utilize the CEPH reference families and other well-characterized families in our collection which have been used for genetic disease linkage work. Preliminary maps and maps of portions of specific regions of 17 and X are provided. We have nearly completed a map of the 1 megabase Mycoplasma arthritidis genome by applying these techniques to a lambda phage library of its genome. We have found bit mapping to be an efficient means to organize a contiguous set of overlapping clones from a larger genome.« less

  6. Maps & minds : mapping through the ages

    USGS Publications Warehouse

    ,

    1984-01-01

    Throughout time, maps have expressed our understanding of our world. Human affairs have been influenced strongly by the quality of maps available to us at the major turning points in our history. "Maps & Minds" traces the ebb and flow of a few central ideas in the mainstream of mapping. Our expanding knowledge of our cosmic neighborhood stems largely from a small number of simple but grand ideas, vigorously pursued.

  7. Navigating 3D electron microscopy maps with EM-SURFER.

    PubMed

    Esquivel-Rodríguez, Juan; Xiong, Yi; Han, Xusi; Guang, Shuomeng; Christoffer, Charles; Kihara, Daisuke

    2015-05-30

    The Electron Microscopy DataBank (EMDB) is growing rapidly, accumulating biological structural data obtained mainly by electron microscopy and tomography, which are emerging techniques for determining large biomolecular complex and subcellular structures. Together with the Protein Data Bank (PDB), EMDB is becoming a fundamental resource of the tertiary structures of biological macromolecules. To take full advantage of this indispensable resource, the ability to search the database by structural similarity is essential. However, unlike high-resolution structures stored in PDB, methods for comparing low-resolution electron microscopy (EM) density maps in EMDB are not well established. We developed a computational method for efficiently searching low-resolution EM maps. The method uses a compact fingerprint representation of EM maps based on the 3D Zernike descriptor, which is derived from a mathematical series expansion for EM maps that are considered as 3D functions. The method is implemented in a web server named EM-SURFER, which allows users to search against the entire EMDB in real-time. EM-SURFER compares the global shapes of EM maps. Examples of search results from different types of query structures are discussed. We developed EM-SURFER, which retrieves structurally relevant matches for query EM maps from EMDB within seconds. The unique capability of EM-SURFER to detect 3D shape similarity of low-resolution EM maps should prove invaluable in structural biology.

  8. Current trends in geomorphological mapping

    NASA Astrophysics Data System (ADS)

    Seijmonsbergen, A. C.

    2012-04-01

    , and in the near future also morphogenetic information. As a result, these new opportunities have changed the workflows for geomorphological mapmaking, and their focus have shifted from field-based techniques to using more computer-based techniques: for example, traditional pre-field air-photo based maps are now replaced by maps prepared in a digital mapping environment, and designated field visits using mobile GIS / digital mapping devices now focus on gathering location information and attribute inventories and are strongly time efficient. The resulting 'modern geomorphological maps' are digital collections of geomorphological information layers consisting of georeferenced vector, raster and tabular data which are stored in a digital environment such as a GIS geodatabase, and are easily visualized as e.g. 'birds' eye' views, as animated 3D displays, on virtual globes, or stored as GeoPDF maps in which georeferenced attribute information can be easily exchanged over the internet. Digital geomorphological information layers are increasingly accessed via web-based services distributed through remote servers. Information can be consulted - or even build using remote geoprocessing servers - by the end user. Therefore, it will not only be the geomorphologist anymore, but also the professional end user that dictates the applied use of digital geomorphological information layers.

  9. MapEdit: solution to continuous raster map creation

    NASA Astrophysics Data System (ADS)

    Rančić, Dejan; Djordjevi-Kajan, Slobodanka

    2003-03-01

    The paper describes MapEdit, MS Windows TM software for georeferencing and rectification of scanned paper maps. The software produces continuous raster maps which can be used as background in geographical information systems. Process of continuous raster map creation using MapEdit "mosaicking" function is also described as well as the georeferencing and rectification algorithms which are used in MapEdit. Our approach for georeferencing and rectification using four control points and two linear transformations for each scanned map part, together with nearest neighbor resampling method, represents low cost—high speed solution that produce continuous raster maps with satisfactory quality for many purposes (±1 pixel). Quality assessment of several continuous raster maps at different scales that have been created using our software and methodology, has been undertaken and results are presented in the paper. For the quality control of the produced raster maps we referred to three wide adopted standards: US Standard for Digital Cartographic Data, National Standard for Spatial Data Accuracy and US National Map Accuracy Standard. The results obtained during the quality assessment process are given in the paper and show that our maps meat all three standards.

  10. Global land cover mapping: a review and uncertainty analysis

    USGS Publications Warehouse

    Congalton, Russell G.; Gu, Jianyu; Yadav, Kamini; Thenkabail, Prasad S.; Ozdogan, Mutlu

    2014-01-01

    Given the advances in remotely sensed imagery and associated technologies, several global land cover maps have been produced in recent times including IGBP DISCover, UMD Land Cover, Global Land Cover 2000 and GlobCover 2009. However, the utility of these maps for specific applications has often been hampered due to considerable amounts of uncertainties and inconsistencies. A thorough review of these global land cover projects including evaluating the sources of error and uncertainty is prudent and enlightening. Therefore, this paper describes our work in which we compared, summarized and conducted an uncertainty analysis of the four global land cover mapping projects using an error budget approach. The results showed that the classification scheme and the validation methodology had the highest error contribution and implementation priority. A comparison of the classification schemes showed that there are many inconsistencies between the definitions of the map classes. This is especially true for the mixed type classes for which thresholds vary for the attributes/discriminators used in the classification process. Examination of these four global mapping projects provided quite a few important lessons for the future global mapping projects including the need for clear and uniform definitions of the classification scheme and an efficient, practical, and valid design of the accuracy assessment.

  11. Flexible, reconfigurable, power efficient transmitter and method

    NASA Technical Reports Server (NTRS)

    Bishop, James W. (Inventor); Zaki, Nazrul H. Mohd (Inventor); Newman, David Childress (Inventor); Bundick, Steven N. (Inventor)

    2011-01-01

    A flexible, reconfigurable, power efficient transmitter device and method is provided. In one embodiment, the method includes receiving outbound data and determining a mode of operation. When operating in a first mode the method may include modulation mapping the outbound data according a modulation scheme to provide first modulation mapped digital data, converting the first modulation mapped digital data to an analog signal that comprises an intermediate frequency (IF) analog signal, upconverting the IF analog signal to produce a first modulated radio frequency (RF) signal based on a local oscillator signal, amplifying the first RF modulated signal to produce a first RF output signal, and outputting the first RF output signal via an isolator. In a second mode of operation method may include modulation mapping the outbound data according a modulation scheme to provide second modulation mapped digital data, converting the second modulation mapped digital data to a first digital baseband signal, conditioning the first digital baseband signal to provide a first analog baseband signal, modulating one or more carriers with the first analog baseband signal to produce a second modulated RF signal based on a local oscillator signal, amplifying the second RF modulated signal to produce a second RF output signal, and outputting the second RF output signal via the isolator. The digital baseband signal may comprise an in-phase (I) digital baseband signal and a quadrature (Q) baseband signal.

  12. A Discrete Global Grid System Programming Language Using MapReduce

    NASA Astrophysics Data System (ADS)

    Peterson, P.; Shatz, I.

    2016-12-01

    A discrete global grid system (DGGS) is a powerful mechanism for storing and integrating geospatial information. As a "pixelization" of the Earth, many image processing techniques lend themselves to the transformation of data values referenced to the DGGS cells. It has been shown that image algebra, as an example, and advanced algebra, like Fast Fourier Transformation, can be used on the DGGS tiling structure for geoprocessing and spatial analysis. MapReduce has been shown to provide advantages for processing and generating large data sets within distributed and parallel computing. The DGGS structure is ideally suited for big distributed Earth data. We proposed that basic expressions could be created to form the atoms of a generalized DGGS language using the MapReduce programming model. We created three very efficient expressions: Selectors (aka filter) - A selection function that generate a set of cells, cell collections, or geometries; Calculators (aka map) - A computational function (including quantization of raw measurements and data sources) that generate values in a DGGS cell; and Aggregators (aka reduce) - A function that generate spatial statistics from cell values within a cell. We found that these three basic MapReduce operations along with a forth function, the Iterator, for horizontal and vertical traversing of any DGGS structure, provided simple building block resulting in very efficient operations and processes that could be used with any DGGS. We provide examples and a demonstration of their effectiveness using the ISEA3H DGGS on the PYXIS Studio.

  13. Direct and accelerated parameter mapping using the unscented Kalman filter.

    PubMed

    Zhao, Li; Feng, Xue; Meyer, Craig H

    2016-05-01

    To accelerate parameter mapping using a new paradigm that combines image reconstruction and model regression as a parameter state-tracking problem. In T2 mapping, the T2 map is first encoded in parameter space by multi-TE measurements and then encoded by Fourier transformation with readout/phase encoding gradients. Using a state transition function and a measurement function, the unscented Kalman filter can describe T2 mapping as a dynamic system and directly estimate the T2 map from the k-space data. The proposed method was validated with a numerical brain phantom and volunteer experiments with a multiple-contrast spin echo sequence. Its performance was compared with a conjugate-gradient nonlinear inversion method at undersampling factors of 2 to 8. An accelerated pulse sequence was developed based on this method to achieve prospective undersampling. Compared with the nonlinear inversion reconstruction, the proposed method had higher precision, improved structural similarity and reduced normalized root mean squared error, with acceleration factors up to 8 in numerical phantom and volunteer studies. This work describes a new perspective on parameter mapping by state tracking. The unscented Kalman filter provides a highly accelerated and efficient paradigm for T2 mapping. © 2015 Wiley Periodicals, Inc.

  14. Assessment of LANDSAT for rangeland mapping, Rush Valley, Utah

    NASA Technical Reports Server (NTRS)

    Ridd, M. K.; Price, K. P.; Douglass, G. E.

    1984-01-01

    The feasibility of using LANDSAT MSS (multispectral scanner) data to identify and map cover types for rangeland, and to determine comparative condition of the ecotypes was assessed. A supporting objective is to assess the utility of various forms of aerial photography in the process. If rangelands can be efficiently mapped with Landsat data, as supported by appropriate aerial photography and field data, then uniform standards of cover classification and condition may be applied across the rangelands of the state. Further, a foundation may be established for long-term monitoring of range trend, using the same satellite system over time.

  15. Single Nucleotide Polymorphism Markers for Genetic Mapping in Drosophila melanogaster

    PubMed Central

    Hoskins, Roger A.; Phan, Alexander C.; Naeemuddin, Mohammed; Mapa, Felipa A.; Ruddy, David A.; Ryan, Jessica J.; Young, Lynn M.; Wells, Trent; Kopczynski, Casey; Ellis, Michael C.

    2001-01-01

    For nearly a century, genetic analysis in Drosophila melanogaster has been a powerful tool for analyzing gene function, yet Drosophila lacks the molecular genetic mapping tools that recently have revolutionized human, mouse, and plant genetics. Here, we describe the systematic characterization of a dense set of molecular markers in Drosophila by using a sequence tagged site-based physical map of the genome. We identify 474 biallelic markers in standard laboratory strains of Drosophila that span the genome. Most of these markers are single nucleotide polymorphisms and sequences for these variants are provided in an accessible format. The average density of the new markers is one per 225 kb on the autosomes and one per megabase on the X chromosome. We include in this survey a set of P-element strains that provide additional use for high-resolution mapping. We show one application of the new markers in a simple set of crosses to map a mutation in the hedgehog gene to an interval of <1 Mb. This new map resource significantly increases the efficiency and resolution of recombination mapping and will be of immediate value to the Drosophila research community. PMID:11381036

  16. Single nucleotide polymorphism markers for genetic mapping in Drosophila melanogaster

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hoskins, Roger A.; Phan, Alexander C.; Naeemuddin, Mohammed

    2001-04-16

    For nearly a century, genetic analysis in Drosophila melanogaster has been a powerful tool for analyzing gene function, yet Drosophila lacks the molecular genetic mapping tools that have recently revolutionized human, mouse and plant genetics. Here, we describe the systematic characterization of a dense set of molecular markers in Drosophila using an STS-based physical map of the genome. We identify 474 biallelic markers in standard laboratory strains of Drosophila that the genome. The majority of these markers are single nucleotide polymorphisms (SNPs) and sequences for these variants are provided in an accessible format. The average density of the new markersmore » is 1 marker per 225 kb on the autosomes and 1 marker per 1 Mb on the X chromosome. We include in this survey a set of P-element strains that provide additional utility for high-resolution mapping. We demonstrate one application of the new markers in a simple set of crosses to map a mutation in the hedgehog gene to an interval of <1 Mb. This new map resource significantly increases the efficiency and resolution of recombination mapping and will be of immediate value to the Drosophila research community.« less

  17. Interactive Mapping on Virtual Terrain Models Using RIMS (Real-time, Interactive Mapping System)

    NASA Astrophysics Data System (ADS)

    Bernardin, T.; Cowgill, E.; Gold, R. D.; Hamann, B.; Kreylos, O.; Schmitt, A.

    2006-12-01

    geophysicists, and planetary scientists. The strength of our system is that it combines interactive rendering with interactive mapping and measurement of features observed in topographic and texture data. Comparison with commercially available software indicates that our system improves mapping accuracy and efficiency. More importantly, it enables Earth scientists to rapidly achieve a deeper level of understanding of remotely sensed data, as observations can be made that are not possible with existing systems.

  18. BatSLAM: Simultaneous localization and mapping using biomimetic sonar.

    PubMed

    Steckel, Jan; Peremans, Herbert

    2013-01-01

    We propose to combine a biomimetic navigation model which solves a simultaneous localization and mapping task with a biomimetic sonar mounted on a mobile robot to address two related questions. First, can robotic sonar sensing lead to intelligent interactions with complex environments? Second, can we model sonar based spatial orientation and the construction of spatial maps by bats? To address these questions we adapt the mapping module of RatSLAM, a previously published navigation system based on computational models of the rodent hippocampus. We analyze the performance of the proposed robotic implementation operating in the real world. We conclude that the biomimetic navigation model operating on the information from the biomimetic sonar allows an autonomous agent to map unmodified (office) environments efficiently and consistently. Furthermore, these results also show that successful navigation does not require the readings of the biomimetic sonar to be interpreted in terms of individual objects/landmarks in the environment. We argue that the system has applications in robotics as well as in the field of biology as a simple, first order, model for sonar based spatial orientation and map building.

  19. BatSLAM: Simultaneous Localization and Mapping Using Biomimetic Sonar

    PubMed Central

    Steckel, Jan; Peremans, Herbert

    2013-01-01

    We propose to combine a biomimetic navigation model which solves a simultaneous localization and mapping task with a biomimetic sonar mounted on a mobile robot to address two related questions. First, can robotic sonar sensing lead to intelligent interactions with complex environments? Second, can we model sonar based spatial orientation and the construction of spatial maps by bats? To address these questions we adapt the mapping module of RatSLAM, a previously published navigation system based on computational models of the rodent hippocampus. We analyze the performance of the proposed robotic implementation operating in the real world. We conclude that the biomimetic navigation model operating on the information from the biomimetic sonar allows an autonomous agent to map unmodified (office) environments efficiently and consistently. Furthermore, these results also show that successful navigation does not require the readings of the biomimetic sonar to be interpreted in terms of individual objects/landmarks in the environment. We argue that the system has applications in robotics as well as in the field of biology as a simple, first order, model for sonar based spatial orientation and map building. PMID:23365647

  20. Chaotic map clustering algorithm for EEG analysis

    NASA Astrophysics Data System (ADS)

    Bellotti, R.; De Carlo, F.; Stramaglia, S.

    2004-03-01

    The non-parametric chaotic map clustering algorithm has been applied to the analysis of electroencephalographic signals, in order to recognize the Huntington's disease, one of the most dangerous pathologies of the central nervous system. The performance of the method has been compared with those obtained through parametric algorithms, as K-means and deterministic annealing, and supervised multi-layer perceptron. While supervised neural networks need a training phase, performed by means of data tagged by the genetic test, and the parametric methods require a prior choice of the number of classes to find, the chaotic map clustering gives a natural evidence of the pathological class, without any training or supervision, thus providing a new efficient methodology for the recognition of patterns affected by the Huntington's disease.

  1. Developing Connective Leadership: Successes with Thinking Maps[R

    ERIC Educational Resources Information Center

    Alper, Larry; Williams, Kimberly; Hyerle, David

    2011-01-01

    "If our best thinking comes by making connections and building patterns, then what would these patterns look like, and what might they be based on?"--ask the authors. Most importantly, how could they be used? Developing Connective Leadership shows you how Thinking Maps[R] are an efficient and eloquent language that can be used to explore and…

  2. Planetary maps

    USGS Publications Warehouse

    ,

    1992-01-01

    An important goal of the USGS planetary mapping program is to systematically map the geology of the Moon, Mars, Venus, and Mercury, and the satellites of the outer planets. These geologic maps are published in the USGS Miscellaneous Investigations (I) Series. Planetary maps on sale at the USGS include shaded-relief maps, topographic maps, geologic maps, and controlled photomosaics. Controlled photomosaics are assembled from two or more photographs or images using a network of points of known latitude and longitude. The images used for most of these planetary maps are electronic images, obtained from orbiting television cameras, various optical-mechanical systems. Photographic film was only used to map Earth's Moon.

  3. Comprehensive Evaluation and Analysis of China's Mainstream Online Map Service Websites

    NASA Astrophysics Data System (ADS)

    Zhang, H.; Jiang, J.; Huang, W.; Wang, Q.; Gu, X.

    2012-08-01

    With the flourish development of China's Internet market, all kinds of users for map service demand is rising continually, within it contains tremendous commercial interests. Many internet giants have got involved in the field of online map service, and defined it as an important strategic product of the company. The main purpose of this research is to evaluate these online map service websites comprehensively with a model, and analyse the problems according to the evaluation results. Then some corresponding solving measures are proposed, which provides a theoretical and application guidance for the future development of fiercely competitive online map websites. The research consists of three stages: (a) the mainstream online map service websites in China are introduced and the present situation of them is analysed through visit, investigation, consultant, analysis and research. (b) a whole comprehensive evaluation quota system of online map service websites from the view of functions, layout, interaction design color position and so on, combining with the data indexes such as time efficiency, accuracy, objectivity and authority. (c) a comprehensive evaluation to these online map service websites is proceeded based on the fuzzy evaluation mathematical model, and the difficulty that measure the map websites quantitatively is solved.

  4. Flood Hazard Mapping by Applying Fuzzy TOPSIS Method

    NASA Astrophysics Data System (ADS)

    Han, K. Y.; Lee, J. Y.; Keum, H.; Kim, B. J.; Kim, T. H.

    2017-12-01

    There are lots of technical methods to integrate various factors for flood hazard mapping. The purpose of this study is to suggest the methodology of integrated flood hazard mapping using MCDM(Multi Criteria Decision Making). MCDM problems involve a set of alternatives that are evaluated on the basis of conflicting and incommensurate criteria. In this study, to apply MCDM to assessing flood risk, maximum flood depth, maximum velocity, and maximum travel time are considered as criterion, and each applied elements are considered as alternatives. The scheme to find the efficient alternative closest to a ideal value is appropriate way to assess flood risk of a lot of element units(alternatives) based on various flood indices. Therefore, TOPSIS which is most commonly used MCDM scheme is adopted to create flood hazard map. The indices for flood hazard mapping(maximum flood depth, maximum velocity, and maximum travel time) have uncertainty concerning simulation results due to various values according to flood scenario and topographical condition. These kind of ambiguity of indices can cause uncertainty of flood hazard map. To consider ambiguity and uncertainty of criterion, fuzzy logic is introduced which is able to handle ambiguous expression. In this paper, we made Flood Hazard Map according to levee breach overflow using the Fuzzy TOPSIS Technique. We confirmed the areas where the highest grade of hazard was recorded through the drawn-up integrated flood hazard map, and then produced flood hazard map can be compared them with those indicated in the existing flood risk maps. Also, we expect that if we can apply the flood hazard map methodology suggested in this paper even to manufacturing the current flood risk maps, we will be able to make a new flood hazard map to even consider the priorities for hazard areas, including more varied and important information than ever before. Keywords : Flood hazard map; levee break analysis; 2D analysis; MCDM; Fuzzy TOPSIS

  5. Classification of fMRI resting-state maps using machine learning techniques: A comparative study

    NASA Astrophysics Data System (ADS)

    Gallos, Ioannis; Siettos, Constantinos

    2017-11-01

    We compare the efficiency of Principal Component Analysis (PCA) and nonlinear learning manifold algorithms (ISOMAP and Diffusion maps) for classifying brain maps between groups of schizophrenia patients and healthy from fMRI scans during a resting-state experiment. After a standard pre-processing pipeline, we applied spatial Independent component analysis (ICA) to reduce (a) noise and (b) spatial-temporal dimensionality of fMRI maps. On the cross-correlation matrix of the ICA components, we applied PCA, ISOMAP and Diffusion Maps to find an embedded low-dimensional space. Finally, support-vector-machines (SVM) and k-NN algorithms were used to evaluate the performance of the algorithms in classifying between the two groups.

  6. Digital mapping techniques '06 - Workshop proceedings

    USGS Publications Warehouse

    Soller, David R.

    2007-01-01

    The Digital Mapping Techniques `06 (DMT`06) workshop was attended by more than 110 technical experts from 51 agencies, universities, and private companies, including representatives from 27 state geological surveys (see Appendix A of these Proceedings). This workshop was similar in nature to the previous nine meetings, which were held in Lawrence, Kansas (Soller, 1997), Champaign, Illinois (Soller, 1998), Madison, Wisconsin (Soller, 1999), Lexington, Kentucky (Soller, 2000), Tuscaloosa, Alabama (Soller, 2001), Salt Lake City, Utah (Soller, 2002), Millersville, Pennsylvania (Soller, 2003), Portland, Oregon (Soller, 2004), and Baton Rouge, Louisiana (Soller, 2005). This year?s meeting was hosted by the Ohio Geological Survey, from June 11-14, 2006, on the Ohio State University campus in Columbus, Ohio. As in the previous meetings, the objective was to foster informal discussion and exchange of technical information. It is with great pleasure that I note that the objective was successfully met, as attendees continued to share and exchange knowledge and information, and renew friendships and collegial work begun at past DMT workshops.Each DMT workshop has been coordinated by the Association of American State Geologists (AASG) and U.S. Geological Survey (USGS) Data Capture Working Group, the latter of which was formed in August 1996 to support the AASG and the USGS in their effort to build a National Geologic Map Database (see Soller, this volume, and http://ngmdb.usgs.gov/info/standards/datacapt/). The Working Group was formed because increased production efficiencies, standardization, and quality of digital map products were needed for the database - and for the State and Federal geological surveys - to provide more high-quality digital maps to the public.At the 2006 meeting, oral and poster presentations and special discussion sessions emphasized: 1) methods for creating and publishing map products (here, "publishing" includes Web-based release); 2) field data

  7. Calculating Higher-Order Moments of Phylogenetic Stochastic Mapping Summaries in Linear Time.

    PubMed

    Dhar, Amrit; Minin, Vladimir N

    2017-05-01

    Stochastic mapping is a simulation-based method for probabilistically mapping substitution histories onto phylogenies according to continuous-time Markov models of evolution. This technique can be used to infer properties of the evolutionary process on the phylogeny and, unlike parsimony-based mapping, conditions on the observed data to randomly draw substitution mappings that do not necessarily require the minimum number of events on a tree. Most stochastic mapping applications simulate substitution mappings only to estimate the mean and/or variance of two commonly used mapping summaries: the number of particular types of substitutions (labeled substitution counts) and the time spent in a particular group of states (labeled dwelling times) on the tree. Fast, simulation-free algorithms for calculating the mean of stochastic mapping summaries exist. Importantly, these algorithms scale linearly in the number of tips/leaves of the phylogenetic tree. However, to our knowledge, no such algorithm exists for calculating higher-order moments of stochastic mapping summaries. We present one such simulation-free dynamic programming algorithm that calculates prior and posterior mapping variances and scales linearly in the number of phylogeny tips. Our procedure suggests a general framework that can be used to efficiently compute higher-order moments of stochastic mapping summaries without simulations. We demonstrate the usefulness of our algorithm by extending previously developed statistical tests for rate variation across sites and for detecting evolutionarily conserved regions in genomic sequences.

  8. SPECTRAL LINE DE-CONFUSION IN AN INTENSITY MAPPING SURVEY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cheng, Yun-Ting; Bock, James; Bradford, C. Matt

    2016-12-01

    Spectral line intensity mapping (LIM) has been proposed as a promising tool to efficiently probe the cosmic reionization and the large-scale structure. Without detecting individual sources, LIM makes use of all available photons and measures the integrated light in the source confusion limit to efficiently map the three-dimensional matter distribution on large scales as traced by a given emission line. One particular challenge is the separation of desired signals from astrophysical continuum foregrounds and line interlopers. Here we present a technique to extract large-scale structure information traced by emission lines from different redshifts, embedded in a three-dimensional intensity mapping data cube.more » The line redshifts are distinguished by the anisotropic shape of the power spectra when projected onto a common coordinate frame. We consider the case where high-redshift [C ii] lines are confused with multiple low-redshift CO rotational lines. We present a semi-analytic model for [C ii] and CO line estimates based on the cosmic infrared background measurements, and show that with a modest instrumental noise level and survey geometry, the large-scale [C ii] and CO power spectrum amplitudes can be successfully extracted from a confusion-limited data set, without external information. We discuss the implications and limits of this technique for possible LIM experiments.« less

  9. Constraining the interaction between dark sectors with future HI intensity mapping observations

    NASA Astrophysics Data System (ADS)

    Xu, Xiaodong; Ma, Yin-Zhe; Weltman, Amanda

    2018-04-01

    We study a model of interacting dark matter and dark energy, in which the two components are coupled. We calculate the predictions for the 21-cm intensity mapping power spectra, and forecast the detectability with future single-dish intensity mapping surveys (BINGO, FAST and SKA-I). Since dark energy is turned on at z ˜1 , which falls into the sensitivity range of these radio surveys, the HI intensity mapping technique is an efficient tool to constrain the interaction. By comparing with current constraints on dark sector interactions, we find that future radio surveys will produce tight and reliable constraints on the coupling parameters.

  10. Evaluation of iconic versus F-map microburst displays

    NASA Technical Reports Server (NTRS)

    Salzberger, Mark; Hansman, R. John; Wanke, Craig

    1994-01-01

    Previous studies have shown graphical presentation methods of hazardous wind shear to be superior to textual or audible warnings alone. Positional information and the strength of the hazard were observed to be and were cited by pilots as the most important factors in a display. In this experiment the use of the three different graphical presentations of hazardous wind shear are examined. Airborne predictive detectors of wind shear enable the dissemination of varying levels of information. The effectiveness of iconic and mapping display modes of different complexities are addressed through simulation and analysis. Different positional and time-varying situations are presented in a 'part-task' Boeing 767 simulator using data from actual microburst events. Experienced airline pilots fly approach profiles using both iconic and F-map wind shear alerting displays. Microburst accompanied each event is also shown to the pilot. Mapping display types are expected to be found exceptionally efficient at conveying location comparison information while iconic displays simplify the threat recognition process. Preliminary results from the simulator study are presented. Recommendations concerning the suitability of multilevel iconic and mapping displays are made. Situational problems with current display prototypes are also addressed.

  11. Acoustic methods for cavitation mapping in biomedical applications

    NASA Astrophysics Data System (ADS)

    Wan, M.; Xu, S.; Ding, T.; Hu, H.; Liu, R.; Bai, C.; Lu, S.

    2015-12-01

    In recent years, cavitation is increasingly utilized in a wide range of applications in biomedical field. Monitoring the spatial-temporal evolution of cavitation bubbles is of great significance for efficiency and safety in biomedical applications. In this paper, several acoustic methods for cavitation mapping proposed or modified on the basis of existing work will be presented. The proposed novel ultrasound line-by-line/plane-by-plane method can depict cavitation bubbles distribution with high spatial and temporal resolution and may be developed as a potential standard 2D/3D cavitation field mapping method. The modified ultrafast active cavitation mapping based upon plane wave transmission and reception as well as bubble wavelet and pulse inversion technique can apparently enhance the cavitation to tissue ratio in tissue and further assist in monitoring the cavitation mediated therapy with good spatial and temporal resolution. The methods presented in this paper will be a foundation to promote the research and development of cavitation imaging in non-transparent medium.

  12. Active edge maps for medical image registration

    NASA Astrophysics Data System (ADS)

    Kerwin, William; Yuan, Chun

    2001-07-01

    Applying edge detection prior to performing image registration yields several advantages over raw intensity- based registration. Advantages include the ability to register multicontrast or multimodality images, immunity to intensity variations, and the potential for computationally efficient algorithms. In this work, a common framework for edge-based image registration is formulated as an adaptation of snakes used in boundary detection. Called active edge maps, the new formulation finds a one-to-one transformation T(x) that maps points in a source image to corresponding locations in a target image using an energy minimization approach. The energy consists of an image component that is small when edge features are well matched in the two images, and an internal term that restricts T(x) to allowable configurations. The active edge map formulation is illustrated here with a specific example developed for affine registration of carotid artery magnetic resonance images. In this example, edges are identified using a magnitude of gradient operator, image energy is determined using a Gaussian weighted distance function, and the internal energy includes separate, adjustable components that control volume preservation and rigidity.

  13. Direct mapping of symbolic DNA sequence into frequency domain in global repeat map algorithm

    PubMed Central

    Glunčić, Matko; Paar, Vladimir

    2013-01-01

    The main feature of global repeat map (GRM) algorithm (www.hazu.hr/grm/software/win/grm2012.exe) is its ability to identify a broad variety of repeats of unbounded length that can be arbitrarily distant in sequences as large as human chromosomes. The efficacy is due to the use of complete set of a K-string ensemble which enables a new method of direct mapping of symbolic DNA sequence into frequency domain, with straightforward identification of repeats as peaks in GRM diagram. In this way, we obtain very fast, efficient and highly automatized repeat finding tool. The method is robust to substitutions and insertions/deletions, as well as to various complexities of the sequence pattern. We present several case studies of GRM use, in order to illustrate its capabilities: identification of α-satellite tandem repeats and higher order repeats (HORs), identification of Alu dispersed repeats and of Alu tandems, identification of Period 3 pattern in exons, implementation of ‘magnifying glass’ effect, identification of complex HOR pattern, identification of inter-tandem transitional dispersed repeat sequences and identification of long segmental duplications. GRM algorithm is convenient for use, in particular, in cases of large repeat units, of highly mutated and/or complex repeats, and of global repeat maps for large genomic sequences (chromosomes and genomes). PMID:22977183

  14. Pure E and B polarization maps via Wiener filtering

    NASA Astrophysics Data System (ADS)

    Bunn, Emory F.; Wandelt, Benjamin

    2017-08-01

    In order to draw scientific conclusions from observations of cosmic microwave background (CMB) polarization, it is necessary to separate the contributions of the E and B components of the data. For data with incomplete sky coverage, there are ambiguous modes, which can be sourced by either E or B signals. Techniques exist for producing "pure" E and B maps, which are guaranteed to be free of cross-contamination, although the standard method, which involves constructing an eigenbasis, has a high computational cost. We show that such pure maps can be thought of as resulting from the application of a Wiener filter to the data. This perspective leads to far more efficient methods of producing pure maps. Moreover, by expressing the idea of purification in the general framework of Wiener filtering (i.e., maximization of a posterior probability), it leads to a variety of generalizations of the notion of pure E and B maps, e.g., accounting for noise or other contaminants in the data as well as correlations with temperature anisotropy.

  15. Unified View of Backward Backtracking in Short Read Mapping

    NASA Astrophysics Data System (ADS)

    Mäkinen, Veli; Välimäki, Niko; Laaksonen, Antti; Katainen, Riku

    Mapping short DNA reads to the reference genome is the core task in the recent high-throughput technologies to study e.g. protein-DNA interactions (ChIP-seq) and alternative splicing (RNA-seq). Several tools for the task (bowtie, bwa, SOAP2, TopHat) have been developed that exploit Burrows-Wheeler transform and the backward backtracking technique on it, to map the reads to their best approximate occurrences in the genome. These tools use different tailored mechanisms for small error-levels to prune the search phase significantly. We propose a new pruning mechanism that can be seen a generalization of the tailored mechanisms used so far. It uses a novel idea of storing all cyclic rotations of fixed length substrings of the reference sequence with a compressed index that is able to exploit the repetitions created to level out the growth of the input set. For RNA-seq we propose a new method that combines dynamic programming with backtracking to map efficiently and correctly all reads that span two exons. Same mechanism can also be used for mapping mate-pair reads.

  16. An integrated approach for automated cover-type mapping of large inaccessible areas in Alaska

    USGS Publications Warehouse

    Fleming, Michael D.

    1988-01-01

    The lack of any detailed cover type maps in the state necessitated that a rapid and accurate approach to be employed to develop maps for 329 million acres of Alaska within a seven-year period. This goal has been addressed by using an integrated approach to computer-aided analysis which combines efficient use of field data with the only consistent statewide spatial data sets available: Landsat multispectral scanner data, digital elevation data derived from 1:250 000-scale maps, and 1:60 000-scale color-infrared aerial photographs.

  17. YouGenMap: a web platform for dynamic multi-comparative mapping and visualization of genetic maps

    Treesearch

    Keith Batesole; Kokulapalan Wimalanathan; Lin Liu; Fan Zhang; Craig S. Echt; Chun Liang

    2014-01-01

    Comparative genetic maps are used in examination of genome organization, detection of conserved gene order, and exploration of marker order variations. YouGenMap is an open-source web tool that offers dynamic comparative mapping capability of users' own genetic mapping between 2 or more map sets. Users' genetic map data and optional gene annotations are...

  18. Open-Source Programming for Automated Generation of Graphene Raman Spectral Maps

    NASA Astrophysics Data System (ADS)

    Vendola, P.; Blades, M.; Pierre, W.; Jedlicka, S.; Rotkin, S. V.

    Raman microscopy is a useful tool for studying the structural characteristics of graphene deposited onto substrates. However, extracting useful information from the Raman spectra requires data processing and 2D map generation. An existing home-built confocal Raman microscope was optimized for graphene samples and programmed to automatically generate Raman spectral maps across a specified area. In particular, an open source data collection scheme was generated to allow the efficient collection and analysis of the Raman spectral data for future use. NSF ECCS-1509786.

  19. Improving IMRT delivery efficiency with reweighted L1-minimization for inverse planning

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kim, Hojin; Becker, Stephen; Lee, Rena

    2013-07-15

    Purpose: This study presents an improved technique to further simplify the fluence-map in intensity modulated radiation therapy (IMRT) inverse planning, thereby reducing plan complexity and improving delivery efficiency, while maintaining the plan quality.Methods: First-order total-variation (TV) minimization (min.) based on L1-norm has been proposed to reduce the complexity of fluence-map in IMRT by generating sparse fluence-map variations. However, with stronger dose sparing to the critical structures, the inevitable increase in the fluence-map complexity can lead to inefficient dose delivery. Theoretically, L0-min. is the ideal solution for the sparse signal recovery problem, yet practically intractable due to its nonconvexity of themore » objective function. As an alternative, the authors use the iteratively reweighted L1-min. technique to incorporate the benefits of the L0-norm into the tractability of L1-min. The weight multiplied to each element is inversely related to the magnitude of the corresponding element, which is iteratively updated by the reweighting process. The proposed penalizing process combined with TV min. further improves sparsity in the fluence-map variations, hence ultimately enhancing the delivery efficiency. To validate the proposed method, this work compares three treatment plans obtained from quadratic min. (generally used in clinic IMRT), conventional TV min., and our proposed reweighted TV min. techniques, implemented by a large-scale L1-solver (template for first-order conic solver), for five patient clinical data. Criteria such as conformation number (CN), modulation index (MI), and estimated treatment time are employed to assess the relationship between the plan quality and delivery efficiency.Results: The proposed method yields simpler fluence-maps than the quadratic and conventional TV based techniques. To attain a given CN and dose sparing to the critical organs for 5 clinical cases, the proposed method reduces the number of

  20. HapMap scanning of novel human minor histocompatibility antigens.

    PubMed

    Kamei, Michi; Nannya, Yasuhito; Torikai, Hiroki; Kawase, Takakazu; Taura, Kenjiro; Inamoto, Yoshihiro; Takahashi, Taro; Yazaki, Makoto; Morishima, Satoko; Tsujimura, Kunio; Miyamura, Koichi; Ito, Tetsuya; Togari, Hajime; Riddell, Stanley R; Kodera, Yoshihisa; Morishima, Yasuo; Takahashi, Toshitada; Kuzushima, Kiyotaka; Ogawa, Seishi; Akatsuka, Yoshiki

    2009-05-21

    Minor histocompatibility antigens (mHags) are molecular targets of allo-immunity associated with hematopoietic stem cell transplantation (HSCT) and involved in graft-versus-host disease, but they also have beneficial antitumor activity. mHags are typically defined by host SNPs that are not shared by the donor and are immunologically recognized by cytotoxic T cells isolated from post-HSCT patients. However, the number of molecularly identified mHags is still too small to allow prospective studies of their clinical importance in transplantation medicine, mostly due to the lack of an efficient method for isolation. Here we show that when combined with conventional immunologic assays, the large data set from the International HapMap Project can be directly used for genetic mapping of novel mHags. Based on the immunologically determined mHag status in HapMap panels, a target mHag locus can be uniquely mapped through whole genome association scanning taking advantage of the unprecedented resolution and power obtained with more than 3 000 000 markers. The feasibility of our approach could be supported by extensive simulations and further confirmed by actually isolating 2 novel mHags as well as 1 previously identified example. The HapMap data set represents an invaluable resource for investigating human variation, with obvious applications in genetic mapping of clinically relevant human traits.

  1. Long Read Alignment with Parallel MapReduce Cloud Platform

    PubMed Central

    Al-Absi, Ahmed Abdulhakim; Kang, Dae-Ki

    2015-01-01

    Genomic sequence alignment is an important technique to decode genome sequences in bioinformatics. Next-Generation Sequencing technologies produce genomic data of longer reads. Cloud platforms are adopted to address the problems arising from storage and analysis of large genomic data. Existing genes sequencing tools for cloud platforms predominantly consider short read gene sequences and adopt the Hadoop MapReduce framework for computation. However, serial execution of map and reduce phases is a problem in such systems. Therefore, in this paper, we introduce Burrows-Wheeler Aligner's Smith-Waterman Alignment on Parallel MapReduce (BWASW-PMR) cloud platform for long sequence alignment. The proposed cloud platform adopts a widely accepted and accurate BWA-SW algorithm for long sequence alignment. A custom MapReduce platform is developed to overcome the drawbacks of the Hadoop framework. A parallel execution strategy of the MapReduce phases and optimization of Smith-Waterman algorithm are considered. Performance evaluation results exhibit an average speed-up of 6.7 considering BWASW-PMR compared with the state-of-the-art Bwasw-Cloud. An average reduction of 30% in the map phase makespan is reported across all experiments comparing BWASW-PMR with Bwasw-Cloud. Optimization of Smith-Waterman results in reducing the execution time by 91.8%. The experimental study proves the efficiency of BWASW-PMR for aligning long genomic sequences on cloud platforms. PMID:26839887

  2. Long Read Alignment with Parallel MapReduce Cloud Platform.

    PubMed

    Al-Absi, Ahmed Abdulhakim; Kang, Dae-Ki

    2015-01-01

    Genomic sequence alignment is an important technique to decode genome sequences in bioinformatics. Next-Generation Sequencing technologies produce genomic data of longer reads. Cloud platforms are adopted to address the problems arising from storage and analysis of large genomic data. Existing genes sequencing tools for cloud platforms predominantly consider short read gene sequences and adopt the Hadoop MapReduce framework for computation. However, serial execution of map and reduce phases is a problem in such systems. Therefore, in this paper, we introduce Burrows-Wheeler Aligner's Smith-Waterman Alignment on Parallel MapReduce (BWASW-PMR) cloud platform for long sequence alignment. The proposed cloud platform adopts a widely accepted and accurate BWA-SW algorithm for long sequence alignment. A custom MapReduce platform is developed to overcome the drawbacks of the Hadoop framework. A parallel execution strategy of the MapReduce phases and optimization of Smith-Waterman algorithm are considered. Performance evaluation results exhibit an average speed-up of 6.7 considering BWASW-PMR compared with the state-of-the-art Bwasw-Cloud. An average reduction of 30% in the map phase makespan is reported across all experiments comparing BWASW-PMR with Bwasw-Cloud. Optimization of Smith-Waterman results in reducing the execution time by 91.8%. The experimental study proves the efficiency of BWASW-PMR for aligning long genomic sequences on cloud platforms.

  3. Mapping Children--Mapping Space.

    ERIC Educational Resources Information Center

    Pick, Herbert L., Jr.

    Research is underway concerning the way the perception, conception, and representation of spatial layout develops. Three concepts are important here--space itself, frame of reference, and cognitive map. Cognitive map refers to a form of representation of the behavioral space, not paired associate or serial response learning. Other criteria…

  4. Maps for the nation: The current federal mapping establishment

    USGS Publications Warehouse

    North, G.W.

    1983-01-01

    The U.S. Government annually produces an estimated 53,000 new maps and charts and distributes about 160 million copies. A large number of these maps are produced under the national mapping program, a decentralized Federal/State cooperative approach to mapping the country at standard scales. Circular A-16, issued by the Office of Management and Budget in 1953 and revised in 1967, delegates the mapping responsibilities to various federal agencies. The U.S. Department of the Interior's Geological Survey is the principal federal agency responsible for implementing the national mapping program. Other major federal map producing agencies include the Departments of Agriculture, Commerce, Defense, Housing and Urban Development, and Transportation, and the Tennessee Valley Authority. To make maps and mapping information more readily available, the National Cartographic Information Center was established in 1974 and an expanded National Map Library Depository Program in 1981. The most recent of many technological advances made under the mapping program are in the areas of digital cartography and video disc and optical disc information storage systems. Future trends and changes in the federal mapping program will involve expanded information and customer service operations, further developments in the production and use of digital cartographic data, and consideration of a Federal Mapping Agency. ?? 1983.

  5. NATIONAL CARTOGRAPHIC INFORMATION CENTER: AN INFORMATION RESOURCE ON MAPPING PRODUCTS FOR THE NATION.

    USGS Publications Warehouse

    Stevens, Alan R.

    1985-01-01

    Since its inception in 1974 the National Cartographic Information Center (NCIC), US Geological Survey, has rapidly developed to become a focal point for providing information on the availability of cartographic data, including maps/charts, aerial photographs, satellite imagery, geodetic control, digital mapping data, map materials and related cartographic products. In early years NCIC concentrated its efforts on encoding and entering several major National Mapping Division record collections into its systems. NCIC is now stressing the acquisition of data from sources outside the National Mapping Division, including 37 Federal agencies and more than a thousand State and private institutions. A critical review has recently been conducted by NCIC of its systems with the aim of improving its efficiency and levels of operation. Several activities which resulted include improving its existing networks, refinement of digital data distribution, study of new storage media and related projects.

  6. Mapping soil water content on golf course greens with GPR

    USDA-ARS?s Scientific Manuscript database

    Ground-penetrating radar (GPR) can be an effective and efficient method for high-resolution mapping of volumetric water content in the sand layer directly beneath the ground surface at a golf course green. This information could potentially be very useful to golf course superintendents for determi...

  7. Optimal mapping of irregular finite element domains to parallel processors

    NASA Technical Reports Server (NTRS)

    Flower, J.; Otto, S.; Salama, M.

    1987-01-01

    Mapping the solution domain of n-finite elements into N-subdomains that may be processed in parallel by N-processors is an optimal one if the subdomain decomposition results in a well-balanced workload distribution among the processors. The problem is discussed in the context of irregular finite element domains as an important aspect of the efficient utilization of the capabilities of emerging multiprocessor computers. Finding the optimal mapping is an intractable combinatorial optimization problem, for which a satisfactory approximate solution is obtained here by analogy to a method used in statistical mechanics for simulating the annealing process in solids. The simulated annealing analogy and algorithm are described, and numerical results are given for mapping an irregular two-dimensional finite element domain containing a singularity onto the Hypercube computer.

  8. Engineering management technologies of increasing energy efficiency processes in the investment and construction projects

    NASA Astrophysics Data System (ADS)

    Borisovich Zelentsov, Leonid; Dmitrievna Mailyan, Liya; Sultanovich Shogenov, Murat

    2017-10-01

    The article deals with the problems of using the energy-efficient materials and engineering technologies during the construction of buildings and structures. As the analysis showed, one of the most important problems in this sphere is the infringement of production technologies working with energy-efficient materials. To improve the given situation, it is offered to set a technological normal at the design stage by means of working out the technological maps studying the set and the succession of operations in details, taking in mind the properties of energy-efficient materials. At Don State Technical University (DSTU) the intelligent systems of management are being developed providing organizational and technological and also informational integration of design and production stages by means of creating the single database of technological maps, volumes of work and resources.

  9. Integrate and fire neural networks, piecewise contractive maps and limit cycles.

    PubMed

    Catsigeras, Eleonora; Guiraud, Pierre

    2013-09-01

    We study the global dynamics of integrate and fire neural networks composed of an arbitrary number of identical neurons interacting by inhibition and excitation. We prove that if the interactions are strong enough, then the support of the stable asymptotic dynamics consists of limit cycles. We also find sufficient conditions for the synchronization of networks containing excitatory neurons. The proofs are based on the analysis of the equivalent dynamics of a piecewise continuous Poincaré map associated to the system. We show that for efficient interactions the Poincaré map is piecewise contractive. Using this contraction property, we prove that there exist a countable number of limit cycles attracting all the orbits dropping into the stable subset of the phase space. This result applies not only to the Poincaré map under study, but also to a wide class of general n-dimensional piecewise contractive maps.

  10. MaMR: High-performance MapReduce programming model for material cloud applications

    NASA Astrophysics Data System (ADS)

    Jing, Weipeng; Tong, Danyu; Wang, Yangang; Wang, Jingyuan; Liu, Yaqiu; Zhao, Peng

    2017-02-01

    With the increasing data size in materials science, existing programming models no longer satisfy the application requirements. MapReduce is a programming model that enables the easy development of scalable parallel applications to process big data on cloud computing systems. However, this model does not directly support the processing of multiple related data, and the processing performance does not reflect the advantages of cloud computing. To enhance the capability of workflow applications in material data processing, we defined a programming model for material cloud applications that supports multiple different Map and Reduce functions running concurrently based on hybrid share-memory BSP called MaMR. An optimized data sharing strategy to supply the shared data to the different Map and Reduce stages was also designed. We added a new merge phase to MapReduce that can efficiently merge data from the map and reduce modules. Experiments showed that the model and framework present effective performance improvements compared to previous work.

  11. Story Map Instruction: A Road Map for Reading Comprehension.

    ERIC Educational Resources Information Center

    Davis, Zephaniah, T.; McPherson, Michael D.

    1989-01-01

    Introduces teachers to the development and use of story maps as a tool for promoting reading comprehension. Presents a definition and review of story map research. Explains how to construct story maps, and offers suggestions for starting story map instruction. Provides variations on the use of story maps. (MG)

  12. Phosphorylation of MAP65-1 by Arabidopsis Aurora Kinases Is Required for Efficient Cell Cycle Progression1[OPEN

    PubMed Central

    Weimer, Annika K.; Stoppin-Mellet, Virginie; Kosetsu, Ken; Cedeño, Cesyen; Jaquinod, Michel; Njo, Maria; De Milde, Liesbeth; Tompa, Peter; Inzé, Dirk; Beeckman, Tom; Vantard, Marylin

    2017-01-01

    Aurora kinases are key effectors of mitosis. Plant Auroras are functionally divided into two clades. The alpha Auroras (Aurora1 and Aurora2) associate with the spindle and the cell plate and are implicated in controlling formative divisions throughout plant development. The beta Aurora (Aurora3) localizes to centromeres and likely functions in chromosome separation. In contrast to the wealth of data available on the role of Aurora in other kingdoms, knowledge on their function in plants is merely emerging. This is exemplified by the fact that only histone H3 and the plant homolog of TPX2 have been identified as Aurora substrates in plants. Here we provide biochemical, genetic, and cell biological evidence that the microtubule-bundling protein MAP65-1—a member of the MAP65/Ase1/PRC1 protein family, implicated in central spindle formation and cytokinesis in animals, yeasts, and plants—is a genuine substrate of alpha Aurora kinases. MAP65-1 interacts with Aurora1 in vivo and is phosphorylated on two residues at its unfolded tail domain. Its overexpression and down-regulation antagonistically affect the alpha Aurora double mutant phenotypes. Phospho-mutant analysis shows that Aurora contributes to the microtubule bundling capacity of MAP65-1 in concert with other mitotic kinases. PMID:27879390

  13. Cost-Efficiencies in Online Learning. ASHE Higher Education Report, Volume 32, Number 1

    ERIC Educational Resources Information Center

    Meyer, Katrina A.

    2006-01-01

    This monograph is divided into five chapters. The first chapter provides a road map for understanding the review of studies on cost-efficiencies of online learning, including understanding why cost-efficiencies are so important to many higher education institutions and the framework used to categorize and discuss these studies. The second chapter…

  14. A study of mapping exogenous knowledge representations into CONFIG

    NASA Technical Reports Server (NTRS)

    Mayfield, Blayne E.

    1992-01-01

    representation was chosen as the focus of this study. A mapping from CC to CONFIG was developed. Due to differences between the two programs, however, the mapping transforms some of the CC knowledge to CONFIG as documentation rather than as knowledge in a form useful to computation. The study suggests that it may be worthwhile to pursue the mappings further. By implementing the mapping as a program, actual comparisons of computational efficiency and quality of results can be made between the QSIM and CONFIG programs. A secondary study may reveal that the results of the two programs augment one another, contradict one another, or differ only slightly. If the latter, the qualitative reasoning techniques may be compared in other areas, such as computational efficiency.

  15. Contour Mapping

    NASA Technical Reports Server (NTRS)

    1995-01-01

    In the early 1990s, the Ohio State University Center for Mapping, a NASA Center for the Commercial Development of Space (CCDS), developed a system for mobile mapping called the GPSVan. While driving, the users can map an area from the sophisticated mapping van equipped with satellite signal receivers, video cameras and computer systems for collecting and storing mapping data. George J. Igel and Company and the Ohio State University Center for Mapping advanced the technology for use in determining the contours of a construction site. The new system reduces the time required for mapping and staking, and can monitor the amount of soil moved.

  16. Preliminary Evaluation of MapReduce for High-Performance Climate Data Analysis

    NASA Technical Reports Server (NTRS)

    Duffy, Daniel Q.; Schnase, John L.; Thompson, John H.; Freeman, Shawn M.; Clune, Thomas L.

    2012-01-01

    MapReduce is an approach to high-performance analytics that may be useful to data intensive problems in climate research. It offers an analysis paradigm that uses clusters of computers and combines distributed storage of large data sets with parallel computation. We are particularly interested in the potential of MapReduce to speed up basic operations common to a wide range of analyses. In order to evaluate this potential, we are prototyping a series of canonical MapReduce operations over a test suite of observational and climate simulation datasets. Our initial focus has been on averaging operations over arbitrary spatial and temporal extents within Modern Era Retrospective- Analysis for Research and Applications (MERRA) data. Preliminary results suggest this approach can improve efficiencies within data intensive analytic workflows.

  17. Calculating Higher-Order Moments of Phylogenetic Stochastic Mapping Summaries in Linear Time

    PubMed Central

    Dhar, Amrit

    2017-01-01

    Abstract Stochastic mapping is a simulation-based method for probabilistically mapping substitution histories onto phylogenies according to continuous-time Markov models of evolution. This technique can be used to infer properties of the evolutionary process on the phylogeny and, unlike parsimony-based mapping, conditions on the observed data to randomly draw substitution mappings that do not necessarily require the minimum number of events on a tree. Most stochastic mapping applications simulate substitution mappings only to estimate the mean and/or variance of two commonly used mapping summaries: the number of particular types of substitutions (labeled substitution counts) and the time spent in a particular group of states (labeled dwelling times) on the tree. Fast, simulation-free algorithms for calculating the mean of stochastic mapping summaries exist. Importantly, these algorithms scale linearly in the number of tips/leaves of the phylogenetic tree. However, to our knowledge, no such algorithm exists for calculating higher-order moments of stochastic mapping summaries. We present one such simulation-free dynamic programming algorithm that calculates prior and posterior mapping variances and scales linearly in the number of phylogeny tips. Our procedure suggests a general framework that can be used to efficiently compute higher-order moments of stochastic mapping summaries without simulations. We demonstrate the usefulness of our algorithm by extending previously developed statistical tests for rate variation across sites and for detecting evolutionarily conserved regions in genomic sequences. PMID:28177780

  18. Survey of MapReduce frame operation in bioinformatics.

    PubMed

    Zou, Quan; Li, Xu-Bin; Jiang, Wen-Rui; Lin, Zi-Yu; Li, Gui-Lin; Chen, Ke

    2014-07-01

    Bioinformatics is challenged by the fact that traditional analysis tools have difficulty in processing large-scale data from high-throughput sequencing. The open source Apache Hadoop project, which adopts the MapReduce framework and a distributed file system, has recently given bioinformatics researchers an opportunity to achieve scalable, efficient and reliable computing performance on Linux clusters and on cloud computing services. In this article, we present MapReduce frame-based applications that can be employed in the next-generation sequencing and other biological domains. In addition, we discuss the challenges faced by this field as well as the future works on parallel computing in bioinformatics. © The Author 2013. Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  19. Design and application of star map simulation system for star sensors

    NASA Astrophysics Data System (ADS)

    Wu, Feng; Shen, Weimin; Zhu, Xifang; Chen, Yuheng; Xu, Qinquan

    2013-12-01

    Modern star sensors are powerful to measure attitude automatically which assure a perfect performance of spacecrafts. They achieve very accurate attitudes by applying algorithms to process star maps obtained by the star camera mounted on them. Therefore, star maps play an important role in designing star cameras and developing procession algorithms. Furthermore, star maps supply significant supports to exam the performance of star sensors completely before their launch. However, it is not always convenient to supply abundant star maps by taking pictures of the sky. Thus, star map simulation with the aid of computer attracts a lot of interests by virtue of its low price and good convenience. A method to simulate star maps by programming and extending the function of the optical design program ZEMAX is proposed. The star map simulation system is established. Firstly, based on analyzing the working procedures of star sensors to measure attitudes and the basic method to design optical system by ZEMAX, the principle of simulating star sensor imaging is given out in detail. The theory about adding false stars and noises, and outputting maps is discussed and the corresponding approaches are proposed. Then, by external programming, the star map simulation program is designed and produced. Its user interference and operation are introduced. Applications of star map simulation method in evaluating optical system, star image extraction algorithm and star identification algorithm, and calibrating system errors are presented completely. It was proved that the proposed simulation method provides magnificent supports to the study on star sensors, and improves the performance of star sensors efficiently.

  20. Cross-genome map based dissection of a nitrogen use efficiency ortho-metaQTL in bread wheat unravels concerted cereal genome evolution.

    PubMed

    Quraishi, Umar Masood; Abrouk, Michael; Murat, Florent; Pont, Caroline; Foucrier, Séverine; Desmaizieres, Gregory; Confolent, Carole; Rivière, Nathalie; Charmet, Gilles; Paux, Etienne; Murigneux, Alain; Guerreiro, Laurent; Lafarge, Stéphane; Le Gouis, Jacques; Feuillet, Catherine; Salse, Jerome

    2011-03-01

    Monitoring nitrogen use efficiency (NUE) in plants is becoming essential to maintain yield while reducing fertilizer usage. Optimized NUE application in major crops is essential for long-term sustainability of agriculture production. Here, we report the precise identification of 11 major chromosomal regions controlling NUE in wheat that co-localise with key developmental genes such as Ppd (photoperiod sensitivity), Vrn (vernalization requirement), Rht (reduced height) and can be considered as robust markers from a molecular breeding perspective. Physical mapping, sequencing, annotation and candidate gene validation of an NUE metaQTL on wheat chromosome 3B allowed us to propose that a glutamate synthase (GoGAT) gene that is conserved structurally and functionally at orthologous positions in rice, sorghum and maize genomes may contribute to NUE in wheat and other cereals. We propose an evolutionary model for the NUE locus in cereals from a common ancestral region, involving species specific shuffling events such as gene deletion, inversion, transposition and the invasion of repetitive elements. © 2011 The Authors. The Plant Journal © 2011 Blackwell Publishing Ltd.

  1. Semi-automated landform classification for hazard mapping of soil liquefaction by earthquake

    NASA Astrophysics Data System (ADS)

    Nakano, Takayuki

    2018-05-01

    Soil liquefaction damages were caused by huge earthquake in Japan, and the similar damages are concerned in near future huge earthquake. On the other hand, a preparation of soil liquefaction risk map (soil liquefaction hazard map) is impeded by the difficulty of evaluation of soil liquefaction risk. Generally, relative soil liquefaction risk should be able to be evaluated from landform classification data by using experimental rule based on the relationship between extent of soil liquefaction damage and landform classification items associated with past earthquake. Therefore, I rearranged the relationship between landform classification items and soil liquefaction risk intelligibly in order to enable the evaluation of soil liquefaction risk based on landform classification data appropriately and efficiently. And I developed a new method of generating landform classification data of 50-m grid size from existing landform classification data of 250-m grid size by using digital elevation model (DEM) data and multi-band satellite image data in order to evaluate soil liquefaction risk in detail spatially. It is expected that the products of this study contribute to efficient producing of soil liquefaction hazard map by local government.

  2. A journey from a SSR-based low density map to a SNP-based high density map for identification of disease resistance quantitative trait loci in peanut

    USDA-ARS?s Scientific Manuscript database

    Mapping and identification of quantitative trait loci (QTLs) are important for efficient marker-assisted breeding. Diseases such as leaf spots and Tomato spotted wilt virus (TSWV) cause significant loses to peanut growers. The U.S. Peanut Genome Initiative (PGI) was launched in 2004, and expanded to...

  3. Uncertainty Assessment and Weight Map Generation for Efficient Fusion of Tandem-X and CARTOSAT-1 Dems

    NASA Astrophysics Data System (ADS)

    Bagheri, H.; Schmitt, M.; Zhu, X. X.

    2017-05-01

    Recently, with InSAR data provided by the German TanDEM-X mission, a new global, high-resolution Digital Elevation Model (DEM) has been produced by the German Aerospace Center (DLR) with unprecedented height accuracy. However, due to SAR-inherent sensor specifics, its quality decreases over urban areas, making additional improvement necessary. On the other hand, DEMs derived from optical remote sensing imagery, such as Cartosat-1 data, have an apparently greater resolution in urban areas, making their fusion with TanDEM-X elevation data a promising perspective. The objective of this paper is two-fold: First, the height accuracies of TanDEM-X and Cartosat-1 elevation data over different land types are empirically evaluated in order to analyze the potential of TanDEM-XCartosat- 1 DEM data fusion. After the quality assessment, urban DEM fusion using weighted averaging is investigated. In this experiment, both weight maps derived from the height error maps delivered with the DEM data, as well as more sophisticated weight maps predicted by a procedure based on artificial neural networks (ANNs) are compared. The ANN framework employs several features that can describe the height residual performance to predict the weights used in the subsequent fusion step. The results demonstrate that especially the ANN-based framework is able to improve the quality of the final DEM through data fusion.

  4. 3D resolved mapping of optical aberrations in thick tissues

    PubMed Central

    Zeng, Jun; Mahou, Pierre; Schanne-Klein, Marie-Claire; Beaurepaire, Emmanuel; Débarre, Delphine

    2012-01-01

    We demonstrate a simple method for mapping optical aberrations with 3D resolution within thick samples. The method relies on the local measurement of the variation in image quality with externally applied aberrations. We discuss the accuracy of the method as a function of the signal strength and of the aberration amplitude and we derive the achievable resolution for the resulting measurements. We then report on measured 3D aberration maps in human skin biopsies and mouse brain slices. From these data, we analyse the consequences of tissue structure and refractive index distribution on aberrations and imaging depth in normal and cleared tissue samples. The aberration maps allow the estimation of the typical aplanetism region size over which aberrations can be uniformly corrected. This method and data pave the way towards efficient correction strategies for tissue imaging applications. PMID:22876353

  5. Hierarchical layered and semantic-based image segmentation using ergodicity map

    NASA Astrophysics Data System (ADS)

    Yadegar, Jacob; Liu, Xiaoqing

    2010-04-01

    Image segmentation plays a foundational role in image understanding and computer vision. Although great strides have been made and progress achieved on automatic/semi-automatic image segmentation algorithms, designing a generic, robust, and efficient image segmentation algorithm is still challenging. Human vision is still far superior compared to computer vision, especially in interpreting semantic meanings/objects in images. We present a hierarchical/layered semantic image segmentation algorithm that can automatically and efficiently segment images into hierarchical layered/multi-scaled semantic regions/objects with contextual topological relationships. The proposed algorithm bridges the gap between high-level semantics and low-level visual features/cues (such as color, intensity, edge, etc.) through utilizing a layered/hierarchical ergodicity map, where ergodicity is computed based on a space filling fractal concept and used as a region dissimilarity measurement. The algorithm applies a highly scalable, efficient, and adaptive Peano- Cesaro triangulation/tiling technique to decompose the given image into a set of similar/homogenous regions based on low-level visual cues in a top-down manner. The layered/hierarchical ergodicity map is built through a bottom-up region dissimilarity analysis. The recursive fractal sweep associated with the Peano-Cesaro triangulation provides efficient local multi-resolution refinement to any level of detail. The generated binary decomposition tree also provides efficient neighbor retrieval mechanisms for contextual topological object/region relationship generation. Experiments have been conducted within the maritime image environment where the segmented layered semantic objects include the basic level objects (i.e. sky/land/water) and deeper level objects in the sky/land/water surfaces. Experimental results demonstrate the proposed algorithm has the capability to robustly and efficiently segment images into layered semantic objects

  6. Mapping landslide susceptibility using data-driven methods.

    PubMed

    Zêzere, J L; Pereira, S; Melo, R; Oliveira, S C; Garcia, R A C

    2017-07-01

    Most epistemic uncertainty within data-driven landslide susceptibility assessment results from errors in landslide inventories, difficulty in identifying and mapping landslide causes and decisions related with the modelling procedure. In this work we evaluate and discuss differences observed on landslide susceptibility maps resulting from: (i) the selection of the statistical method; (ii) the selection of the terrain mapping unit; and (iii) the selection of the feature type to represent landslides in the model (polygon versus point). The work is performed in a single study area (Silveira Basin - 18.2km 2 - Lisbon Region, Portugal) using a unique database of geo-environmental landslide predisposing factors and an inventory of 82 shallow translational slides. The logistic regression, the discriminant analysis and two versions of the information value were used and we conclude that multivariate statistical methods perform better when computed over heterogeneous terrain units and should be selected to assess landslide susceptibility based on slope terrain units, geo-hydrological terrain units or census terrain units. However, evidence was found that the chosen terrain mapping unit can produce greater differences on final susceptibility results than those resulting from the chosen statistical method for modelling. The landslide susceptibility should be assessed over grid cell terrain units whenever the spatial accuracy of landslide inventory is good. In addition, a single point per landslide proved to be efficient to generate accurate landslide susceptibility maps, providing the landslides are of small size, thus minimizing the possible existence of heterogeneities of predisposing factors within the landslide boundary. Although during last years the ROC curves have been preferred to evaluate the susceptibility model's performance, evidence was found that the model with the highest AUC ROC is not necessarily the best landslide susceptibility model, namely when terrain

  7. Molecular mapping of chromosomes 17 and X

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Barker, D.F.

    1989-01-01

    The basic aims of this project are the construction of high density genetic maps of chromosomes 17 and X and the utilization of these maps for the subsequent isolation of a set of physically overlapping DNA segment clones. The strategy depends on the utilization of chromosome specific libraries of small (1--15 kb) segments from each of the two chromosomes. Since the time of submission of our previous progress report, we have refined the genetic map of markers which we had previously isolated for chromosome 17. We have completed our genetic mapping in CEPH reference and NF1 families of 15 markersmore » in the pericentric region of chromosome 17. Physical mapping results with three probes, were shown be in very close genetic proximity to the NF1 gene, with respect to two translocation breakpoints which disrupt the activity of the gene. All three of the probes were found to lie between the centromere and the most proximal translocation breakpoint, providing important genetic markers proximal to the NF1 gene. Our primary focus has shifted to the X chromosome. We have isolated an additional 30 polymorphic markers, bringing the total number we have isolated to over 80. We have invested substantial effort in characterizing the polymorphisms at each of these loci and constructed plasmid subclones which reveal the polymorphisms for nearly all of the loci. These subclones are of practical value in that they produce simpler and stronger patterns on human genomic Southern blots, thus improving the efficiency of the genetic mapping experiments. These subclones may also be of value for deriving DNA sequence information at each locus, necessary for establishing polymerase chain reaction primers specific for each locus. Such information would allow the use of each locus as a sequence tagged site.« less

  8. Digital mapping in extreme and remote environments

    NASA Astrophysics Data System (ADS)

    Andersson, Joel; Bauer, Tobias; Sarlus, Zimer; Zainy, Maher; Brethes, Anais

    2017-04-01

    During the last few years, Luleå University of Technology has performed a series of research projects in remote areas with extreme climatic conditions using digital mapping technologies. The majority of past and ongoing research projects focus on the arctic regions of the Fennoscandian Shield and Greenland but also on the Zagros fold-and-thrust belt in northern Iraq. Currently, we use the Midland Valley application FieldMove on iPad mini devices with ruggedized casings. As all projects have a strong focus on geological field work, harsh climatic conditions are a challenge not only for the geologists but also for the digital mapping hardware. In the arctic regions especially cold temperatures affect battery lifetime and performance of the screens. But also high temperatures are restricting digital mapping. From experience, a typical temperature range where digital mapping, using iPad tablets, is possible lies between -20 and +40 degrees. Furthermore, the remote character of field areas complicates access but also availability of electricity. By a combination of robust solar chargers and ruggedized batteries we are able to work entirely autarkical. Additionally, we are currently installing a drone system that allows us to map outcrops normally inaccessible because of safety reasons or time deficiency. The produced data will subsequently be taken into our Virtual Reality studio for interpretation and processing. There we will be able to work also with high resolution DEM data from Lidar scanning allowing us to interpret structural features such as post-glacial faults in areas that are otherwise only accessible by helicopter. By combining digital field mapping with drone technique and a Virtual Reality studio we are able to work in hardly accessible areas, improve safety during field work and increase efficiency substantially.

  9. Mapping Variables.

    ERIC Educational Resources Information Center

    Stone, Mark H.; Wright, Benjamin D.; Stenner, A. Jackson

    1999-01-01

    Describes mapping variables, the principal technique for planning and constructing a test or rating instrument. A variable map is also useful for interpreting results. Provides several maps to show the importance and value of mapping a variable by person and item data. (Author/SLD)

  10. USGS maps

    USGS Publications Warehouse

    ,

    2005-01-01

    Discover a small sample of the millions of maps produced by the U.S. Geological Survey (USGS) in its mission to map the Nation and survey its resources. This booklet gives a brief overview of the types of maps sold and distributed by the USGS through its Earth Science Information Centers (ESIC) and also available from business partners located in most States. The USGS provides a wide variety of maps, from topographic maps showing the geographic relief and thematic maps displaying the geology and water resources of the United States, to special studies of the moon and planets.

  11. A hybrid short read mapping accelerator

    PubMed Central

    2013-01-01

    Background The rapid growth of short read datasets poses a new challenge to the short read mapping problem in terms of sensitivity and execution speed. Existing methods often use a restrictive error model for computing the alignments to improve speed, whereas more flexible error models are generally too slow for large-scale applications. A number of short read mapping software tools have been proposed. However, designs based on hardware are relatively rare. Field programmable gate arrays (FPGAs) have been successfully used in a number of specific application areas, such as the DSP and communications domains due to their outstanding parallel data processing capabilities, making them a competitive platform to solve problems that are “inherently parallel”. Results We present a hybrid system for short read mapping utilizing both FPGA-based hardware and CPU-based software. The computation intensive alignment and the seed generation operations are mapped onto an FPGA. We present a computationally efficient, parallel block-wise alignment structure (Align Core) to approximate the conventional dynamic programming algorithm. The performance is compared to the multi-threaded CPU-based GASSST and BWA software implementations. For single-end alignment, our hybrid system achieves faster processing speed than GASSST (with a similar sensitivity) and BWA (with a higher sensitivity); for pair-end alignment, our design achieves a slightly worse sensitivity than that of BWA but has a higher processing speed. Conclusions This paper shows that our hybrid system can effectively accelerate the mapping of short reads to a reference genome based on the seed-and-extend approach. The performance comparison to the GASSST and BWA software implementations under different conditions shows that our hybrid design achieves a high degree of sensitivity and requires less overall execution time with only modest FPGA resource utilization. Our hybrid system design also shows that the performance

  12. Seafloor 2030 - Building a Global Ocean Map through International Collaboration

    NASA Astrophysics Data System (ADS)

    Ferrini, V. L.; Wigley, R. A.; Falconer, R. K. H.; Jakobsson, M.; Allen, G.; Mayer, L. A.; Schmitt, T.; Rovere, M.; Weatherall, P.; Marks, K. M.

    2016-12-01

    With more than 85% of the ocean floor unmapped, a huge proportion of our planet remains unexplored. Creating a comprehensive map of seafloor bathymetry remains a true global challenge that can only be accomplished through collaboration and partnership between governments, industry, academia, research organizations and non-government organizations. The objective of Seafloor 2030 is to comprehensively map the global ocean floor to resolutions that enable exploration and improved understanding of ocean processes, while informing maritime policy and supporting the management of natural marine resources for a sustainable Blue Economy. Seafloor 2030 is the outcome of the Forum for Future of Ocean Floor Mapping held in Monaco in June 2016, which was held under the auspices of GEBCO and the Nippon Foundation of Japan. GEBCO is the only international organization mandated to map the global ocean floor and is guided by the International Hydrographic Organization (IHO) and the Intergovernmental Oceanographic Commission of UNESCO. The task of completely mapping the ocean floor will require new global coordination to ensure that both existing data are identified and that new mapping efforts are coordinated to help efficiently "map the gaps." Fundamental to achieving Seafloor 2030 will be greater access to data, tools and technology, particularly for developing and coastal nations. This includes bathymetric post-processing and analysis software, database technology, computing infrastructure and gridding techniques as well as the latest developments in seafloor mapping methods and emerging crowd-sourced bathymetry initiatives. The key to achieving this global bathymetric map is capacity building and education - including greater coordination between scientific research and industry and the effective engagement of international organizations such as the United Nations.

  13. Efficiently Communicating Rich Heterogeneous Geospatial Data from the FeMO2008 Dive Cruise with FlashMap on EarthRef.org

    NASA Astrophysics Data System (ADS)

    Minnett, R. C.; Koppers, A. A.; Staudigel, D.; Staudigel, H.

    2008-12-01

    EarthRef.org is comprehensive and convenient resource for Earth Science reference data and models. It encompasses four main portals: the Geochemical Earth Reference Model (GERM), the Magnetics Information Consortium (MagIC), the Seamount Biogeosciences Network (SBN), and the Enduring Resources for Earth Science Education (ERESE). Their underlying databases are publically available and the scientific community has contributed widely and is urged to continue to do so. However, the net result is a vast and largely heterogeneous warehouse of geospatial data ranging from carefully prepared maps of seamounts to geochemical data/metadata, daily reports from seagoing expeditions, large volumes of raw and processed multibeam data, images of paleomagnetic sampling sites, etc. This presents a considerable obstacle for integrating other rich media content, such as videos, images, data files, cruise tracks, and interoperable database results, without overwhelming the web user. The four EarthRef.org portals clearly lend themselves to a more intuitive user interface and has, therefore, been an invaluable test bed for the design and implementation of FlashMap, a versatile KML-driven geospatial browser written for reliability and speed in Adobe Flash. FlashMap allows layers of content to be loaded and displayed over a streaming high-resolution map which can be zoomed and panned similarly to Google Maps and Google Earth. Many organizations, from National Geographic to the USGS, have begun using Google Earth software to display geospatial content. However, Google Earth, as a desktop application, does not integrate cleanly with existing websites requiring the user to navigate away from the browser and focus on a separate application and Google Maps, written in Java Script, does not scale up reliably to large datasets. FlashMap remedies these problems as a web-based application that allows for seamless integration of the real-time display power of Google Earth and the flexibility of

  14. Mapping a battlefield simulation onto message-passing parallel architectures

    NASA Technical Reports Server (NTRS)

    Nicol, David M.

    1987-01-01

    Perhaps the most critical problem in distributed simulation is that of mapping: without an effective mapping of workload to processors the speedup potential of parallel processing cannot be realized. Mapping a simulation onto a message-passing architecture is especially difficult when the computational workload dynamically changes as a function of time and space; this is exactly the situation faced by battlefield simulations. This paper studies an approach where the simulated battlefield domain is first partitioned into many regions of equal size; typically there are more regions than processors. The regions are then assigned to processors; a processor is responsible for performing all simulation activity associated with the regions. The assignment algorithm is quite simple and attempts to balance load by exploiting locality of workload intensity. The performance of this technique is studied on a simple battlefield simulation implemented on the Flex/32 multiprocessor. Measurements show that the proposed method achieves reasonable processor efficiencies. Furthermore, the method shows promise for use in dynamic remapping of the simulation.

  15. Evaluating and mapping sources and temporary storage areas of sediment

    Treesearch

    Leslie M. Reid

    1982-01-01

    Legislation to regulate forest practices, water quality, and management of federal lands has increased the land managers' need for efficient methods of identifying and mapping sources of sediment in forested basins. At the same time, theoretical analysis of landscape evolution has led research geomorphologists to the consideration of many of the same...

  16. TSPmap, a tool making use of traveling salesperson problem solvers in the efficient and accurate construction of high-density genetic linkage maps.

    PubMed

    Monroe, J Grey; Allen, Zachariah A; Tanger, Paul; Mullen, Jack L; Lovell, John T; Moyers, Brook T; Whitley, Darrell; McKay, John K

    2017-01-01

    Recent advances in nucleic acid sequencing technologies have led to a dramatic increase in the number of markers available to generate genetic linkage maps. This increased marker density can be used to improve genome assemblies as well as add much needed resolution for loci controlling variation in ecologically and agriculturally important traits. However, traditional genetic map construction methods from these large marker datasets can be computationally prohibitive and highly error prone. We present TSPmap , a method which implements both approximate and exact Traveling Salesperson Problem solvers to generate linkage maps. We demonstrate that for datasets with large numbers of genomic markers (e.g. 10,000) and in multiple population types generated from inbred parents, TSPmap can rapidly produce high quality linkage maps with low sensitivity to missing and erroneous genotyping data compared to two other benchmark methods, JoinMap and MSTmap . TSPmap is open source and freely available as an R package. With the advancement of low cost sequencing technologies, the number of markers used in the generation of genetic maps is expected to continue to rise. TSPmap will be a useful tool to handle such large datasets into the future, quickly producing high quality maps using a large number of genomic markers.

  17. New geospatial approaches for efficiently mapping forest biomass logistics at high resolution over large areas

    Treesearch

    John Hogland; Nathaniel Anderson; Woodam Chung

    2018-01-01

    Adequate biomass feedstock supply is an important factor in evaluating the financial feasibility of alternative site locations for bioenergy facilities and for maintaining profitability once a facility is built. We used newly developed spatial analysis and logistics software to model the variables influencing feedstock supply and to estimate and map two components of...

  18. Jane Pusch | NREL

    Science.gov Websites

    Pusch facilitates travel, meetings, and conferences for the Integrated Applications Center. Prior to , Business Administration, University of Rhode Island B. Ed., French and Mathematics, Queens University of

  19. Insect Flight: Computation and Biomimetic Design

    DTIC Science & Technology

    2008-05-31

    Mechanics, 37, 183-210 (2005). • Z. Jane Wang, ”Insect Flight”, McGraw Hill Year Book of Science and Technology, 2006. • Anders Andersen, Umberto Pesavento ...Umberto Pesavento , and Z. Jane Wang, ’Analysis of transitions between fluttering, tumbling and steady descent of falling cards’, Journal of Fluid

  20. Hartmann-Hahn 2D-map to optimize the RAMP-CPMAS NMR experiment for pharmaceutical materials.

    PubMed

    Suzuki, Kazuko; Martineau, Charlotte; Fink, Gerhard; Steuernagel, Stefan; Taulelle, Francis

    2012-02-01

    Cross polarization-magic angle spinning (CPMAS) is the most used experiment for solid-state NMR measurements in the pharmaceutical industry, with the well-known variant RAMP-CPMAS its dominant implementation. The experimental work presented in this contribution focuses on the entangled effects of the main parameters of such an experiment. The shape of the RAMP-CP pulse has been considered as well as the contact time duration, and a particular attention also has been devoted to the radio-frequency (RF) field inhomogeneity. (13)C CPMAS NMR spectra have been recorded with a systematic variation of (13)C and (1)H constant radiofrequency field pair values and represented as a Hartmann-Hahn matching two-dimensional map. Such a map yields a rational overview of the intricate optimal conditions necessary to achieve an efficient CP magnetization transfer. The map also highlights the effects of sweeping the RF by the RAMP-CP pulse on the number of Hartmann-Hahn matches crossed and how RF field inhomogeneity helps in increasing the CP efficiency by using a larger fraction of the sample. In the light of the results, strategies for optimal RAMP-CPMAS measurements are suggested, which lead to a much higher efficiency than constant amplitude CP experiment. Copyright © 2012 John Wiley & Sons, Ltd.

  1. Mapping of Coral Reef Environment in the Arabian Gulf Using Multispectral Remote Sensing

    NASA Astrophysics Data System (ADS)

    Ben-Romdhane, H.; Marpu, P. R.; Ghedira, H.; Ouarda, T. B. M. J.

    2016-06-01

    Coral reefs of the Arabian Gulf are subject to several pressures, thus requiring conservation actions. Well-designed conservation plans involve efficient mapping and monitoring systems. Satellite remote sensing is a cost-effective tool for seafloor mapping at large scales. Multispectral remote sensing of coastal habitats, like those of the Arabian Gulf, presents a special challenge due to their complexity and heterogeneity. The present study evaluates the potential of multispectral sensor DubaiSat-2 in mapping benthic communities of United Arab Emirates. We propose to use a spectral-spatial method that includes multilevel segmentation, nonlinear feature analysis and ensemble learning methods. Support Vector Machine (SVM) is used for comparison of classification performances. Comparative data were derived from the habitat maps published by the Environment Agency-Abu Dhabi. The spectral-spatial method produced 96.41% mapping accuracy. SVM classification is assessed to be 94.17% accurate. The adaptation of these methods can help achieving well-designed coastal management plans in the region.

  2. Integrated Georeferencing of Stereo Image Sequences Captured with a Stereovision Mobile Mapping System - Approaches and Practical Results

    NASA Astrophysics Data System (ADS)

    Eugster, H.; Huber, F.; Nebiker, S.; Gisi, A.

    2012-07-01

    Stereovision based mobile mapping systems enable the efficient capturing of directly georeferenced stereo pairs. With today's camera and onboard storage technologies imagery can be captured at high data rates resulting in dense stereo sequences. These georeferenced stereo sequences provide a highly detailed and accurate digital representation of the roadside environment which builds the foundation for a wide range of 3d mapping applications and image-based geo web-services. Georeferenced stereo images are ideally suited for the 3d mapping of street furniture and visible infrastructure objects, pavement inspection, asset management tasks or image based change detection. As in most mobile mapping systems, the georeferencing of the mapping sensors and observations - in our case of the imaging sensors - normally relies on direct georeferencing based on INS/GNSS navigation sensors. However, in urban canyons the achievable direct georeferencing accuracy of the dynamically captured stereo image sequences is often insufficient or at least degraded. Furthermore, many of the mentioned application scenarios require homogeneous georeferencing accuracy within a local reference frame over the entire mapping perimeter. To achieve these demands georeferencing approaches are presented and cost efficient workflows are discussed which allows validating and updating the INS/GNSS based trajectory with independently estimated positions in cases of prolonged GNSS signal outages in order to increase the georeferencing accuracy up to the project requirements.

  3. Probabilistic, Seismically-Induced Landslide Hazard Mapping of Western Oregon

    NASA Astrophysics Data System (ADS)

    Olsen, M. J.; Sharifi Mood, M.; Gillins, D. T.; Mahalingam, R.

    2015-12-01

    Earthquake-induced landslides can generate significant damage within urban communities by damaging structures, obstructing lifeline connection routes and utilities, generating various environmental impacts, and possibly resulting in loss of life. Reliable hazard and risk maps are important to assist agencies in efficiently allocating and managing limited resources to prepare for such events. This research presents a new methodology in order to communicate site-specific landslide hazard assessments in a large-scale, regional map. Implementation of the proposed methodology results in seismic-induced landslide hazard maps that depict the probabilities of exceeding landslide displacement thresholds (e.g. 0.1, 0.3, 1.0 and 10 meters). These maps integrate a variety of data sources including: recent landslide inventories, LIDAR and photogrammetric topographic data, geology map, mapped NEHRP site classifications based on available shear wave velocity data in each geologic unit, and USGS probabilistic seismic hazard curves. Soil strength estimates were obtained by evaluating slopes present along landslide scarps and deposits for major geologic units. Code was then developed to integrate these layers to perform a rigid, sliding block analysis to determine the amount and associated probabilities of displacement based on each bin of peak ground acceleration in the seismic hazard curve at each pixel. The methodology was applied to western Oregon, which contains weak, weathered, and often wet soils at steep slopes. Such conditions have a high landslide hazard even without seismic events. A series of landslide hazard maps highlighting the probabilities of exceeding the aforementioned thresholds were generated for the study area. These output maps were then utilized in a performance based design framework enabling them to be analyzed in conjunction with other hazards for fully probabilistic-based hazard evaluation and risk assessment. a) School of Civil and Construction

  4. Development of an Efficient Protein Extraction Method Compatible with LC-MS/MS for Proteome Mapping in Two Australian Seagrasses Zostera muelleri and Posidonia australis.

    PubMed

    Jiang, Zhijian; Kumar, Manoj; Padula, Matthew P; Pernice, Mathieu; Kahlke, Tim; Kim, Mikael; Ralph, Peter J

    2017-01-01

    The availability of the first complete genome sequence of the marine flowering plant Zostera marina (commonly known as seagrass) in early 2016, is expected to significantly raise the impact of seagrass proteomics. Seagrasses are marine ecosystem engineers that are currently declining worldwide at an alarming rate due to both natural and anthropogenic disturbances. Seagrasses (especially species of the genus Zostera ) are compromised for proteomic studies primarily due to the lack of efficient protein extraction methods because of their recalcitrant cell wall which is rich in complex polysaccharides and a high abundance of secondary metabolites in their cells. In the present study, three protein extraction methods that are commonly used in plant proteomics i.e., phenol (P); trichloroacetic acid/acetone/SDS/phenol (TASP); and borax/polyvinyl-polypyrrolidone/phenol (BPP) extraction, were evaluated quantitatively and qualitatively based on two dimensional isoelectric focusing (2D-IEF) maps and LC-MS/MS analysis using the two most abundant Australian seagrass species, namely Zostera muelleri and Posidonia australis . All three tested methods produced high quality protein extracts with excellent 2D-IEF maps in P. australis . However, the BPP method produces better results in Z. muelleri compared to TASP and P. Therefore, we further modified the BPP method (M-BPP) by homogenizing the tissue in a modified protein extraction buffer containing both ionic and non-ionic detergents (0.5% SDS; 1.5% Triton X-100), 2% PVPP and protease inhibitors. Further, the extracted proteins were solubilized in 0.5% of zwitterionic detergent (C7BzO) instead of 4% CHAPS. This slight modification to the BPP method resulted in a higher protein yield, and good quality 2-DE maps with a higher number of protein spots in both the tested seagrasses. Further, the M-BPP method was successfully utilized in western-blot analysis of phosphoenolpyruvate carboxylase (PEPC-a key enzyme for carbon metabolism

  5. Using high-resolution digital aerial imagery to map land cover

    USGS Publications Warehouse

    Dieck, J.J.; Robinson, Larry

    2014-01-01

    The Upper Midwest Environmental Sciences Center (UMESC) has used aerial photography to map land cover/land use on federally owned and managed lands for over 20 years. Until recently, that process used 23- by 23-centimeter (9- by 9-inch) analog aerial photos to classify vegetation along the Upper Mississippi River System, on National Wildlife Refuges, and in National Parks. With digital aerial cameras becoming more common and offering distinct advantages over analog film, UMESC transitioned to an entirely digital mapping process in 2009. Though not without challenges, this method has proven to be much more accurate and efficient when compared to the analog process.

  6. Hydrochromic Approaches to Mapping Human Sweat Pores.

    PubMed

    Park, Dong-Hoon; Park, Bum Jun; Kim, Jong-Man

    2016-06-21

    colorimetric change near body temperature. This feature enables the use of this technique to generate high-quality images of sweat pores. This Account also focuses on the results of the most recent phase of this investigation, which led to the development of a simple yet efficient and reliable technique for sweat pore mapping. The method utilizes a hydrophilic polymer composite film containing fluorescein, a commercially available dye that undergoes a fluorometric response as a result of water-dependent interconversion between its ring-closed spirolactone (nonfluorescent) and ring-opened fluorone (fluorescent) forms. Surface-modified carbon nanodots (CDs) have also been found to be efficient for hydrochromic mapping of human sweat pores. The results discovered by Lou et al. [ Adv. Mater. 2015 , 27 , 1389 ] are also included in this Account. Sweat pore maps obtained from fingertips using these materials were found to be useful for fingerprint analysis. In addition, this hydrochromism-based approach is sufficiently sensitive to enable differentiation between sweat-secreting active pores and inactive pores. As a result, the techniques can be applied to clinical diagnosis of malfunctioning sweat pores. The directions that future research in this area will follow are also discussed.

  7. Efficiency-wage competition and nonlinear dynamics

    NASA Astrophysics Data System (ADS)

    Guerrazzi, Marco; Sodini, Mauro

    2018-05-01

    In this paper we develop a nonlinear version of the efficiency-wage competition model pioneered by Hahn (1987) [27]. Under the assumption that the strategic relationship among optimal wage bids put forward by competing firms is non-monotonic, we show that market wage offers can actually display persistent fluctuations described by a piece-wise non-invertible map. Thereafter, assuming that employers are never constrained in the labour market, we give evidence that in the parameter region of chaotic dynamics, the model is able to reproduce the business cycle regularity according to which in the short-run average wages fluctuate less than aggregate employment. In addition, we show that the efficiency-wage competition among firms leads to some inefficiencies in the wage setting process.

  8. a Hadoop-Based Distributed Framework for Efficient Managing and Processing Big Remote Sensing Images

    NASA Astrophysics Data System (ADS)

    Wang, C.; Hu, F.; Hu, X.; Zhao, S.; Wen, W.; Yang, C.

    2015-07-01

    Various sensors from airborne and satellite platforms are producing large volumes of remote sensing images for mapping, environmental monitoring, disaster management, military intelligence, and others. However, it is challenging to efficiently storage, query and process such big data due to the data- and computing- intensive issues. In this paper, a Hadoop-based framework is proposed to manage and process the big remote sensing data in a distributed and parallel manner. Especially, remote sensing data can be directly fetched from other data platforms into the Hadoop Distributed File System (HDFS). The Orfeo toolbox, a ready-to-use tool for large image processing, is integrated into MapReduce to provide affluent image processing operations. With the integration of HDFS, Orfeo toolbox and MapReduce, these remote sensing images can be directly processed in parallel in a scalable computing environment. The experiment results show that the proposed framework can efficiently manage and process such big remote sensing data.

  9. Detection And Mapping (DAM) package. Volume 4A: Software System Manual, part 1

    NASA Technical Reports Server (NTRS)

    Schlosser, E. H.

    1980-01-01

    The package is an integrated set of manual procedures, computer programs, and graphic devices designed for efficient production of precisely registered and formatted maps from digital LANDSAT multispectral scanner (MSS) data. The software can be readily implemented on any Univac 1100 series computer with standard peripheral equipment. This version of the software includes predefined spectral limits for use in classifying and mapping surface water for LANDSAT-1, LANDSAT-2, and LANDSAT-3. Tape formats supported include X, AM, and PM.

  10. Sonochemical cleaning efficiencies in dental instruments

    NASA Astrophysics Data System (ADS)

    Tiong, T. Joyce; Walmsley, A. Damien; Price, Gareth J.

    2012-05-01

    Ultrasound has been widely used for cleaning purposes in a variety of situations, including in dental practice. Cleaning is achieved through a combination of acoustically driven streaming effects and sonochemical effects arising from the production of inertial cavitation in a liquid. In our work, various dental instruments used for endodontic (root canal) treatment have been evaluated for their efficiency in producing sonochemical effects in an in-vitro cleaning environment. The areas where cavitation was produced were mapped by monitoring chemiluminescence from luminol solutions and this was correlated with their cleaning efficiencies - assessed by the ability to bleach a dye, to form an emulsion by mixing immiscible components and also to remove ink from a glass surface. The results showed good correlation (Pearson's coefficient > 0.9) between the cavitation and cleaning efficiencies, suggesting that the former plays an important role in cleaning. The methods developed and the results will be beneficial in endodontics research in order to optimise future root canal instruments and treatments.

  11. Molecular mapping of chromosomes 17 and X. Progress report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Barker, D.F.

    1991-01-15

    Progress toward the construction of high density genetic maps of chromosomes 17 and X has been made by isolating and characterizing a relatively large set of polymorphic probes for each chromosome and using these probes to construct genetic maps. We have mapped the same polymorphic probes against a series of chromosome breakpoints on X and 17. The probes could be assigned to over 30 physical intervals on the X chromosome and 7 intervals on 17. In many cases, this process resulted in improved characterization of the relative locations of the breakpoints with respect to each other and the definition ofmore » new physical intervals. The strategy for isolation of the polymorphic clones utilized chromosome specific libraries of 1--15 kb segments from each of the two chromosomes. From these libraries, clones were screened for those detecting restriction fragment length polymorphisms. The markers were further characterized, the chromosomal assignments confirmed and in most cases segments of the original probes were subcloned into plasmids to produce probes with improved signal to noise ratios for use in the genetic marker studies. The linkage studies utilize the CEPH reference families and other well-characterized families in our collection which have been used for genetic disease linkage work. Preliminary maps and maps of portions of specific regions of 17 and X are provided. We have nearly completed a map of the 1 megabase Mycoplasma arthritidis genome by applying these techniques to a lambda phage library of its genome. We have found bit mapping to be an efficient means to organize a contiguous set of overlapping@ clones from a larger genome.« less

  12. Mapping of information and identification of construction waste at project life cycle

    NASA Astrophysics Data System (ADS)

    Wibowo, Mochamad Agung; Handayani, Naniek Utami; Nurdiana, Asri; Sholeh, Moh Nur; Pamungkas, Gita Silvia

    2018-03-01

    The development of construction project towards green construction is needed in order to improve the efficiency of construction projects. One that needs to be minimized is construction waste. Construction waste is waste generated from construction project activities, both solid waste and non solid waste. More specifically, the waste happens at every phase of the project life cycle. Project life cycle are the stage of idea, design, construction, and operation/maintenance. Each phase is managed by different stakeholders. Therefore it requires special handling from the involved stakeholders. The objective of the study is to map the information and identify the waste at each phase of the project life cycle. The purpose of mapping is to figure out the process of information and product flow and with its timeline. This mapping used Value Stream Mapping (VSM). Identification of waste was done by distributing questionnaire to respondents to know the waste according to owner, consultant planner, contractor, and supervisory consultant. The result of the study is the mapping of information flow and product flow at the phases of idea, design, construction, and operation/ maintenance.

  13. Rain use efficiency across a precipitation gradient on the Tibetan Plateau

    USDA-ARS?s Scientific Manuscript database

    Rain use efficiency (RUE), commonly described as the ratio of aboveground net primary production (ANPP) to mean annual precipitation (MAP), is a critical indicator for predicting potential responses of grassland ecosystems to changing precipitation regimes. However, current understanding on patterns...

  14. Efficient global fiber tracking on multi-dimensional diffusion direction maps

    NASA Astrophysics Data System (ADS)

    Klein, Jan; Köhler, Benjamin; Hahn, Horst K.

    2012-02-01

    Global fiber tracking algorithms have recently been proposed which were able to compute results of unprecedented quality. They account for avoiding accumulation errors by a global optimization process at the cost of a high computation time of several hours or even days. In this paper, we introduce a novel global fiber tracking algorithm which, for the first time, globally optimizes the underlying diffusion direction map obtained from DTI or HARDI data, instead of single fiber segments. As a consequence, the number of iterations in the optimization process can drastically be reduced by about three orders of magnitude. Furthermore, in contrast to all previous algorithms, the density of the tracked fibers can be adjusted after the optimization within a few seconds. We evaluated our method for diffusion-weighted images obtained from software phantoms, healthy volunteers, and tumor patients. We show that difficult fiber bundles, e.g., the visual pathways or tracts for different motor functions can be determined and separated in an excellent quality. Furthermore, crossing and kissing bundles are correctly resolved. On current standard hardware, a dense fiber tracking result of a whole brain can be determined in less than half an hour which is a strong improvement compared to previous work.

  15. Shifted Transversal Design smart-pooling for high coverage interactome mapping

    PubMed Central

    Xin, Xiaofeng; Rual, Jean-François; Hirozane-Kishikawa, Tomoko; Hill, David E.; Vidal, Marc; Boone, Charles; Thierry-Mieg, Nicolas

    2009-01-01

    “Smart-pooling,” in which test reagents are multiplexed in a highly redundant manner, is a promising strategy for achieving high efficiency, sensitivity, and specificity in systems-level projects. However, previous applications relied on low redundancy designs that do not leverage the full potential of smart-pooling, and more powerful theoretical constructions, such as the Shifted Transversal Design (STD), lack experimental validation. Here we evaluate STD smart-pooling in yeast two-hybrid (Y2H) interactome mapping. We employed two STD designs and two established methods to perform ORFeome-wide Y2H screens with 12 baits. We found that STD pooling achieves similar levels of sensitivity and specificity as one-on-one array-based Y2H, while the costs and workloads are divided by three. The screening-sequencing approach is the most cost- and labor-efficient, yet STD identifies about twofold more interactions. Screening-sequencing remains an appropriate method for quickly producing low-coverage interactomes, while STD pooling appears as the method of choice for obtaining maps with higher coverage. PMID:19447967

  16. Interpersonal psychotherapy for the prevention of excess weight gain and eating disorders: A brief case study.

    PubMed

    Tanofsky-Kraff, Marian; Shomaker, Lauren B; Young, Jami F; Wilfley, Denise E

    2016-06-01

    This article presents a brief case study of "Jane Doe," a 13-year-old, non-Hispanic White girl 2 participating in a clinical research trial of interpersonal psychotherapy-weight gain (IPT-WG). Girls at-risk for adult obesity and binge eating disorder (BED) were randomly assigned to take part in 12 weeks of preventative group treatment. Jane's IPT-WG group included five other early adolescent girls (mostly aged 12-13) at risk for adult obesity and BED. The case of Jane illustrates a successful example of IPT-WG for the prevention of excessive weight gain and for the prevention of BED. (PsycINFO Database Record (c) 2016 APA, all rights reserved).

  17. A limit-cycle self-organizing map architecture for stable arm control.

    PubMed

    Huang, Di-Wei; Gentili, Rodolphe J; Katz, Garrett E; Reggia, James A

    2017-01-01

    Inspired by the oscillatory nature of cerebral cortex activity, we recently proposed and studied self-organizing maps (SOMs) based on limit cycle neural activity in an attempt to improve the information efficiency and robustness of conventional single-node, single-pattern representations. Here we explore for the first time the use of limit cycle SOMs to build a neural architecture that controls a robotic arm by solving inverse kinematics in reach-and-hold tasks. This multi-map architecture integrates open-loop and closed-loop controls that learn to self-organize oscillatory neural representations and to harness non-fixed-point neural activity even for fixed-point arm reaching tasks. We show through computer simulations that our architecture generalizes well, achieves accurate, fast, and smooth arm movements, and is robust in the face of arm perturbations, map damage, and variations of internal timing parameters controlling the flow of activity. A robotic implementation is evaluated successfully without further training, demonstrating for the first time that limit cycle maps can control a physical robot arm. We conclude that architectures based on limit cycle maps can be organized to function effectively as neural controllers. Copyright © 2016 Elsevier Ltd. All rights reserved.

  18. OPTIMA: sensitive and accurate whole-genome alignment of error-prone genomic maps by combinatorial indexing and technology-agnostic statistical analysis.

    PubMed

    Verzotto, Davide; M Teo, Audrey S; Hillmer, Axel M; Nagarajan, Niranjan

    2016-01-01

    Resolution of complex repeat structures and rearrangements in the assembly and analysis of large eukaryotic genomes is often aided by a combination of high-throughput sequencing and genome-mapping technologies (for example, optical restriction mapping). In particular, mapping technologies can generate sparse maps of large DNA fragments (150 kilo base pairs (kbp) to 2 Mbp) and thus provide a unique source of information for disambiguating complex rearrangements in cancer genomes. Despite their utility, combining high-throughput sequencing and mapping technologies has been challenging because of the lack of efficient and sensitive map-alignment algorithms for robustly aligning error-prone maps to sequences. We introduce a novel seed-and-extend glocal (short for global-local) alignment method, OPTIMA (and a sliding-window extension for overlap alignment, OPTIMA-Overlap), which is the first to create indexes for continuous-valued mapping data while accounting for mapping errors. We also present a novel statistical model, agnostic with respect to technology-dependent error rates, for conservatively evaluating the significance of alignments without relying on expensive permutation-based tests. We show that OPTIMA and OPTIMA-Overlap outperform other state-of-the-art approaches (1.6-2 times more sensitive) and are more efficient (170-200 %) and precise in their alignments (nearly 99 % precision). These advantages are independent of the quality of the data, suggesting that our indexing approach and statistical evaluation are robust, provide improved sensitivity and guarantee high precision.

  19. Use of habitats as surrogates of biodiversity for efficient coral reef conservation planning in Pacific Ocean islands.

    PubMed

    Dalleau, Mayeul; Andréfouët, Serge; Wabnitz, Colette C C; Payri, Claude; Wantiez, Laurent; Pichon, Michel; Friedman, Kim; Vigliola, Laurent; Benzoni, Francesca

    2010-04-01

    Marine protected areas (MPAs) have been highlighted as a means toward effective conservation of coral reefs. New strategies are required to more effectively select MPA locations and increase the pace of their implementation. Many criteria exist to design MPA networks, but generally, it is recommended that networks conserve a diversity of species selected for, among other attributes, their representativeness, rarity, or endemicity. Because knowledge of species' spatial distribution remains scarce, efficient surrogates are urgently needed. We used five different levels of habitat maps and six spatial scales of analysis to identify under which circumstances habitat data used to design MPA networks for Wallis Island provided better representation of species than random choice alone. Protected-area site selections were derived from a rarity-complementarity algorithm. Habitat surrogacy was tested for commercial fish species, all fish species, commercially harvested invertebrates, corals, and algae species. Efficiency of habitat surrogacy varied by species group, type of habitat map, and spatial scale of analysis. Maps with the highest habitat thematic complexity provided better surrogates than simpler maps and were more robust to changes in spatial scales. Surrogates were most efficient for commercial fishes, corals, and algae but not for commercial invertebrates. Conversely, other measurements of species-habitat associations, such as richness congruence and composition similarities provided weak results. We provide, in part, a habitat-mapping methodology for designation of MPAs for Pacific Ocean islands that are characterized by habitat zonations similar to Wallis. Given the increasing availability and affordability of space-borne imagery to map habitats, our approach could appreciably facilitate and improve current approaches to coral reef conservation and enhance MPA implementation.

  20. Robust Nonrigid Multimodal Image Registration using Local Frequency Maps*

    PubMed Central

    Jian, Bing; Vemuri, Baba C.; Marroquin, José L.

    2008-01-01

    Automatic multi-modal image registration is central to numerous tasks in medical imaging today and has a vast range of applications e.g., image guidance, atlas construction, etc. In this paper, we present a novel multi-modal 3D non-rigid registration algorithm where in 3D images to be registered are represented by their corresponding local frequency maps efficiently computed using the Riesz transform as opposed to the popularly used Gabor filters. The non-rigid registration between these local frequency maps is formulated in a statistically robust framework involving the minimization of the integral squared error a.k.a. L2E (L2 error). This error is expressed as the squared difference between the true density of the residual (which is the squared difference between the non-rigidly transformed reference and the target local frequency representations) and a Gaussian or mixture of Gaussians density approximation of the same. The non-rigid transformation is expressed in a B-spline basis to achieve the desired smoothness in the transformation as well as computational efficiency. The key contributions of this work are (i) the use of Riesz transform to achieve better efficiency in computing the local frequency representation in comparison to Gabor filter-based approaches, (ii) new mathematical model for local-frequency based non-rigid registration, (iii) analytic computation of the gradient of the robust non-rigid registration cost function to achieve efficient and accurate registration. The proposed non-rigid L2E-based registration is a significant extension of research reported in literature to date. We present experimental results for registering several real data sets with synthetic and real non-rigid misalignments. PMID:17354721

  1. Identification of Candidate Genes Underlying an Iron Efficiency Quantitative Trait Locus in Soybean1

    PubMed Central

    Peiffer, Gregory A.; King, Keith E.; Severin, Andrew J.; May, Gregory D.; Cianzio, Silvia R.; Lin, Shun Fu; Lauter, Nicholas C.; Shoemaker, Randy C.

    2012-01-01

    Prevalent on calcareous soils in the United States and abroad, iron deficiency is among the most common and severe nutritional stresses in plants. In soybean (Glycine max) commercial plantings, the identification and use of iron-efficient genotypes has proven to be the best form of managing this soil-related plant stress. Previous studies conducted in soybean identified a significant iron efficiency quantitative trait locus (QTL) explaining more than 70% of the phenotypic variation for the trait. In this research, we identified candidate genes underlying this QTL through molecular breeding, mapping, and transcriptome sequencing. Introgression mapping was performed using two related near-isogenic lines in which a region located on soybean chromosome 3 required for iron efficiency was identified. The region corresponds to the previously reported iron efficiency QTL. The location was further confirmed through QTL mapping conducted in this study. Transcriptome sequencing and quantitative real-time-polymerase chain reaction identified two genes encoding transcription factors within the region that were significantly induced in soybean roots under iron stress. The two induced transcription factors were identified as homologs of the subgroup lb basic helix-loop-helix (bHLH) genes that are known to regulate the strategy I response in Arabidopsis (Arabidopsis thaliana). Resequencing of these differentially expressed genes unveiled a significant deletion within a predicted dimerization domain. We hypothesize that this deletion disrupts the Fe-DEFICIENCY-INDUCED TRANSCRIPTION FACTOR (FIT)/bHLH heterodimer that has been shown to induce known iron acquisition genes. PMID:22319075

  2. Mapping Partners Master Drug Dictionary to RxNorm using an NLP-based approach.

    PubMed

    Zhou, Li; Plasek, Joseph M; Mahoney, Lisa M; Chang, Frank Y; DiMaggio, Dana; Rocha, Roberto A

    2012-08-01

    To develop an automated method based on natural language processing (NLP) to facilitate the creation and maintenance of a mapping between RxNorm and a local medication terminology for interoperability and meaningful use purposes. We mapped 5961 terms from Partners Master Drug Dictionary (MDD) and 99 of the top prescribed medications to RxNorm. The mapping was conducted at both term and concept levels using an NLP tool, called MTERMS, followed by a manual review conducted by domain experts who created a gold standard mapping. The gold standard was used to assess the overall mapping between MDD and RxNorm and evaluate the performance of MTERMS. Overall, 74.7% of MDD terms and 82.8% of the top 99 terms had an exact semantic match to RxNorm. Compared to the gold standard, MTERMS achieved a precision of 99.8% and a recall of 73.9% when mapping all MDD terms, and a precision of 100% and a recall of 72.6% when mapping the top prescribed medications. The challenges and gaps in mapping MDD to RxNorm are mainly due to unique user or application requirements for representing drug concepts and the different modeling approaches inherent in the two terminologies. An automated approach based on NLP followed by human expert review is an efficient and feasible way for conducting dynamic mapping. Copyright © 2011 Elsevier Inc. All rights reserved.

  3. Efficient Kinematic Computations For 7-DOF Manipulators

    NASA Technical Reports Server (NTRS)

    Seraji, Homayoun; Long, Mark K.; Kreutz-Delgado, Kenneth

    1994-01-01

    Efficient algorithms for forward kinematic mappings of seven-degree-of-freedom (7-DOF) robotic manipulator having revolute joints developed on basis of representation of redundant DOF in terms of parameter called "arm angle." Continuing effort to exploit redundancy in manipulator according to concept of basic and additional tasks. Concept also discussed in "Configuration-Control Scheme Copes With Singularities" (NPO-18556) and "Increasing the Dexterity of Redundant Robots" (NPO-17801).

  4. Canopy Density Mapping on Ultracam-D Aerial Imagery in Zagros Woodlands, Iran

    NASA Astrophysics Data System (ADS)

    Erfanifard, Y.; Khodaee, Z.

    2013-09-01

    Canopy density maps express different characteristics of forest stands, especially in woodlands. Obtaining such maps by field measurements is so expensive and time-consuming. It seems necessary to find suitable techniques to produce these maps to be used in sustainable management of woodland ecosystems. In this research, a robust procedure was suggested to obtain these maps by very high spatial resolution aerial imagery. It was aimed to produce canopy density maps by UltraCam-D aerial imagery, newly taken in Zagros woodlands by Iran National Geographic Organization (NGO), in this study. A 30 ha plot of Persian oak (Quercus persica) coppice trees was selected in Zagros woodlands, Iran. The very high spatial resolution aerial imagery of the plot purchased from NGO, was classified by kNN technique and the tree crowns were extracted precisely. The canopy density was determined in each cell of different meshes with different sizes overlaid on the study area map. The accuracy of the final maps was investigated by the ground truth obtained by complete field measurements. The results showed that the proposed method of obtaining canopy density maps was efficient enough in the study area. The final canopy density map obtained by a mesh with 30 Ar (3000 m2) cell size had 80% overall accuracy and 0.61 KHAT coefficient of agreement which shows a great agreement with the observed samples. This method can also be tested in other case studies to reveal its capability in canopy density map production in woodlands.

  5. Use of LANDSAT-1 data for the detection and mapping of saline seeps in Montana

    NASA Technical Reports Server (NTRS)

    May, G. A. (Principal Investigator); Petersen, G. W.

    1976-01-01

    The author has identified the following significant results. April, May, and August are the best times to detect saline seeps. Specific times within these months would be dependent upon weather, phenology, and growth conditions. Saline seeps can be efficiently and accurately mapped, within resolution capabilities, from merged May and August LANDSAT 1 data. Seeps were mapped by detecting salt crusts in the spring and indicator plants in the fall. These indicator plants were kochia, inkweed, and foxtail barley. The total hectares of the mapped saline seeps were calculated and tabulated. Saline seeps less than two hectares in size or that have linear configurations less than 200 meters in width were not mapped using the LANDSAT 1 data. Saline seep signatures developed in the Coffee Creek test site were extended to map saline seeps located outside this area.

  6. Sampling and mapping forest volume and biomass using airborne LIDARs

    Treesearch

    Erik Naesset; Terje Gobakken; Ross Nelson

    2009-01-01

    Since around 1995, extensive research efforts have been made in Scandinavia to develop airborne Light Detection and Ranging (LIDAR) as an operational tool for wall-to-wall mapping of forest stands for planning purposes. Scanning LIDAR has the ability to capture the entire three-dimensional structure of forest canopies and has therefore proved to be a very efficient...

  7. ANNIT - An Efficient Inversion Algorithm based on Prediction Principles

    NASA Astrophysics Data System (ADS)

    Růžek, B.; Kolář, P.

    2009-04-01

    Solution of inverse problems represents meaningful job in geophysics. The amount of data is continuously increasing, methods of modeling are being improved and the computer facilities are also advancing great technical progress. Therefore the development of new and efficient algorithms and computer codes for both forward and inverse modeling is still up to date. ANNIT is contributing to this stream since it is a tool for efficient solution of a set of non-linear equations. Typical geophysical problems are based on parametric approach. The system is characterized by a vector of parameters p, the response of the system is characterized by a vector of data d. The forward problem is usually represented by unique mapping F(p)=d. The inverse problem is much more complex and the inverse mapping p=G(d) is available in an analytical or closed form only exceptionally and generally it may not exist at all. Technically, both forward and inverse mapping F and G are sets of non-linear equations. ANNIT solves such situation as follows: (i) joint subspaces {pD, pM} of original data and model spaces D, M, resp. are searched for, within which the forward mapping F is sufficiently smooth that the inverse mapping G does exist, (ii) numerical approximation of G in subspaces {pD, pM} is found, (iii) candidate solution is predicted by using this numerical approximation. ANNIT is working in an iterative way in cycles. The subspaces {pD, pM} are searched for by generating suitable populations of individuals (models) covering data and model spaces. The approximation of the inverse mapping is made by using three methods: (a) linear regression, (b) Radial Basis Function Network technique, (c) linear prediction (also known as "Kriging"). The ANNIT algorithm has built in also an archive of already evaluated models. Archive models are re-used in a suitable way and thus the number of forward evaluations is minimized. ANNIT is now implemented both in MATLAB and SCILAB. Numerical tests show good

  8. Using high-throughput barcode sequencing to efficiently map connectomes

    PubMed Central

    Peikon, Ian D.; Kebschull, Justus M.; Vagin, Vasily V.; Ravens, Diana I.; Sun, Yu-Chi; Brouzes, Eric; Corrêa, Ivan R.; Bressan, Dario

    2017-01-01

    Abstract The function of a neural circuit is determined by the details of its synaptic connections. At present, the only available method for determining a neural wiring diagram with single synapse precision—a ‘connectome’—is based on imaging methods that are slow, labor-intensive and expensive. Here, we present SYNseq, a method for converting the connectome into a form that can exploit the speed and low cost of modern high-throughput DNA sequencing. In SYNseq, each neuron is labeled with a unique random nucleotide sequence—an RNA ‘barcode’—which is targeted to the synapse using engineered proteins. Barcodes in pre- and postsynaptic neurons are then associated through protein-protein crosslinking across the synapse, extracted from the tissue, and joined into a form suitable for sequencing. Although our failure to develop an efficient barcode joining scheme precludes the widespread application of this approach, we expect that with further development SYNseq will enable tracing of complex circuits at high speed and low cost. PMID:28449067

  9. A computational linguistics motivated mapping of ICPC-2 PLUS to SNOMED CT.

    PubMed

    Wang, Yefeng; Patrick, Jon; Miller, Graeme; O'Hallaran, Julie

    2008-10-27

    . Automating as much as is possible of this process turns the searching and mapping task into a validation task, which can effectively reduce the cost and increase the efficiency and accuracy of this task over manual methods.

  10. Mapping Inhibitory Neuronal Circuits by Laser Scanning Photostimulation

    PubMed Central

    Ikrar, Taruna; Olivas, Nicholas D.; Shi, Yulin; Xu, Xiangmin

    2011-01-01

    Inhibitory neurons are crucial to cortical function. They comprise about 20% of the entire cortical neuronal population and can be further subdivided into diverse subtypes based on their immunochemical, morphological, and physiological properties1-4. Although previous research has revealed much about intrinsic properties of individual types of inhibitory neurons, knowledge about their local circuit connections is still relatively limited3,5,6. Given that each individual neuron's function is shaped by its excitatory and inhibitory synaptic input within cortical circuits, we have been using laser scanning photostimulation (LSPS) to map local circuit connections to specific inhibitory cell types. Compared to conventional electrical stimulation or glutamate puff stimulation, LSPS has unique advantages allowing for extensive mapping and quantitative analysis of local functional inputs to individually recorded neurons3,7-9. Laser photostimulation via glutamate uncaging selectively activates neurons perisomatically, without activating axons of passage or distal dendrites, which ensures a sub-laminar mapping resolution. The sensitivity and efficiency of LSPS for mapping inputs from many stimulation sites over a large region are well suited for cortical circuit analysis. Here we introduce the technique of LSPS combined with whole-cell patch clamping for local inhibitory circuit mapping. Targeted recordings of specific inhibitory cell types are facilitated by use of transgenic mice expressing green fluorescent proteins (GFP) in limited inhibitory neuron populations in the cortex3,10, which enables consistent sampling of the targeted cell types and unambiguous identification of the cell types recorded. As for LSPS mapping, we outline the system instrumentation, describe the experimental procedure and data acquisition, and present examples of circuit mapping in mouse primary somatosensory cortex. As illustrated in our experiments, caged glutamate is activated in a spatially

  11. Geovisualization in the HydroProg web map service

    NASA Astrophysics Data System (ADS)

    Spallek, Waldemar; Wieczorek, Malgorzata; Szymanowski, Mariusz; Niedzielski, Tomasz; Swierczynska, Malgorzata

    2016-04-01

    The HydroProg system, built at the University of Wroclaw (Poland) in frame of the research project no. 2011/01/D/ST10/04171 financed by the National Science Centre of Poland, has been designed for computing predictions of river stages in real time on a basis of multimodelling. This experimental system works on the upper Nysa Klodzka basin (SW Poland) above the gauge in the town of Bardo, with the catchment area of 1744 square kilometres. The system operates in association with the Local System for Flood Monitoring of Klodzko County (LSOP), and produces hydrograph prognoses as well as inundation predictions. For presenting the up-to-date predictions and their statistics in the online mode, the dedicated real-time web map service has been designed. Geovisualisation in the HydroProg map service concerns: interactive maps of study area, interactive spaghetti hydrograms of water level forecasts along with observed river stages, animated images of inundation. The LSOP network offers a high spatial and temporal resolution of observations, as the length of the sampling interval is equal to 15 minutes. The main environmental elements related to hydrological modelling are shown on the main map. This includes elevation data (hillshading and hypsometric tints), rivers and reservoirs as well as catchment boundaries. Furthermore, we added main towns, roads as well as political and administrative boundaries for better map understanding. The web map was designed as a multi-scale representation, with levels of detail and zooming according to scales: 1:100 000, 1:250 000 and 1:500 000. Observations of water level in LSOP are shown on interactive hydrographs for each gauge. Additionally, predictions and some of their statistical characteristics (like prediction errors and Nash-Sutcliffe efficiency) are shown for selected gauges. Finally, predictions of inundation are presented on animated maps which have been added for four experimental sites. The HydroProg system is a strictly

  12. A novel algorithm for fast and efficient multifocus wavefront shaping

    NASA Astrophysics Data System (ADS)

    Fayyaz, Zahra; Nasiriavanaki, Mohammadreza

    2018-02-01

    Wavefront shaping using spatial light modulator (SLM) is a popular method for focusing light through a turbid media, such as biological tissues. Usually, in iterative optimization methods, due to the very large number of pixels in SLM, larger pixels are formed, bins, and the phase value of the bins are changed to obtain an optimum phase map, hence a focus. In this study an efficient optimization algorithm is proposed to obtain an arbitrary map of focus utilizing all the SLM pixels or small bin sizes. The application of such methodology in dermatology, hair removal in particular, is explored and discussed

  13. Method and system for efficiently searching an encoded vector index

    DOEpatents

    Bui, Thuan Quang; Egan, Randy Lynn; Kathmann, Kevin James

    2001-09-04

    Method and system aspects for efficiently searching an encoded vector index are provided. The aspects include the translation of a search query into a candidate bitmap, and the mapping of data from the candidate bitmap into a search result bitmap according to entry values in the encoded vector index. Further, the translation includes the setting of a bit in the candidate bitmap for each entry in a symbol table that corresponds to candidate of the search query. Also included in the mapping is the identification of a bit value in the candidate bitmap pointed to by an entry in an encoded vector.

  14. IntegratedMap: a Web interface for integrating genetic map data.

    PubMed

    Yang, Hongyu; Wang, Hongyu; Gingle, Alan R

    2005-05-01

    IntegratedMap is a Web application and database schema for storing and interactively displaying genetic map data. Its Web interface includes a menu for direct chromosome/linkage group selection, a search form for selection based on mapped object location and linkage group displays. An overview display provides convenient access to the full range of mapped and anchored object types with genetic locus details, such as numbers, types and names of mapped/anchored objects displayed in a compact scrollable list box that automatically updates based on selected map location and object type. Also, multilinkage group and localized map views are available along with links that can be configured for integration with other Web resources. IntegratedMap is implemented in C#/ASP.NET and the package, including a MySQL schema creation script, is available from http://cggc.agtec.uga.edu/Data/download.asp

  15. Efficient morse decompositions of vector fields.

    PubMed

    Chen, Guoning; Mischaikow, Konstantin; Laramee, Robert S; Zhang, Eugene

    2008-01-01

    Existing topology-based vector field analysis techniques rely on the ability to extract the individual trajectories such as fixed points, periodic orbits, and separatrices that are sensitive to noise and errors introduced by simulation and interpolation. This can make such vector field analysis unsuitable for rigorous interpretations. We advocate the use of Morse decompositions, which are robust with respect to perturbations, to encode the topological structures of a vector field in the form of a directed graph, called a Morse connection graph (MCG). While an MCG exists for every vector field, it need not be unique. Previous techniques for computing MCG's, while fast, are overly conservative and usually results in MCG's that are too coarse to be useful for the applications. To address this issue, we present a new technique for performing Morse decomposition based on the concept of tau-maps, which typically provides finer MCG's than existing techniques. Furthermore, the choice of tau provides a natural tradeoff between the fineness of the MCG's and the computational costs. We provide efficient implementations of Morse decomposition based on tau-maps, which include the use of forward and backward mapping techniques and an adaptive approach in constructing better approximations of the images of the triangles in the meshes used for simulation.. Furthermore, we propose the use of spatial tau-maps in addition to the original temporal tau-maps. These techniques provide additional trade-offs between the quality of the MCGs and the speed of computation. We demonstrate the utility of our technique with various examples in the plane and on surfaces including engine simulation data sets.

  16. Occupancy Grid Map Merging Using Feature Maps

    DTIC Science & Technology

    2010-11-01

    each robot begins exploring at different starting points, once two robots can communicate, they send their odometry data, LIDAR observations, and maps...robots [11]. Moreover, it is relevant to mention that significant success has been achieved in solving SLAM problems when using hybrid maps [12...represents the environment by parametric features. Our method is capable of representing a LIDAR scanned environment map in a parametric fashion. In general

  17. CloudAligner: A fast and full-featured MapReduce based tool for sequence mapping.

    PubMed

    Nguyen, Tung; Shi, Weisong; Ruden, Douglas

    2011-06-06

    Research in genetics has developed rapidly recently due to the aid of next generation sequencing (NGS). However, massively-parallel NGS produces enormous amounts of data, which leads to storage, compatibility, scalability, and performance issues. The Cloud Computing and MapReduce framework, which utilizes hundreds or thousands of shared computers to map sequencing reads quickly and efficiently to reference genome sequences, appears to be a very promising solution for these issues. Consequently, it has been adopted by many organizations recently, and the initial results are very promising. However, since these are only initial steps toward this trend, the developed software does not provide adequate primary functions like bisulfite, pair-end mapping, etc., in on-site software such as RMAP or BS Seeker. In addition, existing MapReduce-based applications were not designed to process the long reads produced by the most recent second-generation and third-generation NGS instruments and, therefore, are inefficient. Last, it is difficult for a majority of biologists untrained in programming skills to use these tools because most were developed on Linux with a command line interface. To urge the trend of using Cloud technologies in genomics and prepare for advances in second- and third-generation DNA sequencing, we have built a Hadoop MapReduce-based application, CloudAligner, which achieves higher performance, covers most primary features, is more accurate, and has a user-friendly interface. It was also designed to be able to deal with long sequences. The performance gain of CloudAligner over Cloud-based counterparts (35 to 80%) mainly comes from the omission of the reduce phase. In comparison to local-based approaches, the performance gain of CloudAligner is from the partition and parallel processing of the huge reference genome as well as the reads. The source code of CloudAligner is available at http://cloudaligner.sourceforge.net/ and its web version is at http

  18. Mapping Gnss Restricted Environments with a Drone Tandem and Indirect Position Control

    NASA Astrophysics Data System (ADS)

    Cledat, E.; Cucci, D. A.

    2017-08-01

    The problem of autonomously mapping highly cluttered environments, such as urban and natural canyons, is intractable with the current UAV technology. The reason lies in the absence or unreliability of GNSS signals due to partial sky occlusion or multi-path effects. High quality carrier-phase observations are also required in efficient mapping paradigms, such as Assisted Aerial Triangulation, to achieve high ground accuracy without the need of dense networks of ground control points. In this work we consider a drone tandem in which the first drone flies outside the canyon, where GNSS constellation is ideal, visually tracks the second drone and provides an indirect position control for it. This enables both autonomous guidance and accurate mapping of GNSS restricted environments without the need of ground control points. We address the technical feasibility of this concept considering preliminary real-world experiments in comparable conditions and we perform a mapping accuracy prediction based on a simulation scenario.

  19. Making Socialists: Mary Bridges Adams and the Fight for Knowledge and Power, 1855-1939

    ERIC Educational Resources Information Center

    Weiler, Kathleen

    2012-01-01

    This article presents a review of "Making socialists: Mary Bridges Adams and the fight for knowledge and power, 1855-1939," by Jane Martin. Jane Martin has explored the history of late-nineteenth-century and early-twentieth century-British women educational activists in numerous publications over the past two decades. Her first book,…

  20. Exploration of depth modeling mode one lossless wedgelets storage strategies for 3D-high efficiency video coding

    NASA Astrophysics Data System (ADS)

    Sanchez, Gustavo; Marcon, César; Agostini, Luciano Volcan

    2018-01-01

    The 3D-high efficiency video coding has introduced tools to obtain higher efficiency in 3-D video coding, and most of them are related to the depth maps coding. Among these tools, the depth modeling mode-1 (DMM-1) focuses on better encoding edges regions of depth maps. The large memory required for storing all wedgelet patterns is one of the bottlenecks in the DMM-1 hardware design of both encoder and decoder since many patterns must be stored. Three algorithms to reduce the DMM-1 memory requirements and a hardware design targeting the most efficient among these algorithms are presented. Experimental results demonstrate that the proposed solutions surpass related works reducing up to 78.8% of the wedgelet memory, without degrading the encoding efficiency. Synthesis results demonstrate that the proposed algorithm reduces almost 75% of the power dissipation when compared to the standard approach.

  1. Efficient methods and readily customizable libraries for managing complexity of large networks.

    PubMed

    Dogrusoz, Ugur; Karacelik, Alper; Safarli, Ilkin; Balci, Hasan; Dervishi, Leonard; Siper, Metin Can

    2018-01-01

    One common problem in visualizing real-life networks, including biological pathways, is the large size of these networks. Often times, users find themselves facing slow, non-scaling operations due to network size, if not a "hairball" network, hindering effective analysis. One extremely useful method for reducing complexity of large networks is the use of hierarchical clustering and nesting, and applying expand-collapse operations on demand during analysis. Another such method is hiding currently unnecessary details, to later gradually reveal on demand. Major challenges when applying complexity reduction operations on large networks include efficiency and maintaining the user's mental map of the drawing. We developed specialized incremental layout methods for preserving a user's mental map while managing complexity of large networks through expand-collapse and hide-show operations. We also developed open-source JavaScript libraries as plug-ins to the web based graph visualization library named Cytsocape.js to implement these methods as complexity management operations. Through efficient specialized algorithms provided by these extensions, one can collapse or hide desired parts of a network, yielding potentially much smaller networks, making them more suitable for interactive visual analysis. This work fills an important gap by making efficient implementations of some already known complexity management techniques freely available to tool developers through a couple of open source, customizable software libraries, and by introducing some heuristics which can be applied upon such complexity management techniques to ensure preserving mental map of users.

  2. The World Karst Aquifer Mapping project: concept, mapping procedure and map of Europe

    NASA Astrophysics Data System (ADS)

    Chen, Zhao; Auler, Augusto S.; Bakalowicz, Michel; Drew, David; Griger, Franziska; Hartmann, Jens; Jiang, Guanghui; Moosdorf, Nils; Richts, Andrea; Stevanovic, Zoran; Veni, George; Goldscheider, Nico

    2017-05-01

    Karst aquifers contribute substantially to freshwater supplies in many regions of the world, but are vulnerable to contamination and difficult to manage because of their unique hydrogeological characteristics. Many karst systems are hydraulically connected over wide areas and require transboundary exploration, protection and management. In order to obtain a better global overview of karst aquifers, to create a basis for sustainable international water-resources management, and to increase the awareness in the public and among decision makers, the World Karst Aquifer Mapping (WOKAM) project was established. The goal is to create a world map and database of karst aquifers, as a further development of earlier maps. This paper presents the basic concepts and the detailed mapping procedure, using France as an example to illustrate the step-by-step workflow, which includes generalization, differentiation of continuous and discontinuous carbonate and evaporite rock areas, and the identification of non-exposed karst aquifers. The map also shows selected caves and karst springs, which are collected in an associated global database. The draft karst aquifer map of Europe shows that 21.6% of the European land surface is characterized by the presence of (continuous or discontinuous) carbonate rocks; about 13.8% of the land surface is carbonate rock outcrop.

  3. Do turning visited routes in black maps into white promote sightseeing?

    NASA Astrophysics Data System (ADS)

    Izumi, Tomoko; Nakatani, Yoshio

    2017-07-01

    In this paper, we propose a new approach for promoting visiting various areas during a sightseeing based on the "FUrther BENEfit of a Kind of Inconvenience" (FUBEN-EKI) theory. FUBEN-EKI is a theory such that an inconvenient system in some aspect enables users to obtain more hidden benefits than convenient systems. Familiar navigation systems lead tourists to time efficient routes, and so the systems may limit their opportunities for new discoveries during their sightseeing. It is supposed that if tourists walk freely in various areas then they will have more change to find new discoveries. To promote visiting various areas, we propose a blacked-out map: The map is hidden by a black filter initially, but only the area where the user moves becomes clear. Since the user cannot see the map of unvisited area, the user thinks that the unvisited area might have interesting sightseeing spots. Moreover, to make all of the area clear, the user have to go to all of the area on the map.

  4. A strategy to load balancing for non-connectivity MapReduce job

    NASA Astrophysics Data System (ADS)

    Zhou, Huaping; Liu, Guangzong; Gui, Haixia

    2017-09-01

    MapReduce has been widely used in large scale and complex datasets as a kind of distributed programming model. Original Hash partitioning function in MapReduce often results the problem of data skew when data distribution is uneven. To solve the imbalance of data partitioning, we proposes a strategy to change the remaining partitioning index when data is skewed. In Map phase, we count the amount of data which will be distributed to each reducer, then Job Tracker monitor the global partitioning information and dynamically modify the original partitioning function according to the data skew model, so the Partitioner can change the index of these partitioning which will cause data skew to the other reducer that has less load in the next partitioning process, and can eventually balance the load of each node. Finally, we experimentally compare our method with existing methods on both synthetic and real datasets, the experimental results show our strategy can solve the problem of data skew with better stability and efficiency than Hash method and Sampling method for non-connectivity MapReduce task.

  5. Spectral edge: gradient-preserving spectral mapping for image fusion.

    PubMed

    Connah, David; Drew, Mark S; Finlayson, Graham D

    2015-12-01

    This paper describes a novel approach to image fusion for color display. Our goal is to generate an output image whose gradient matches that of the input as closely as possible. We achieve this using a constrained contrast mapping paradigm in the gradient domain, where the structure tensor of a high-dimensional gradient representation is mapped exactly to that of a low-dimensional gradient field which is then reintegrated to form an output. Constraints on output colors are provided by an initial RGB rendering. Initially, we motivate our solution with a simple "ansatz" (educated guess) for projecting higher-D contrast onto color gradients, which we expand to a more rigorous theorem to incorporate color constraints. The solution to these constrained optimizations is closed-form, allowing for simple and hence fast and efficient algorithms. The approach can map any N-D image data to any M-D output and can be used in a variety of applications using the same basic algorithm. In this paper, we focus on the problem of mapping N-D inputs to 3D color outputs. We present results in five applications: hyperspectral remote sensing, fusion of color and near-infrared or clear-filter images, multilighting imaging, dark flash, and color visualization of magnetic resonance imaging diffusion-tensor imaging.

  6. Secure Chaotic Map Based Block Cryptosystem with Application to Camera Sensor Networks

    PubMed Central

    Guo, Xianfeng; Zhang, Jiashu; Khan, Muhammad Khurram; Alghathbar, Khaled

    2011-01-01

    Recently, Wang et al. presented an efficient logistic map based block encryption system. The encryption system employs feedback ciphertext to achieve plaintext dependence of sub-keys. Unfortunately, we discovered that their scheme is unable to withstand key stream attack. To improve its security, this paper proposes a novel chaotic map based block cryptosystem. At the same time, a secure architecture for camera sensor network is constructed. The network comprises a set of inexpensive camera sensors to capture the images, a sink node equipped with sufficient computation and storage capabilities and a data processing server. The transmission security between the sink node and the server is gained by utilizing the improved cipher. Both theoretical analysis and simulation results indicate that the improved algorithm can overcome the flaws and maintain all the merits of the original cryptosystem. In addition, computational costs and efficiency of the proposed scheme are encouraging for the practical implementation in the real environment as well as camera sensor network. PMID:22319371

  7. Secure chaotic map based block cryptosystem with application to camera sensor networks.

    PubMed

    Guo, Xianfeng; Zhang, Jiashu; Khan, Muhammad Khurram; Alghathbar, Khaled

    2011-01-01

    Recently, Wang et al. presented an efficient logistic map based block encryption system. The encryption system employs feedback ciphertext to achieve plaintext dependence of sub-keys. Unfortunately, we discovered that their scheme is unable to withstand key stream attack. To improve its security, this paper proposes a novel chaotic map based block cryptosystem. At the same time, a secure architecture for camera sensor network is constructed. The network comprises a set of inexpensive camera sensors to capture the images, a sink node equipped with sufficient computation and storage capabilities and a data processing server. The transmission security between the sink node and the server is gained by utilizing the improved cipher. Both theoretical analysis and simulation results indicate that the improved algorithm can overcome the flaws and maintain all the merits of the original cryptosystem. In addition, computational costs and efficiency of the proposed scheme are encouraging for the practical implementation in the real environment as well as camera sensor network.

  8. Mapping the Indonesian territory, based on pollution, social demography and geographical data, using self organizing feature map

    NASA Astrophysics Data System (ADS)

    Hernawati, Kuswari; Insani, Nur; Bambang S. H., M.; Nur Hadi, W.; Sahid

    2017-08-01

    This research aims to mapping the 33 (thirty-three) provinces in Indonesia, based on the data on air, water and soil pollution, as well as social demography and geography data, into a clustered model. The method used in this study was unsupervised method that combines the basic concept of Kohonen or Self-Organizing Feature Maps (SOFM). The method is done by providing the design parameters for the model based on data related directly/ indirectly to pollution, which are the demographic and social data, pollution levels of air, water and soil, as well as the geographical situation of each province. The parameters used consists of 19 features/characteristics, including the human development index, the number of vehicles, the availability of the plant's water absorption and flood prevention, as well as geographic and demographic situation. The data used were secondary data from the Central Statistics Agency (BPS), Indonesia. The data are mapped into SOFM from a high-dimensional vector space into two-dimensional vector space according to the closeness of location in term of Euclidean distance. The resulting outputs are represented in clustered grouping. Thirty-three provinces are grouped into five clusters, where each cluster has different features/characteristics and level of pollution. The result can used to help the efforts on prevention and resolution of pollution problems on each cluster in an effective and efficient way.

  9. Performance Measurements and Mapping of a R-407C Vapor Injection Scroll Compressor

    NASA Astrophysics Data System (ADS)

    Lumpkin, Domenique; Spielbauer, Niklas; Groll, Eckhard

    2017-08-01

    Environmental conditions significantly define the performance of HVAC&R systems. Vapor compression systems in hot climates tend to operate at higher pressure ratios, leading to increased discharge temperatures. Higher discharge temperatures can lead to higher irreversibilities in the compression process, lower specific enthalpies differences across the evaporator, and possibly a reduction in the compressor life due to the breakdown of the oil used for lubrication. To counter these effects, the use of economized, vapor injection compressors is proposed for vapor compression systems in high temperature climates. Such compressors are commercially available for refrigeration applications, in particular, supermarket refrigeration systems. However, compressor maps for vapor injection compressors are limited and none exist for R-407C. Through calorimeter testing, a compressor map for a single-port vapor injection compressor using R-407C was developed. A standard correlation for mapping single-port vapor injection compressors is proposed and validated using the compressor test results. The system and compressor performance with and without vapor injection was considered. As expected, with vapor injection there was a reduction in compressor discharge temperatures and an increase in the system coefficient of performance. The proposed dimensionless correlation is more accurate than the AHRI polynomial for mapping the injection ratio, discharge temperature, and compressor heat loss. The predicted volumetric efficiency values from the dimensionless correlation is within 1% of the measured valued. Similarly, the predicted isentropic efficiency values are within 2% of the measured values.

  10. Regional gene mapping using mixed radiation hybrids and reverse chromosome painting.

    PubMed

    Lin, J Y; Bedford, J S

    1997-11-01

    We describe a new approach for low-resolution physical mapping using pooled DNA probe from mixed (non-clonal) populations of human-CHO cell hybrids and reverse chromosome painting. This mapping method is based on a process in which the human chromosome fragments bearing a complementing gene were selectively retained in a large non-clonal population of CHO-human hybrid cells during a series of 12- to 15-Gy gamma irradiations each followed by continuous growth selection. The location of the gene could then be identified by reverse chromosome painting on normal human metaphase spreads using biotinylated DNA from this population of "enriched" hybrid cells. We tested the validity of this method by correctly mapping the complementing human HPRT gene, whose location is well established. We then demonstrated the method's usefulness by mapping the chromosome location of a human gene which complemented the defect responsible for the hypersensitivity to ionizing radiation in CHO irs-20 cells. This method represents an efficient alternative to conventional concordance analysis in somatic cell hybrids where detailed chromosome analysis of numerous hybrid clones is necessary. Using this approach, it is possible to localize a gene for which there is no prior sequence or linkage information to a subchromosomal region, thus facilitating association with known mapping landmarks (e.g. RFLP, YAC or STS contigs) for higher-resolution mapping.

  11. The Performance Analysis of AN Indoor Mobile Mapping System with Rgb-D Sensor

    NASA Astrophysics Data System (ADS)

    Tsai, G. J.; Chiang, K. W.; Chu, C. H.; Chen, Y. L.; El-Sheimy, N.; Habib, A.

    2015-08-01

    Over the years, Mobile Mapping Systems (MMSs) have been widely applied to urban mapping, path management and monitoring and cyber city, etc. The key concept of mobile mapping is based on positioning technology and photogrammetry. In order to achieve the integration, multi-sensor integrated mapping technology has clearly established. In recent years, the robotic technology has been rapidly developed. The other mapping technology that is on the basis of low-cost sensor has generally used in robotic system, it is known as the Simultaneous Localization and Mapping (SLAM). The objective of this study is developed a prototype of indoor MMS for mobile mapping applications, especially to reduce the costs and enhance the efficiency of data collection and validation of direct georeferenced (DG) performance. The proposed indoor MMS is composed of a tactical grade Inertial Measurement Unit (IMU), the Kinect RGB-D sensor and light detection, ranging (LIDAR) and robot. In summary, this paper designs the payload for indoor MMS to generate the floor plan. In first session, it concentrates on comparing the different positioning algorithms in the indoor environment. Next, the indoor plans are generated by two sensors, Kinect RGB-D sensor LIDAR on robot. Moreover, the generated floor plan will compare with the known plan for both validation and verification.

  12. LOD map--A visual interface for navigating multiresolution volume visualization.

    PubMed

    Wang, Chaoli; Shen, Han-Wei

    2006-01-01

    In multiresolution volume visualization, a visual representation of level-of-detail (LOD) quality is important for us to examine, compare, and validate different LOD selection algorithms. While traditional methods rely on ultimate images for quality measurement, we introduce the LOD map--an alternative representation of LOD quality and a visual interface for navigating multiresolution data exploration. Our measure for LOD quality is based on the formulation of entropy from information theory. The measure takes into account the distortion and contribution of multiresolution data blocks. A LOD map is generated through the mapping of key LOD ingredients to a treemap representation. The ordered treemap layout is used for relative stable update of the LOD map when the view or LOD changes. This visual interface not only indicates the quality of LODs in an intuitive way, but also provides immediate suggestions for possible LOD improvement through visually-striking features. It also allows us to compare different views and perform rendering budget control. A set of interactive techniques is proposed to make the LOD adjustment a simple and easy task. We demonstrate the effectiveness and efficiency of our approach on large scientific and medical data sets.

  13. Lava flow risk maps at Mount Cameroon volcano

    NASA Astrophysics Data System (ADS)

    Favalli, M.; Fornaciai, A.; Papale, P.; Tarquini, S.

    2009-04-01

    . Maps of type (2) and (3) are useful at the onset of a new eruption, when a vent forms. The combined use of these maps provides an efficient tool for lava flow risk assessment at Mount Cameroon.

  14. Wide-Baseline Stereo-Based Obstacle Mapping for Unmanned Surface Vehicles

    PubMed Central

    Mou, Xiaozheng; Wang, Han

    2018-01-01

    This paper proposes a wide-baseline stereo-based static obstacle mapping approach for unmanned surface vehicles (USVs). The proposed approach eliminates the complicated calibration work and the bulky rig in our previous binocular stereo system, and raises the ranging ability from 500 to 1000 m with a even larger baseline obtained from the motion of USVs. Integrating a monocular camera with GPS and compass information in this proposed system, the world locations of the detected static obstacles are reconstructed while the USV is traveling, and an obstacle map is then built. To achieve more accurate and robust performance, multiple pairs of frames are leveraged to synthesize the final reconstruction results in a weighting model. Experimental results based on our own dataset demonstrate the high efficiency of our system. To the best of our knowledge, we are the first to address the task of wide-baseline stereo-based obstacle mapping in a maritime environment. PMID:29617293

  15. 29 CFR 4006.6 - Definition of “participant.”

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... of $5,000 or less leaves employment, the benefit will be immediately cashed out. On December 30, 2013... treated as not having an accrued benefit on December 31, 2013 (the participant count date for the 2014 premium), because Jane's benefit is treated as having been paid on December 30, 2013. Thus, Jane is not...

  16. Unmanned Aircraft Systems for Logistics Applications

    DTIC Science & Technology

    2011-01-01

    Jane’s Air-Launched Weapons, “ GBU - 39 /B Small Diameter Bomb (SDB) and SDB II (United States),” June 5, 2009. As of September 11, 2009: http...www.janes.com/articles/Janes-Air-Launched-Weapons/ GBU - 39 -B-Small- Diameter-Bomb-SDB-and-SDB-II-United-States.html [subscription required] Jane’s Electro

  17. Efficient LIDAR Point Cloud Data Managing and Processing in a Hadoop-Based Distributed Framework

    NASA Astrophysics Data System (ADS)

    Wang, C.; Hu, F.; Sha, D.; Han, X.

    2017-10-01

    Light Detection and Ranging (LiDAR) is one of the most promising technologies in surveying and mapping city management, forestry, object recognition, computer vision engineer and others. However, it is challenging to efficiently storage, query and analyze the high-resolution 3D LiDAR data due to its volume and complexity. In order to improve the productivity of Lidar data processing, this study proposes a Hadoop-based framework to efficiently manage and process LiDAR data in a distributed and parallel manner, which takes advantage of Hadoop's storage and computing ability. At the same time, the Point Cloud Library (PCL), an open-source project for 2D/3D image and point cloud processing, is integrated with HDFS and MapReduce to conduct the Lidar data analysis algorithms provided by PCL in a parallel fashion. The experiment results show that the proposed framework can efficiently manage and process big LiDAR data.

  18. NeatMap--non-clustering heat map alternatives in R.

    PubMed

    Rajaram, Satwik; Oono, Yoshi

    2010-01-22

    The clustered heat map is the most popular means of visualizing genomic data. It compactly displays a large amount of data in an intuitive format that facilitates the detection of hidden structures and relations in the data. However, it is hampered by its use of cluster analysis which does not always respect the intrinsic relations in the data, often requiring non-standardized reordering of rows/columns to be performed post-clustering. This sometimes leads to uninformative and/or misleading conclusions. Often it is more informative to use dimension-reduction algorithms (such as Principal Component Analysis and Multi-Dimensional Scaling) which respect the topology inherent in the data. Yet, despite their proven utility in the analysis of biological data, they are not as widely used. This is at least partially due to the lack of user-friendly visualization methods with the visceral impact of the heat map. NeatMap is an R package designed to meet this need. NeatMap offers a variety of novel plots (in 2 and 3 dimensions) to be used in conjunction with these dimension-reduction techniques. Like the heat map, but unlike traditional displays of such results, it allows the entire dataset to be displayed while visualizing relations between elements. It also allows superimposition of cluster analysis results for mutual validation. NeatMap is shown to be more informative than the traditional heat map with the help of two well-known microarray datasets. NeatMap thus preserves many of the strengths of the clustered heat map while addressing some of its deficiencies. It is hoped that NeatMap will spur the adoption of non-clustering dimension-reduction algorithms.

  19. A Fast Approximate Algorithm for Mapping Long Reads to Large Reference Databases.

    PubMed

    Jain, Chirag; Dilthey, Alexander; Koren, Sergey; Aluru, Srinivas; Phillippy, Adam M

    2018-04-30

    Emerging single-molecule sequencing technologies from Pacific Biosciences and Oxford Nanopore have revived interest in long-read mapping algorithms. Alignment-based seed-and-extend methods demonstrate good accuracy, but face limited scalability, while faster alignment-free methods typically trade decreased precision for efficiency. In this article, we combine a fast approximate read mapping algorithm based on minimizers with a novel MinHash identity estimation technique to achieve both scalability and precision. In contrast to prior methods, we develop a mathematical framework that defines the types of mapping targets we uncover, establish probabilistic estimates of p-value and sensitivity, and demonstrate tolerance for alignment error rates up to 20%. With this framework, our algorithm automatically adapts to different minimum length and identity requirements and provides both positional and identity estimates for each mapping reported. For mapping human PacBio reads to the hg38 reference, our method is 290 × faster than Burrows-Wheeler Aligner-MEM with a lower memory footprint and recall rate of 96%. We further demonstrate the scalability of our method by mapping noisy PacBio reads (each ≥5 kbp in length) to the complete NCBI RefSeq database containing 838 Gbp of sequence and >60,000 genomes.

  20. Human Mind Maps

    ERIC Educational Resources Information Center

    Glass, Tom

    2016-01-01

    When students generate mind maps, or concept maps, the maps are usually on paper, computer screens, or a blackboard. Human Mind Maps require few resources and little preparation. The main requirements are space where students can move around and a little creativity and imagination. Mind maps can be used for a variety of purposes, and Human Mind…

  1. Zernike analysis of all-sky night brightness maps.

    PubMed

    Bará, Salvador; Nievas, Miguel; Sánchez de Miguel, Alejandro; Zamorano, Jaime

    2014-04-20

    All-sky night brightness maps (calibrated images of the night sky with hemispherical field-of-view (FOV) taken at standard photometric bands) provide useful data to assess the light pollution levels at any ground site. We show that these maps can be efficiently described and analyzed using Zernike circle polynomials. The relevant image information can be compressed into a low-dimensional coefficients vector, giving an analytical expression for the sky brightness and alleviating the effects of noise. Moreover, the Zernike expansions allow us to quantify in a straightforward way the average and zenithal sky brightness and its variation across the FOV, providing a convenient framework to study the time course of these magnitudes. We apply this framework to analyze the results of a one-year campaign of night sky brightness measurements made at the UCM observatory in Madrid.

  2. Lithographed spectrometers for tomographic line mapping of the Epoch of Reionization

    NASA Astrophysics Data System (ADS)

    O'Brient, R.; Bock, J. J.; Bradford, C. M.; Crites, A.; Duan, R.; Hailey-Dunsheath, S.; Hunacek, J.; LeDuc, R.; Shirokoff, E.; Staniszewski, Z.; Turner, A.; Zemcov, M.

    2014-08-01

    The Tomographic Ionized carbon Mapping Experiment (TIME) is a multi-phased experiment that will topographically map [CII] emission from the Epoch of Reionization. We are developing lithographed spectrometers that couple to TES bolometers in anticipation of the second generation instrument. Our design intentionally mirrors many features of the parallel SuperSpec project, inductively coupling power from a trunk-line microstrip onto half-wave resonators. The resonators couple to a rat-race hybrids that feeds TES bolometers. Our 25 channel prototype shows spectrally positioned lines roughly matching design with a receiver optical efficiency of 15-20%, a level that is dominated by loss in components outside the spectrometer.

  3. Ningaloo Reef: Shallow Marine Habitats Mapped Using a Hyperspectral Sensor

    PubMed Central

    Kobryn, Halina T.; Wouters, Kristin; Beckley, Lynnath E.; Heege, Thomas

    2013-01-01

    Research, monitoring and management of large marine protected areas require detailed and up-to-date habitat maps. Ningaloo Marine Park (including the Muiron Islands) in north-western Australia (stretching across three degrees of latitude) was mapped to 20 m depth using HyMap airborne hyperspectral imagery (125 bands) at 3.5 m resolution across the 762 km2 of reef environment between the shoreline and reef slope. The imagery was corrected for atmospheric, air-water interface and water column influences to retrieve bottom reflectance and bathymetry using the physics-based Modular Inversion and Processing System. Using field-validated, image-derived spectra from a representative range of cover types, the classification combined a semi-automated, pixel-based approach with fuzzy logic and derivative techniques. Five thematic classification levels for benthic cover (with probability maps) were generated with varying degrees of detail, ranging from a basic one with three classes (biotic, abiotic and mixed) to the most detailed with 46 classes. The latter consisted of all abiotic and biotic seabed components and hard coral growth forms in dominant or mixed states. The overall accuracy of mapping for the most detailed maps was 70% for the highest classification level. Macro-algal communities formed most of the benthic cover, while hard and soft corals represented only about 7% of the mapped area (58.6 km2). Dense tabulate coral was the largest coral mosaic type (37% of all corals) and the rest of the corals were a mix of tabulate, digitate, massive and soft corals. Our results show that for this shallow, fringing reef environment situated in the arid tropics, hyperspectral remote sensing techniques can offer an efficient and cost-effective approach to mapping and monitoring reef habitats over large, remote and inaccessible areas. PMID:23922921

  4. CrowdMapping: A Crowdsourcing-Based Terminology Mapping Method for Medical Data Standardization.

    PubMed

    Mao, Huajian; Chi, Chenyang; Huang, Boyu; Meng, Haibin; Yu, Jinghui; Zhao, Dongsheng

    2017-01-01

    Standardized terminology is the prerequisite of data exchange in analysis of clinical processes. However, data from different electronic health record systems are based on idiosyncratic terminology systems, especially when the data is from different hospitals and healthcare organizations. Terminology standardization is necessary for the medical data analysis. We propose a crowdsourcing-based terminology mapping method, CrowdMapping, to standardize the terminology in medical data. CrowdMapping uses a confidential model to determine how terminologies are mapped to a standard system, like ICD-10. The model uses mappings from different health care organizations and evaluates the diversity of the mapping to determine a more sophisticated mapping rule. Further, the CrowdMapping model enables users to rate the mapping result and interact with the model evaluation. CrowdMapping is a work-in-progress system, we present initial results mapping terminologies.

  5. On-board computational efficiency in real time UAV embedded terrain reconstruction

    NASA Astrophysics Data System (ADS)

    Partsinevelos, Panagiotis; Agadakos, Ioannis; Athanasiou, Vasilis; Papaefstathiou, Ioannis; Mertikas, Stylianos; Kyritsis, Sarantis; Tripolitsiotis, Achilles; Zervos, Panagiotis

    2014-05-01

    In the last few years, there is a surge of applications for object recognition, interpretation and mapping using unmanned aerial vehicles (UAV). Specifications in constructing those UAVs are highly diverse with contradictory characteristics including cost-efficiency, carrying weight, flight time, mapping precision, real time processing capabilities, etc. In this work, a hexacopter UAV is employed for near real time terrain mapping. The main challenge addressed is to retain a low cost flying platform with real time processing capabilities. The UAV weight limitation affecting the overall flight time, makes the selection of the on-board processing components particularly critical. On the other hand, surface reconstruction, as a computational demanding task, calls for a highly demanding processing unit on board. To merge these two contradicting aspects along with customized development, a System on a Chip (SoC) integrated circuit is proposed as a low-power, low-cost processor, which natively supports camera sensors and positioning and navigation systems. Modern SoCs, such as Omap3530 or Zynq, are classified as heterogeneous devices and provide a versatile platform, allowing access to both general purpose processors, such as the ARM11, as well as specialized processors, such as a digital signal processor and floating field-programmable gate array. A UAV equipped with the proposed embedded processors, allows on-board terrain reconstruction using stereo vision in near real time. Furthermore, according to the frame rate required, additional image processing may concurrently take place, such as image rectification andobject detection. Lastly, the onboard positioning and navigation (e.g., GNSS) chip may further improve the quality of the generated map. The resulting terrain maps are compared to ground truth geodetic measurements in order to access the accuracy limitations of the overall process. It is shown that with our proposed novel system,there is much potential in

  6. Using high-throughput barcode sequencing to efficiently map connectomes.

    PubMed

    Peikon, Ian D; Kebschull, Justus M; Vagin, Vasily V; Ravens, Diana I; Sun, Yu-Chi; Brouzes, Eric; Corrêa, Ivan R; Bressan, Dario; Zador, Anthony M

    2017-07-07

    The function of a neural circuit is determined by the details of its synaptic connections. At present, the only available method for determining a neural wiring diagram with single synapse precision-a 'connectome'-is based on imaging methods that are slow, labor-intensive and expensive. Here, we present SYNseq, a method for converting the connectome into a form that can exploit the speed and low cost of modern high-throughput DNA sequencing. In SYNseq, each neuron is labeled with a unique random nucleotide sequence-an RNA 'barcode'-which is targeted to the synapse using engineered proteins. Barcodes in pre- and postsynaptic neurons are then associated through protein-protein crosslinking across the synapse, extracted from the tissue, and joined into a form suitable for sequencing. Although our failure to develop an efficient barcode joining scheme precludes the widespread application of this approach, we expect that with further development SYNseq will enable tracing of complex circuits at high speed and low cost. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  7. USGS Maps

    USGS Publications Warehouse

    ,

    1994-01-01

    Most USGS topographic maps use brown contours to show the shape and elevation of the terrain. Elevations are usually shown in feet, but on some maps they are in meters. Contour intervals vary, depending mainly on the scale of the map and the type of terrain.

  8. Impact of population structure, effective bottleneck time, and allele frequency on linkage disequilibrium maps

    PubMed Central

    Zhang, Weihua; Collins, Andrew; Gibson, Jane; Tapper, William J.; Hunt, Sarah; Deloukas, Panos; Bentley, David R.; Morton, Newton E.

    2004-01-01

    Genetic maps in linkage disequilibrium (LD) units play the same role for association mapping as maps in centimorgans provide at much lower resolution for linkage mapping. Association mapping of genes determining disease susceptibility and other phenotypes is based on the theory of LD, here applied to relations with three phenomena. To test the theory, markers at high density along a 10-Mb continuous segment of chromosome 20q were studied in African-American, Asian, and Caucasian samples. Population structure, whether created by pooling samples from divergent populations or by the mating pattern in a mixed population, is accurately bioassayed from genotype frequencies. The effective bottleneck time for Eurasians is substantially less than for migration out of Africa, reflecting later bottlenecks. The classical dependence of allele frequency on mutation age does not hold for the generally shorter time span of inbreeding and LD. Limitation of the classical theory to mutation age justifies the assumption of constant time in a LD map, except for alleles that were rare at the effective bottleneck time or have arisen since. This assumption is derived from the Malecot model and verified in all samples. Tested measures of relative efficiency, support intervals, and localization error determine the operating characteristics of LD maps that are applicable to every sexually reproducing species, with implications for association mapping, high-resolution linkage maps, evolutionary inference, and identification of recombinogenic sequences. PMID:15604137

  9. Impact of population structure, effective bottleneck time, and allele frequency on linkage disequilibrium maps.

    PubMed

    Zhang, Weihua; Collins, Andrew; Gibson, Jane; Tapper, William J; Hunt, Sarah; Deloukas, Panos; Bentley, David R; Morton, Newton E

    2004-12-28

    Genetic maps in linkage disequilibrium (LD) units play the same role for association mapping as maps in centimorgans provide at much lower resolution for linkage mapping. Association mapping of genes determining disease susceptibility and other phenotypes is based on the theory of LD, here applied to relations with three phenomena. To test the theory, markers at high density along a 10-Mb continuous segment of chromosome 20q were studied in African-American, Asian, and Caucasian samples. Population structure, whether created by pooling samples from divergent populations or by the mating pattern in a mixed population, is accurately bioassayed from genotype frequencies. The effective bottleneck time for Eurasians is substantially less than for migration out of Africa, reflecting later bottlenecks. The classical dependence of allele frequency on mutation age does not hold for the generally shorter time span of inbreeding and LD. Limitation of the classical theory to mutation age justifies the assumption of constant time in a LD map, except for alleles that were rare at the effective bottleneck time or have arisen since. This assumption is derived from the Malecot model and verified in all samples. Tested measures of relative efficiency, support intervals, and localization error determine the operating characteristics of LD maps that are applicable to every sexually reproducing species, with implications for association mapping, high-resolution linkage maps, evolutionary inference, and identification of recombinogenic sequences.

  10. Polder maps: Improving OMIT maps by excluding bulk solvent

    DOE PAGES

    Liebschner, Dorothee; Afonine, Pavel V.; Moriarty, Nigel W.; ...

    2017-02-01

    The crystallographic maps that are routinely used during the structure-solution workflow are almost always model-biased because model information is used for their calculation. As these maps are also used to validate the atomic models that result from model building and refinement, this constitutes an immediate problem: anything added to the model will manifest itself in the map and thus hinder the validation. OMIT maps are a common tool to verify the presence of atoms in the model. The simplest way to compute an OMIT map is to exclude the atoms in question from the structure, update the corresponding structure factorsmore » and compute a residual map. It is then expected that if these atoms are present in the crystal structure, the electron density for the omitted atoms will be seen as positive features in this map. This, however, is complicated by the flat bulk-solvent model which is almost universally used in modern crystallographic refinement programs. This model postulates constant electron density at any voxel of the unit-cell volume that is not occupied by the atomic model. Consequently, if the density arising from the omitted atoms is weak then the bulk-solvent model may obscure it further. A possible solution to this problem is to prevent bulk solvent from entering the selected OMIT regions, which may improve the interpretative power of residual maps. This approach is called a polder (OMIT) map. Polder OMIT maps can be particularly useful for displaying weak densities of ligands, solvent molecules, side chains, alternative conformations and residues both in terminal regions and in loops. As a result, the tools described in this manuscript have been implemented and are available in PHENIX.« less

  11. Integrating Databases with Maps: The Delivery of Cultural Data through TimeMap.

    ERIC Educational Resources Information Center

    Johnson, Ian

    TimeMap is a unique integration of database management, metadata and interactive maps, designed to contextualise and deliver cultural data through maps. TimeMap extends conventional maps with the time dimension, creating and animating maps "on-the-fly"; delivers them as a kiosk application or embedded in Web pages; links flexibly to…

  12. Apriori Versions Based on MapReduce for Mining Frequent Patterns on Big Data.

    PubMed

    Luna, Jose Maria; Padillo, Francisco; Pechenizkiy, Mykola; Ventura, Sebastian

    2017-09-27

    Pattern mining is one of the most important tasks to extract meaningful and useful information from raw data. This task aims to extract item-sets that represent any type of homogeneity and regularity in data. Although many efficient algorithms have been developed in this regard, the growing interest in data has caused the performance of existing pattern mining techniques to be dropped. The goal of this paper is to propose new efficient pattern mining algorithms to work in big data. To this aim, a series of algorithms based on the MapReduce framework and the Hadoop open-source implementation have been proposed. The proposed algorithms can be divided into three main groups. First, two algorithms [Apriori MapReduce (AprioriMR) and iterative AprioriMR] with no pruning strategy are proposed, which extract any existing item-set in data. Second, two algorithms (space pruning AprioriMR and top AprioriMR) that prune the search space by means of the well-known anti-monotone property are proposed. Finally, a last algorithm (maximal AprioriMR) is also proposed for mining condensed representations of frequent patterns. To test the performance of the proposed algorithms, a varied collection of big data datasets have been considered, comprising up to 3 · 10#x00B9;⁸ transactions and more than 5 million of distinct single-items. The experimental stage includes comparisons against highly efficient and well-known pattern mining algorithms. Results reveal the interest of applying MapReduce versions when complex problems are considered, and also the unsuitability of this paradigm when dealing with small data.

  13. 78 FR 24781 - 60-Day Notice of Proposed Information Collection: U.S. Passport Renewal Application for Eligible...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-04-26

    ..., Line 1: Replaced ``Street/RFD , P.O. Box, or URB'' with ``(Street/RFD , URB, or P.O. Box, Apt/Unit, In Care Of or Attn (e.g., In Care Of--Jane Doe, Apt 100))''. [cir] Mailing Address, Address Line 2... applicable. (e.g., In Care Of--Jane Doe, Apt 100)'' with ``(If applicable)''. In the photograph box, the...

  14. Inter-Agency Approaches to the Development of a School-Based Student Health Service

    ERIC Educational Resources Information Center

    Harrison, Jane; Bullock, Jane

    2005-01-01

    Young people living in rural areas lack opportunities for accessing health advice and care without reference to a parent, carer, or other adult. In this article Jane Harrison and Jane Bullock provide the rationale for the development in 1997 of Bodyzone, a school-based health service to address this problem. Presented here as a case study,…

  15. Using Text Sets to Facilitate Critical Thinking in Sixth Graders

    ERIC Educational Resources Information Center

    Scales, Roya Q.; Tracy, Kelly N.

    2017-01-01

    This case study examines features and processes of a sixth grade teacher (Jane) utilizing text sets as a tool for facilitating critical thinking. Jane's strong vision and student-centered beliefs informed her use of various texts to teach language arts as she worked to address demands of the Common Core State Standards. Text sets promoted multiple…

  16. Regional Disparities in Online Map User Access Volume and Determining Factors

    NASA Astrophysics Data System (ADS)

    Li, R.; Yang, N.; Li, R.; Huang, W.; Wu, H.

    2017-09-01

    The regional disparities of online map user access volume (use `user access volume' in this paper to indicate briefly) is a topic of growing interest with the increment of popularity in public users, which helps to target the construction of geographic information services for different areas. At first place we statistically analysed the online map user access logs and quantified these regional access disparities on different scales. The results show that the volume of user access is decreasing from east to the west in China as a whole, while East China produces the most access volume; these cities are also the crucial economic and transport centres. Then Principal Component Regression (PCR) is applied to explore the regional disparities of user access volume. A determining model for Online Map access volume is proposed afterwards, which indicates that area scale is the primary determining factor for regional disparities, followed by public transport development level and public service development level. Other factors like user quality index and financial index have very limited influence on the user access volume. According to the study of regional disparities in user access volume, map providers can reasonably dispatch and allocate the data resources and service resources in each area and improve the operational efficiency of the Online Map server cluster.

  17. MROrchestrator: A Fine-Grained Resource Orchestration Framework for MapReduce Clusters

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sharma, Bikash; Prabhakar, Ramya; Kandemir, Mahmut

    2012-01-01

    Efficient resource management in data centers and clouds running large distributed data processing frameworks like MapReduce is crucial for enhancing the performance of hosted applications and boosting resource utilization. However, existing resource scheduling schemes in Hadoop MapReduce allocate resources at the granularity of fixed-size, static portions of nodes, called slots. In this work, we show that MapReduce jobs have widely varying demands for multiple resources, making the static and fixed-size slot-level resource allocation a poor choice both from the performance and resource utilization standpoints. Furthermore, lack of co-ordination in the management of mul- tiple resources across nodes prevents dynamic slotmore » reconfigura- tion, and leads to resource contention. Motivated by this, we propose MROrchestrator, a MapReduce resource Orchestrator framework, which can dynamically identify resource bottlenecks, and resolve them through fine-grained, co-ordinated, and on- demand resource allocations. We have implemented MROrches- trator on two 24-node native and virtualized Hadoop clusters. Experimental results with a suite of representative MapReduce benchmarks demonstrate up to 38% reduction in job completion times, and up to 25% increase in resource utilization. We further show how popular resource managers like NGM and Mesos when augmented with MROrchestrator can hike up their performance.« less

  18. Land use survey and mapping and water resources investigation in Korea

    NASA Technical Reports Server (NTRS)

    Choi, J. H.; Kim, W. I.; Son, D. S. (Principal Investigator)

    1978-01-01

    The author has identified the following significant results. Land use imagery is applicable to land use classification for small scale land use mapping less than 1:250,000. Land use mapping by satellite is more efficient and more cost-effective than land use mapping from conventional medium altitude aerial photographs. Six categories of level 1 land use classification are recognizable from MSS imagery. A hydrogeomorphological study of the Han River basin indicates that band 7 is useful for recognizing the soil and the weathering part of bed rock. The morphological change of the main river is accurately recognized and the drainage system in the area observed is easily classified because of the more or less simple rock type. Although the direct hydrological characteristics are not obtained from the MSS imagery, the indirect information such as the permeability of the soil and the vegetation cover, is helpful in interpreting the hydrological aspects.

  19. Mapping Van

    NASA Technical Reports Server (NTRS)

    1994-01-01

    A NASA Center for the Commercial Development of Space (CCDS) - developed system for satellite mapping has been commercialized for the first time. Global Visions, Inc. maps an area while driving along a road in a sophisticated mapping van equipped with satellite signal receivers, video cameras and computer systems for collecting and storing mapping data. Data is fed into a computerized geographic information system (GIS). The resulting amps can be used for tax assessment purposes, emergency dispatch vehicles and fleet delivery companies as well as other applications.

  20. Single-shot T2 mapping using overlapping-echo detachment planar imaging and a deep convolutional neural network.

    PubMed

    Cai, Congbo; Wang, Chao; Zeng, Yiqing; Cai, Shuhui; Liang, Dong; Wu, Yawen; Chen, Zhong; Ding, Xinghao; Zhong, Jianhui

    2018-04-24

    An end-to-end deep convolutional neural network (CNN) based on deep residual network (ResNet) was proposed to efficiently reconstruct reliable T 2 mapping from single-shot overlapping-echo detachment (OLED) planar imaging. The training dataset was obtained from simulations that were carried out on SPROM (Simulation with PRoduct Operator Matrix) software developed by our group. The relationship between the original OLED image containing two echo signals and the corresponding T 2 mapping was learned by ResNet training. After the ResNet was trained, it was applied to reconstruct the T 2 mapping from simulation and in vivo human brain data. Although the ResNet was trained entirely on simulated data, the trained network was generalized well to real human brain data. The results from simulation and in vivo human brain experiments show that the proposed method significantly outperforms the echo-detachment-based method. Reliable T 2 mapping with higher accuracy is achieved within 30 ms after the network has been trained, while the echo-detachment-based OLED reconstruction method took approximately 2 min. The proposed method will facilitate real-time dynamic and quantitative MR imaging via OLED sequence, and deep convolutional neural network has the potential to reconstruct maps from complex MRI sequences efficiently. © 2018 International Society for Magnetic Resonance in Medicine.

  1. Smart "geomorphological" map browsing - a tale about geomorphological maps and the internet

    NASA Astrophysics Data System (ADS)

    Geilhausen, M.; Otto, J.-C.

    2012-04-01

    With the digital production of geomorphological maps, the dissemination of research outputs now extends beyond simple paper products. Internet technologies can contribute to both, the dissemination of geomorphological maps and access to geomorphologic data and help to make geomorphological knowledge available to a greater public. Indeed, many national geological surveys employ end-to-end digital workflows from data capture in the field to final map production and dissemination. This paper deals with the potential of web mapping applications and interactive, portable georeferenced PDF maps for the distribution of geomorphological information. Web mapping applications such as Google Maps have become very popular and widespread and increased the interest and access to mapping. They link the Internet with GIS technology and are a common way of presenting dynamic maps online. The GIS processing is performed online and maps are visualised in interactive web viewers characterised by different capabilities such as zooming, panning or adding further thematic layers, with the map refreshed after each task. Depending on the system architecture and the components used, advanced symbology, map overlays from different applications and sources and their integration into a Desktop GIS are possible. This interoperability is achieved through the use of international open standards that include mechanisms for the integration and visualisation of information from multiple sources. The portable document format (PDF) is commonly used for printing and is a standard format that can be processed by many graphic software and printers without loss of information. A GeoPDF enables the sharing of geospatial maps and data in PDF documents. Multiple, independent map frames with individual spatial reference systems are possible within a GeoPDF, for example, for map overlays or insets. Geospatial functionality of a GeoPDF includes scalable map display, layer visibility control, access to attribute

  2. BOREAS Hardcopy Maps

    NASA Technical Reports Server (NTRS)

    Hall, Forrest G. (Editor); Nelson, Elizabeth; Newcomer, Jeffrey A.

    2000-01-01

    Boreal Ecosystem-Atmospheric Study (BOREAS) hardcopy maps are a collection of approximately 1,000 hardcopy maps representing the physical, climatological, and historical attributes of areas covering primarily the Manitoba and Saskatchewan provinces of Canada. These maps were collected by BOREAS Information System (BORIS) and Canada for Remote Sensing (CCRS) staff to provide basic information about site positions, manmade features, topography, geology, hydrology, land cover types, fire history, climate, and soils of the BOREAS study region. These maps are not available for distribution through the BOREAS project but may be used as an on-site resource. Information is provided within this document for individuals who want to order copies of these maps from the original map source. Note that the maps are not contained on the BOREAS CD-ROM set. An inventory listing file is supplied on the CD-ROM to inform users of the maps that are available. This inventory listing is available from the Earth Observing System Data and Information System (EOSDIS) Oak Ridge National Laboratory (ORNL) Distributed Active Archive Center (DAAC). For hardcopies of the individual maps, contact the sources provided.

  3. ActionMap: A web-based software that automates loci assignments to framework maps.

    PubMed

    Albini, Guillaume; Falque, Matthieu; Joets, Johann

    2003-07-01

    Genetic linkage computation may be a repetitive and time consuming task, especially when numerous loci are assigned to a framework map. We thus developed ActionMap, a web-based software that automates genetic mapping on a fixed framework map without adding the new markers to the map. Using this tool, hundreds of loci may be automatically assigned to the framework in a single process. ActionMap was initially developed to map numerous ESTs with a small plant mapping population and is limited to inbred lines and backcrosses. ActionMap is highly configurable and consists of Perl and PHP scripts that automate command steps for the MapMaker program. A set of web forms were designed for data import and mapping settings. Results of automatic mapping can be displayed as tables or drawings of maps and may be exported. The user may create personal access-restricted projects to store raw data, settings and mapping results. All data may be edited, updated or deleted. ActionMap may be used either online or downloaded for free (http://moulon.inra.fr/~bioinfo/).

  4. ActionMap: a web-based software that automates loci assignments to framework maps

    PubMed Central

    Albini, Guillaume; Falque, Matthieu; Joets, Johann

    2003-01-01

    Genetic linkage computation may be a repetitive and time consuming task, especially when numerous loci are assigned to a framework map. We thus developed ActionMap, a web-based software that automates genetic mapping on a fixed framework map without adding the new markers to the map. Using this tool, hundreds of loci may be automatically assigned to the framework in a single process. ActionMap was initially developed to map numerous ESTs with a small plant mapping population and is limited to inbred lines and backcrosses. ActionMap is highly configurable and consists of Perl and PHP scripts that automate command steps for the MapMaker program. A set of web forms were designed for data import and mapping settings. Results of automatic mapping can be displayed as tables or drawings of maps and may be exported. The user may create personal access-restricted projects to store raw data, settings and mapping results. All data may be edited, updated or deleted. ActionMap may be used either online or downloaded for free (http://moulon.inra.fr/~bioinfo/). PMID:12824426

  5. Cartographic mapping study

    NASA Technical Reports Server (NTRS)

    Wilson, C.; Dye, R.; Reed, L.

    1982-01-01

    The errors associated with planimetric mapping of the United States using satellite remote sensing techniques are analyzed. Assumptions concerning the state of the art achievable for satellite mapping systems and platforms in the 1995 time frame are made. An analysis of these performance parameters is made using an interactive cartographic satellite computer model, after first validating the model using LANDSAT 1 through 3 performance parameters. An investigation of current large scale (1:24,000) US National mapping techniques is made. Using the results of this investigation, and current national mapping accuracy standards, the 1995 satellite mapping system is evaluated for its ability to meet US mapping standards for planimetric and topographic mapping at scales of 1:24,000 and smaller.

  6. A filtering approach to edge preserving MAP estimation of images.

    PubMed

    Humphrey, David; Taubman, David

    2011-05-01

    The authors present a computationally efficient technique for maximum a posteriori (MAP) estimation of images in the presence of both blur and noise. The image is divided into statistically independent regions. Each region is modelled with a WSS Gaussian prior. Classical Wiener filter theory is used to generate a set of convex sets in the solution space, with the solution to the MAP estimation problem lying at the intersection of these sets. The proposed algorithm uses an underlying segmentation of the image, and a means of determining the segmentation and refining it are described. The algorithm is suitable for a range of image restoration problems, as it provides a computationally efficient means to deal with the shortcomings of Wiener filtering without sacrificing the computational simplicity of the filtering approach. The algorithm is also of interest from a theoretical viewpoint as it provides a continuum of solutions between Wiener filtering and Inverse filtering depending upon the segmentation used. We do not attempt to show here that the proposed method is the best general approach to the image reconstruction problem. However, related work referenced herein shows excellent performance in the specific problem of demosaicing.

  7. A technology mapping based on graph of excitations and outputs for finite state machines

    NASA Astrophysics Data System (ADS)

    Kania, Dariusz; Kulisz, Józef

    2017-11-01

    A new, efficient technology mapping method of FSMs, dedicated for PAL-based PLDs is proposed. The essence of the method consists in searching for the minimal set of PAL-based logic blocks that cover a set of multiple-output implicants describing the transition and output functions of an FSM. The method is based on a new concept of graph: the Graph of Excitations and Outputs. The proposed algorithm was tested using the FSM benchmarks. The obtained results were compared with the classical technology mapping of FSM.

  8. Genome-wide mapping reveals single-origin chromosome replication in Leishmania, a eukaryotic microbe.

    PubMed

    Marques, Catarina A; Dickens, Nicholas J; Paape, Daniel; Campbell, Samantha J; McCulloch, Richard

    2015-10-19

    DNA replication initiates on defined genome sites, termed origins. Origin usage appears to follow common rules in the eukaryotic organisms examined to date: all chromosomes are replicated from multiple origins, which display variations in firing efficiency and are selected from a larger pool of potential origins. To ask if these features of DNA replication are true of all eukaryotes, we describe genome-wide origin mapping in the parasite Leishmania. Origin mapping in Leishmania suggests a striking divergence in origin usage relative to characterized eukaryotes, since each chromosome appears to be replicated from a single origin. By comparing two species of Leishmania, we find evidence that such origin singularity is maintained in the face of chromosome fusion or fission events during evolution. Mapping Leishmania origins suggests that all origins fire with equal efficiency, and that the genomic sites occupied by origins differ from related non-origins sites. Finally, we provide evidence that origin location in Leishmania displays striking conservation with Trypanosoma brucei, despite the latter parasite replicating its chromosomes from multiple, variable strength origins. The demonstration of chromosome replication for a single origin in Leishmania, a microbial eukaryote, has implications for the evolution of origin multiplicity and associated controls, and may explain the pervasive aneuploidy that characterizes Leishmania chromosome architecture.

  9. Mapping Applications Center, National Mapping Division, U.S. Geological Survey

    USGS Publications Warehouse

    ,

    1996-01-01

    The Mapping Applications Center (MAC), National Mapping Division (NMD), is the eastern regional center for coordinating the production, distribution, and sale of maps and digital products of the U.S. Geological Survey (USGS). It is located in the John Wesley Powell Federal Building in Reston, Va. The MAC's major functions are to (1) establish and manage cooperative mapping programs with State and Federal agencies; (2) perform new research in preparing and applying geospatial information; (3) prepare digital cartographic data, special purpose maps, and standard maps from traditional and classified source materials; (4) maintain the domestic names program of the United States; (5) manage the National Aerial Photography Program (NAPP); (6) coordinate the NMD's publications and outreach programs; and (7) direct the USGS mapprinting operations.

  10. Rigorous Numerical Study of Low-Period Windows for the Quadratic Map

    NASA Astrophysics Data System (ADS)

    Galias, Zbigniew

    An efficient method to find all low-period windows for the quadratic map is proposed. The method is used to obtain very accurate rigorous bounds of positions of all periodic windows with periods p ≤ 32. The contribution of period-doubling windows on the total width of periodic windows is discussed. Properties of periodic windows are studied numerically.

  11. Point-Cloud Compression for Vehicle-Based Mobile Mapping Systems Using Portable Network Graphics

    NASA Astrophysics Data System (ADS)

    Kohira, K.; Masuda, H.

    2017-09-01

    A mobile mapping system is effective for capturing dense point-clouds of roads and roadside objects Point-clouds of urban areas, residential areas, and arterial roads are useful for maintenance of infrastructure, map creation, and automatic driving. However, the data size of point-clouds measured in large areas is enormously large. A large storage capacity is required to store such point-clouds, and heavy loads will be taken on network if point-clouds are transferred through the network. Therefore, it is desirable to reduce data sizes of point-clouds without deterioration of quality. In this research, we propose a novel point-cloud compression method for vehicle-based mobile mapping systems. In our compression method, point-clouds are mapped onto 2D pixels using GPS time and the parameters of the laser scanner. Then, the images are encoded in the Portable Networking Graphics (PNG) format and compressed using the PNG algorithm. In our experiments, our method could efficiently compress point-clouds without deteriorating the quality.

  12. Mapping alteration minerals at prospect, outcrop and drill core scales using imaging spectrometry

    PubMed Central

    Kruse, Fred A.; L. Bedell, Richard; Taranik, James V.; Peppin, William A.; Weatherbee, Oliver; Calvin, Wendy M.

    2011-01-01

    Imaging spectrometer data (also known as ‘hyperspectral imagery’ or HSI) are well established for detailed mineral mapping from airborne and satellite systems. Overhead data, however, have substantial additional potential when used together with ground-based measurements. An imaging spectrometer system was used to acquire airborne measurements and to image in-place outcrops (mine walls) and boxed drill core and rock chips using modified sensor-mounting configurations. Data were acquired at 5 nm nominal spectral resolution in 360 channels from 0.4 to 2.45 μm. Analysis results using standardized hyperspectral methodologies demonstrate rapid extraction of representative mineral spectra and mapping of mineral distributions and abundances in map-plan, with core depth, and on the mine walls. The examples shown highlight the capabilities of these data for mineral mapping. Integration of these approaches promotes improved understanding of relations between geology, alteration and spectral signatures in three dimensions and should lead to improved efficiency of mine development, operations and ultimately effective mine closure. PMID:25937681

  13. Enriching the national map database for multi-scale use: Introducing the visibilityfilter attribution

    USGS Publications Warehouse

    Stauffer, Andrew J.; Webinger, Seth; Roche, Brittany

    2016-01-01

    The US Geological Survey’s (USGS) National Geospatial Technical Operations Center is prototyping and evaluating the ability to filter data through a range of scales using 1:24,000-scale The National Map (TNM) datasets as the source. A “VisibilityFilter” attribute is under evaluation that can be added to all TNM vector data themes and will permit filtering of data to eight target scales between 1:24,000 and 1:5,000,000, thus defining each feature’s smallest applicable scale-of-use. For a prototype implementation, map specifications for 1:100,000- and 1:250,000-scale USGS Topographic Map Series are being utilized to define feature content appropriate at fixed mapping scales to guide generalization decisions that are documented in a ScaleMaster diagram. This paper defines the VisibilityFilter attribute, the generalization decisions made for each TNM data theme, and how these decisions are embedded into the data to support efficient data filtering.

  14. Motor transfer from map ocular exploration to locomotion during spatial navigation from memory.

    PubMed

    Demichelis, Alixia; Olivier, Gérard; Berthoz, Alain

    2013-02-01

    Spatial navigation from memory can rely on two different strategies: a mental simulation of a kinesthetic spatial navigation (egocentric route strategy) or visual-spatial memory using a mental map (allocentric survey strategy). We hypothesized that a previously performed "oculomotor navigation" on a map could be used by the brain to perform a locomotor memory task. Participants were instructed to (1) learn a path on a map through a sequence of vertical and horizontal eyes movements and (2) walk on the slabs of a "magic carpet" to recall this path. The main results showed that the anisotropy of ocular movements (horizontal ones being more efficient than vertical ones) influenced performances of participants when they change direction on the central slab of the magic carpet. These data suggest that, to find their way through locomotor space, subjects mentally repeated their past ocular exploration of the map, and this visuo-motor memory was used as a template for the locomotor performance.

  15. Using concept mapping to design an indicator framework for addiction treatment centres.

    PubMed

    Nabitz, Udo; van Den Brink, Wim; Jansen, Paul

    2005-06-01

    The objective of this study is to determine an indicator framework for addiction treatment centres based on the demands of stakeholders and in alignment with the European Foundation for Quality Management (EFQM) Excellence Model. The setting is the Jellinek Centre based in Amsterdam, the Netherlands, which serves as a prototype for an addiction treatment centre. Concept mapping was used in the construction of the indicator framework. During the 1-day workshop, 16 stakeholders generated, prioritized and sorted 73 items concerning quality and performance. Multidimensional scaling and cluster analysis was applied in constructing a framework consisting of two dimensions and eight clusters. The horizontal axis of the indicator framework is named 'Organization' and has two poles, namely, 'Processes' and 'Results'. The vertical axis is named ' Task' and the poles are named 'Efficient treatment' and 'Prevention programs'. The eight clusters in the two-dimensional framework are arranged in the following, prioritized sequence: 'Efficient treatment network', 'Effective service', ' Target group', 'Quality of life', 'Efficient service', 'Knowledge transfer', 'Reducing addiction related problems', and 'Prevention programs'. The most important items in the framework are: 'patients are satisfied with their treatment', 'early interventions', and 'efficient treatment chain'. The indicator framework aligns with three clusters of the results criteria of the EFQM Excellence Model. It is based on the stakeholders' perspectives and is believed to be specific for addiction treatment centres. The study demonstrates that concept mapping is a suitable strategy for generating indicator frameworks.

  16. Development of an Efficient Protein Extraction Method Compatible with LC-MS/MS for Proteome Mapping in Two Australian Seagrasses Zostera muelleri and Posidonia australis

    PubMed Central

    Jiang, Zhijian; Kumar, Manoj; Padula, Matthew P.; Pernice, Mathieu; Kahlke, Tim; Kim, Mikael; Ralph, Peter J.

    2017-01-01

    The availability of the first complete genome sequence of the marine flowering plant Zostera marina (commonly known as seagrass) in early 2016, is expected to significantly raise the impact of seagrass proteomics. Seagrasses are marine ecosystem engineers that are currently declining worldwide at an alarming rate due to both natural and anthropogenic disturbances. Seagrasses (especially species of the genus Zostera) are compromised for proteomic studies primarily due to the lack of efficient protein extraction methods because of their recalcitrant cell wall which is rich in complex polysaccharides and a high abundance of secondary metabolites in their cells. In the present study, three protein extraction methods that are commonly used in plant proteomics i.e., phenol (P); trichloroacetic acid/acetone/SDS/phenol (TASP); and borax/polyvinyl-polypyrrolidone/phenol (BPP) extraction, were evaluated quantitatively and qualitatively based on two dimensional isoelectric focusing (2D-IEF) maps and LC-MS/MS analysis using the two most abundant Australian seagrass species, namely Zostera muelleri and Posidonia australis. All three tested methods produced high quality protein extracts with excellent 2D-IEF maps in P. australis. However, the BPP method produces better results in Z. muelleri compared to TASP and P. Therefore, we further modified the BPP method (M-BPP) by homogenizing the tissue in a modified protein extraction buffer containing both ionic and non-ionic detergents (0.5% SDS; 1.5% Triton X-100), 2% PVPP and protease inhibitors. Further, the extracted proteins were solubilized in 0.5% of zwitterionic detergent (C7BzO) instead of 4% CHAPS. This slight modification to the BPP method resulted in a higher protein yield, and good quality 2-DE maps with a higher number of protein spots in both the tested seagrasses. Further, the M-BPP method was successfully utilized in western-blot analysis of phosphoenolpyruvate carboxylase (PEPC—a key enzyme for carbon metabolism

  17. Mapping or Tracing? Rethinking Curriculum Mapping in Higher Education

    ERIC Educational Resources Information Center

    Wang, Chia-Ling

    2015-01-01

    Curriculum mapping has been emphasized in recent curriculum innovations in higher education in the drive for global competitiveness. This paper begins by providing an outline of current discourses of curriculum mapping in higher education. Curriculum mapping is frequently associated with outcome-based learning and work readiness, and guiding the…

  18. Linkage disequilibrium interval mapping of quantitative trait loci.

    PubMed

    Boitard, Simon; Abdallah, Jihad; de Rochambeau, Hubert; Cierco-Ayrolles, Christine; Mangin, Brigitte

    2006-03-16

    For many years gene mapping studies have been performed through linkage analyses based on pedigree data. Recently, linkage disequilibrium methods based on unrelated individuals have been advocated as powerful tools to refine estimates of gene location. Many strategies have been proposed to deal with simply inherited disease traits. However, locating quantitative trait loci is statistically more challenging and considerable research is needed to provide robust and computationally efficient methods. Under a three-locus Wright-Fisher model, we derived approximate expressions for the expected haplotype frequencies in a population. We considered haplotypes comprising one trait locus and two flanking markers. Using these theoretical expressions, we built a likelihood-maximization method, called HAPim, for estimating the location of a quantitative trait locus. For each postulated position, the method only requires information from the two flanking markers. Over a wide range of simulation scenarios it was found to be more accurate than a two-marker composite likelihood method. It also performed as well as identity by descent methods, whilst being valuable in a wider range of populations. Our method makes efficient use of marker information, and can be valuable for fine mapping purposes. Its performance is increased if multiallelic markers are available. Several improvements can be developed to account for more complex evolution scenarios or provide robust confidence intervals for the location estimates.

  19. A first generation BAC-based physical map of the rainbow trout genome

    PubMed Central

    Palti, Yniv; Luo, Ming-Cheng; Hu, Yuqin; Genet, Carine; You, Frank M; Vallejo, Roger L; Thorgaard, Gary H; Wheeler, Paul A; Rexroad, Caird E

    2009-01-01

    Background Rainbow trout (Oncorhynchus mykiss) are the most-widely cultivated cold freshwater fish in the world and an important model species for many research areas. Coupling great interest in this species as a research model with the need for genetic improvement of aquaculture production efficiency traits justifies the continued development of genomics research resources. Many quantitative trait loci (QTL) have been identified for production and life-history traits in rainbow trout. A bacterial artificial chromosome (BAC) physical map is needed to facilitate fine mapping of QTL and the selection of positional candidate genes for incorporation in marker-assisted selection (MAS) for improving rainbow trout aquaculture production. This resource will also facilitate efforts to obtain and assemble a whole-genome reference sequence for this species. Results The physical map was constructed from DNA fingerprinting of 192,096 BAC clones using the 4-color high-information content fingerprinting (HICF) method. The clones were assembled into physical map contigs using the finger-printing contig (FPC) program. The map is composed of 4,173 contigs and 9,379 singletons. The total number of unique fingerprinting fragments (consensus bands) in contigs is 1,185,157, which corresponds to an estimated physical length of 2.0 Gb. The map assembly was validated by 1) comparison with probe hybridization results and agarose gel fingerprinting contigs; and 2) anchoring large contigs to the microsatellite-based genetic linkage map. Conclusion The production and validation of the first BAC physical map of the rainbow trout genome is described in this paper. We are currently integrating this map with the NCCCWA genetic map using more than 200 microsatellites isolated from BAC end sequences and by identifying BACs that harbor more than 300 previously mapped markers. The availability of an integrated physical and genetic map will enable detailed comparative genome analyses, fine mapping of

  20. Exploring the dusty star-formation in the early Universe using intensity mapping

    NASA Astrophysics Data System (ADS)

    Lagache, Guilaine

    2018-05-01

    In the last decade, it has become clear that the dust-enshrouded star formation contributes significantly to early galaxy evolution. Detection of dust is therefore essential in determining the properties of galaxies in the high-redshift universe. This requires observations at the (sub-)millimeter wavelengths. Unfortunately, sensitivity and background confusion of single dish observations on the one hand, and mapping efficiency of interferometers on the other hand, pose unique challenges to observers. One promising route to overcome these difficulties is intensity mapping of fluctuations which exploits the confusion-limited regime and measures the collective light emission from all sources, including unresolved faint galaxies. We discuss in this contribution how 2D and 3D intensity mapping can measure the dusty star formation at high redshift, through the Cosmic Infrared Background (2D) and [CII] fine structure transition (3D) anisotropies.

  1. Simian virus 40 major late promoter: an upstream DNA sequence required for efficient in vitro transcription.

    PubMed Central

    Brady, J; Radonovich, M; Thoren, M; Das, G; Salzman, N P

    1984-01-01

    We have previously identified an 11-base DNA sequence, 5'-G-G-T-A-C-C-T-A-A-C-C-3' (simian virus 40 [SV40] map position 294 to 304), which is important in the control of SV40 late RNA expression in vitro and in vivo (Brady et al., Cell 31:625-633, 1982). We report here the identification of another domain of the SV40 late promoter. A series of mutants with deletions extending from SV40 map position 0 to 300 was prepared by nuclease BAL 31 treatment. The cloned templates were then analyzed for efficiency and accuracy of late SV40 RNA expression in the Manley in vitro transcription system. Our studies showed that, in addition to the promoter domain near map position 300, there are essential DNA sequences between nucleotide positions 74 and 95 that are required for efficient expression of late SV40 RNA. Included in this SV40 DNA sequence were two of the six GGGCGG SV40 repeat sequences and an 11-nucleotide segment which showed strong homology with the upstream sequences required for the efficient in vitro and in vivo expression of the histone H2A gene. This upstream promoter sequence supported transcription with the same efficiency even when it was moved 72 nucleotides closer to the major late cap site. In vitro promoter competition analysis demonstrated that the upstream promoter sequence, independent of the 294 to 304 promoter element, is capable of binding polymerase-transcription factors required for SV40 late gene transcription. Finally, we show that DNA sequences which control the specificity of RNA initiation at nucleotide 325 lie downstream of map position 294. Images PMID:6321950

  2. Map Separates

    USGS Publications Warehouse

    ,

    2001-01-01

    U.S. Geological Survey (USGS) topographic maps are printed using up to six colors (black, blue, green, red, brown, and purple). To prepare your own maps or artwork based on maps, you can order separate black-and-white film positives or negatives for any color printed on a USGS topographic map, or for one or more of the groups of related features printed in the same color on the map (such as drainage and drainage names from the blue plate.) In this document, examples are shown with appropriate ink color to illustrate the various separates. When purchased, separates are black-and-white film negatives or positives. After you receive a film separate or composite from the USGS, you can crop, enlarge or reduce, and edit to add or remove details to suit your special needs. For example, you can adapt the separates for making regional and local planning maps or for doing many kinds of studies or promotions by using the features you select and then printing them in colors of your choice.

  3. Reformulating Constraints for Compilability and Efficiency

    NASA Technical Reports Server (NTRS)

    Tong, Chris; Braudaway, Wesley; Mohan, Sunil; Voigt, Kerstin

    1992-01-01

    KBSDE is a knowledge compiler that uses a classification-based approach to map solution constraints in a task specification onto particular search algorithm components that will be responsible for satisfying those constraints (e.g., local constraints are incorporated in generators; global constraints are incorporated in either testers or hillclimbing patchers). Associated with each type of search algorithm component is a subcompiler that specializes in mapping constraints into components of that type. Each of these subcompilers in turn uses a classification-based approach, matching a constraint passed to it against one of several schemas, and applying a compilation technique associated with that schema. While much progress has occurred in our research since we first laid out our classification-based approach [Ton91], we focus in this paper on our reformulation research. Two important reformulation issues that arise out of the choice of a schema-based approach are: (1) compilability-- Can a constraint that does not directly match any of a particular subcompiler's schemas be reformulated into one that does? and (2) Efficiency-- If the efficiency of the compiled search algorithm depends on the compiler's performance, and the compiler's performance depends on the form in which the constraint was expressed, can we find forms for constraints which compile better, or reformulate constraints whose forms can be recognized as ones that compile poorly? In this paper, we describe a set of techniques we are developing for partially addressing these issues.

  4. Snow water equivalent mapping in Norway

    NASA Astrophysics Data System (ADS)

    Tveito, O. E.; Udnæs, H.-C.; Engeset, R.; Førland, E. J.; Isaksen, K.; Mengistu, Z.

    2003-04-01

    In high latitude area snow covers the ground large parts of the year. Information about the water volume as snow is of major importance in many respects. Flood forecasters at NVE need it in order to assess possible flood risks. Hydropower producers need it to plan the most efficient production of the water in their reservoirs, traders to estimate the potential energy available for the market. Meteorologists on their side use the information as boundary conditions in weather forecasting models. The Norwegian meteorological institute has provided snow accumulation maps for Norway for more than 50 years. These maps are now produced twice a month in the winter season. They show the accumulated precipitation in the winter season from the day the permanent snow cover is established. They do however not take melting into account, and do therefore not give a good description of the actual snow amounts during and after periods with snowmelt. Due to an increased need for a direct measure of water volumes as snow cover, met.no and NVE initialized a joint project in order to establish maps of the actual snow cover expressed in water equivalents. The project utilizes recent developments in the use of GIS in spatial modeling. Daily precipitation and temperature are distributed in space by using objective spatial interpolation methods. The interpolation considers topographical and other geographical parameters as well as weather type information. A degree-day model is used at each modeling point to calculate snow-accumulation and snowmelt. The maps represent a spatial scale of 1x1 km2. The modeled snow reservoir is validated by snow pillow values as well traditional snow depth observations. Preliminary results show that the new snow modeling approach reproduces the snow water equivalent well. The spatial approach also opens for a wide use in the terms of areal analysis.

  5. Big Data: A Parallel Particle Swarm Optimization-Back-Propagation Neural Network Algorithm Based on MapReduce.

    PubMed

    Cao, Jianfang; Cui, Hongyan; Shi, Hao; Jiao, Lijuan

    2016-01-01

    A back-propagation (BP) neural network can solve complicated random nonlinear mapping problems; therefore, it can be applied to a wide range of problems. However, as the sample size increases, the time required to train BP neural networks becomes lengthy. Moreover, the classification accuracy decreases as well. To improve the classification accuracy and runtime efficiency of the BP neural network algorithm, we proposed a parallel design and realization method for a particle swarm optimization (PSO)-optimized BP neural network based on MapReduce on the Hadoop platform using both the PSO algorithm and a parallel design. The PSO algorithm was used to optimize the BP neural network's initial weights and thresholds and improve the accuracy of the classification algorithm. The MapReduce parallel programming model was utilized to achieve parallel processing of the BP algorithm, thereby solving the problems of hardware and communication overhead when the BP neural network addresses big data. Datasets on 5 different scales were constructed using the scene image library from the SUN Database. The classification accuracy of the parallel PSO-BP neural network algorithm is approximately 92%, and the system efficiency is approximately 0.85, which presents obvious advantages when processing big data. The algorithm proposed in this study demonstrated both higher classification accuracy and improved time efficiency, which represents a significant improvement obtained from applying parallel processing to an intelligent algorithm on big data.

  6. Digital Mapping Techniques '08—Workshop Proceedings, Moscow, Idaho, May 18–21, 2008

    USGS Publications Warehouse

    Soller, David R.

    2009-01-01

    The Digital Mapping Techniques '08 (DMT'08) workshop was attended by more than 100 technical experts from 40 agencies, universities, and private companies, including representatives from 24 State geological surveys. This year's meeting, the twelfth in the annual series, was hosted by the Idaho Geological Survey, from May 18-21, 2008, on the University of Idaho campus in Moscow, Idaho. Each DMT workshop has been coordinated by the U.S. Geological Survey's National Geologic Map Database Project and the Association of American State Geologists (AASG). As in previous years' meetings, the objective was to foster informal discussion and exchange of technical information, principally in order to develop more efficient methods for digital mapping, cartography, GIS analysis, and information management. At this meeting, oral and poster presentations and special discussion sessions emphasized (1) methods for creating and publishing map products (here, "publishing" includes Web-based release); (2) field data capture software and techniques, including the use of LiDAR; (3) digital cartographic techniques; (4) migration of digital maps into ArcGIS Geodatabase format; (5) analytical GIS techniques; and (6) continued development of the National Geologic Map Database.

  7. Validation of a novel mapping system and utility for mapping complex atrial tachycardias.

    PubMed

    Honarbakhsh, S; Hunter, R J; Dhillon, G; Ullah, W; Keating, E; Providencia, R; Chow, A; Earley, M J; Schilling, R J

    2018-03-01

    This study sought to validate a novel wavefront mapping system utilizing whole-chamber basket catheters (CARTOFINDER, Biosense Webster). The system was validated in terms of (1) mapping atrial-paced beats and (2) mapping complex wavefront patterns in atrial tachycardia (AT). Patients undergoing catheter ablation for AT and persistent AF were included. A 64-pole-basket catheter was used to acquire unipolar signals that were processed by CARTOFINDER mapping system to generate dynamic wavefront propagation maps. The left atrium was paced from four sites to demonstrate focal activation. ATs were mapped with the mechanism confirmed by conventional mapping, entrainment, and response to ablation. Twenty-two patients were included in the study (16 with AT and 6 with AF initially who terminated to AT during ablation). In total, 172 maps were created with the mapping system. It correctly identified atrial-pacing sites in all paced maps. It accurately mapped 9 focal/microreentrant and 18 macroreentrant ATs both in the left and right atrium. A third and fourth observer independently identified the sites of atrial pacing and the AT mechanism from the CARTOFINDER maps, while being blinded to the conventional activation maps. This novel mapping system was effectively validated by mapping focal activation patterns from atrial-paced beats. The system was also effective in mapping complex wavefront patterns in a range of ATs in patients with scarred atria. The system may therefore be of practical use in the mapping and ablation of AT and could have potential for mapping wavefront activations in AF. © 2018 Wiley Periodicals, Inc.

  8. Comparative visualization of genetic and physical maps with Strudel.

    PubMed

    Bayer, Micha; Milne, Iain; Stephen, Gordon; Shaw, Paul; Cardle, Linda; Wright, Frank; Marshall, David

    2011-05-01

    Data visualization can play a key role in comparative genomics, for example, underpinning the investigation of conserved synteny patterns. Strudel is a desktop application that allows users to easily compare both genetic and physical maps interactively and efficiently. It can handle large datasets from several genomes simultaneously, and allows all-by-all comparisons between these. Installers for Strudel are available for Windows, Linux, Solaris and Mac OS X at http://bioinf.scri.ac.uk/strudel/.

  9. Efficient nonparametric n -body force fields from machine learning

    NASA Astrophysics Data System (ADS)

    Glielmo, Aldo; Zeni, Claudio; De Vita, Alessandro

    2018-05-01

    We provide a definition and explicit expressions for n -body Gaussian process (GP) kernels, which can learn any interatomic interaction occurring in a physical system, up to n -body contributions, for any value of n . The series is complete, as it can be shown that the "universal approximator" squared exponential kernel can be written as a sum of n -body kernels. These recipes enable the choice of optimally efficient force models for each target system, as confirmed by extensive testing on various materials. We furthermore describe how the n -body kernels can be "mapped" on equivalent representations that provide database-size-independent predictions and are thus crucially more efficient. We explicitly carry out this mapping procedure for the first nontrivial (three-body) kernel of the series, and we show that this reproduces the GP-predicted forces with meV /Å accuracy while being orders of magnitude faster. These results pave the way to using novel force models (here named "M-FFs") that are computationally as fast as their corresponding standard parametrized n -body force fields, while retaining the nonparametric character, the ease of training and validation, and the accuracy of the best recently proposed machine-learning potentials.

  10. Highly Efficient, All-Dielectric Huygens Metasurfaces

    NASA Astrophysics Data System (ADS)

    Ollanik, Adam; Farrar-Foley, Nick; Smith, Jake; Escarra, Matthew

    Demonstration of the control of light by the introduction of abrupt phase discontinuities across a subwavelength scale has opened the doors to a new level of wavefront control. All-dielectric Huygens metasurfaces hold significant promise due to their dramatically improved efficiency over plasmonic approaches. We present the successful design, computational modeling, and experimental realization of all-dielectric transmissive Huygens metasurfaces capable of deflection efficiency >90%. Dielectric Huygens sources, taking advantage of spectrally aligned electric and magnetic dipole resonances, are capable of tunable phase delay for transmitted light with near unity efficiency of forward scattering. Using ellipsoidal cylinder nanoantennas, we are able to manipulate the phase response and engineer a metasurface with a spatially gradient phase profile. Through careful design and optimization we mitigate the effects of inter-antenna coupling. We have designed and modeled metasurfaces demonstrating anomalous refraction with very high efficiency (>80%) for wavelength bands from the UV to the near-IR. These surfaces were designed using three distinct nanoantenna materials, Si, TiO2, and GaP, to demonstrate the flexibility of the technique. Experimentally, Si nanoantennas are fabricated using a combination of electron beam lithography and ICP/RIE-etching. Metasurfaces are characterized using a goniospectrometer capable of mapping light intensity on a cylindrical shell surrounding the metasurface.

  11. Compositional cokriging for mapping the probability risk of groundwater contamination by nitrates.

    PubMed

    Pardo-Igúzquiza, Eulogio; Chica-Olmo, Mario; Luque-Espinar, Juan A; Rodríguez-Galiano, Víctor

    2015-11-01

    Contamination by nitrates is an important cause of groundwater pollution and represents a potential risk to human health. Management decisions must be made using probability maps that assess the nitrate concentration potential of exceeding regulatory thresholds. However these maps are obtained with only a small number of sparse monitoring locations where the nitrate concentrations have been measured. It is therefore of great interest to have an efficient methodology for obtaining those probability maps. In this paper, we make use of the fact that the discrete probability density function is a compositional variable. The spatial discrete probability density function is estimated by compositional cokriging. There are several advantages in using this approach: (i) problems of classical indicator cokriging, like estimates outside the interval (0,1) and order relations, are avoided; (ii) secondary variables (e.g. aquifer parameters) can be included in the estimation of the probability maps; (iii) uncertainty maps of the probability maps can be obtained; (iv) finally there are modelling advantages because the variograms and cross-variograms of real variables that do not have the restrictions of indicator variograms and indicator cross-variograms. The methodology was applied to the Vega de Granada aquifer in Southern Spain and the advantages of the compositional cokriging approach were demonstrated. Copyright © 2015 Elsevier B.V. All rights reserved.

  12. 14. Photocopy of 1872 photograph by Eadweard Muybridge in Stanford ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    14. Photocopy of 1872 photograph by Eadweard Muybridge in Stanford University Archives, PC 6. SEWING ROOM ('BIRD ROOM').LEFT TO RIGHT, ANNA MARIA LATHROP (MRS. STANFORD'S SISTER), MRS. JANE ANN (DYER) LATHROP (MRS. STANFORD'S MOTHER), ELIZABETH PHILLIPS (MRS. JOSIAH) STANFORD (GOV. STANFORD'S MOTHER), JANE LATHROP (MRS. LELAND) STANFORD AND HER SON, LELAND, JR. - Leland Stanford House, 800 N Street, Sacramento, Sacramento County, CA

  13. Home and Away: A Schoolmistress in Lowland Scotland and Colonial Australia in the Second Half of the Nineteenth Century

    ERIC Educational Resources Information Center

    McDermid, Jane

    2011-01-01

    In this paper, the author discusses the life of Jane Hay Brown, later Hamilton (1827-1898), who worked as a governess and schoolmistress from the late 1840s to the mid 1880s. She was a woman whose life would have remained largely unknown without emigration which resulted in a rich collection of family letters. Jane's letters provide insight into…

  14. Colombia: A Country Under Constant Threat of Disasters

    DTIC Science & Technology

    2014-05-22

    disasters strike every nation in the world , and although these events do not occur with uniformity of distribution, developing nations suffer the greatest...strike every nation in the world , and although these events do not occur with uniformity of distribution, developing nations suffer the greatest...have been victims 4IHS Janes, “Jane’s World Insurgency and Terrorism.” Fuerzas Armadas

  15. Estimating floodwater depths from flood inundation maps and topography

    USGS Publications Warehouse

    Cohen, Sagy; Brakenridge, G. Robert; Kettner, Albert; Bates, Bradford; Nelson, Jonathan M.; McDonald, Richard R.; Huang, Yu-Fen; Munasinghe, Dinuke; Zhang, Jiaqi

    2018-01-01

    Information on flood inundation extent is important for understanding societal exposure, water storage volumes, flood wave attenuation, future flood hazard, and other variables. A number of organizations now provide flood inundation maps based on satellite remote sensing. These data products can efficiently and accurately provide the areal extent of a flood event, but do not provide floodwater depth, an important attribute for first responders and damage assessment. Here we present a new methodology and a GIS-based tool, the Floodwater Depth Estimation Tool (FwDET), for estimating floodwater depth based solely on an inundation map and a digital elevation model (DEM). We compare the FwDET results against water depth maps derived from hydraulic simulation of two flood events, a large-scale event for which we use medium resolution input layer (10 m) and a small-scale event for which we use a high-resolution (LiDAR; 1 m) input. Further testing is performed for two inundation maps with a number of challenging features that include a narrow valley, a large reservoir, and an urban setting. The results show FwDET can accurately calculate floodwater depth for diverse flooding scenarios but also leads to considerable bias in locations where the inundation extent does not align well with the DEM. In these locations, manual adjustment or higher spatial resolution input is required.

  16. Gradient Magnitude Similarity Deviation: A Highly Efficient Perceptual Image Quality Index.

    PubMed

    Xue, Wufeng; Zhang, Lei; Mou, Xuanqin; Bovik, Alan C

    2014-02-01

    It is an important task to faithfully evaluate the perceptual quality of output images in many applications, such as image compression, image restoration, and multimedia streaming. A good image quality assessment (IQA) model should not only deliver high quality prediction accuracy, but also be computationally efficient. The efficiency of IQA metrics is becoming particularly important due to the increasing proliferation of high-volume visual data in high-speed networks. We present a new effective and efficient IQA model, called gradient magnitude similarity deviation (GMSD). The image gradients are sensitive to image distortions, while different local structures in a distorted image suffer different degrees of degradations. This motivates us to explore the use of global variation of gradient based local quality map for overall image quality prediction. We find that the pixel-wise gradient magnitude similarity (GMS) between the reference and distorted images combined with a novel pooling strategy-the standard deviation of the GMS map-can predict accurately perceptual image quality. The resulting GMSD algorithm is much faster than most state-of-the-art IQA methods, and delivers highly competitive prediction accuracy. MATLAB source code of GMSD can be downloaded at http://www4.comp.polyu.edu.hk/~cslzhang/IQA/GMSD/GMSD.htm.

  17. Women and Chemistry in Regency England: New Light on the Marcet Circle.

    PubMed

    Leigh, G Jeffery; Rocke, Alan J

    2016-02-01

    Jane Marcet's Conversations on Chemistry (first edition, 1806) was possibly the best-selling English-language chemistry book of the first half of the nineteenth century. Recent scholarship has explored the degree to which her husband assisted in the writing of the book, without diminishing the high merits of the author. Previously unpublished correspondence, some of which appears here for the first time, casts new light on the social and professional circle of Jane and Alexander Marcet, including its influence on Jane's book. One of the members of that circle was a hitherto unrecognised but highly capable young female chemist, Frederica Sebright. The story told here underlines the tensions in elite circles in early nineteenth-century England between broad-minded acceptance and patronising limitations for women in science.

  18. Zirconium phosphatidylcholine-based nanocapsules as an in vivo degradable drug delivery system of MAP30, a momordica anti-HIV protein.

    PubMed

    Caizhen, Guo; Yan, Gao; Ronron, Chang; Lirong, Yang; Panpan, Chu; Xuemei, Hu; Yuanbiao, Qiao; Qingshan, Li

    2015-04-10

    An essential in vivo drug delivery system of a momordica anti-HIV protein, MAP30, was developed through encapsulating in chemically synthesized matrices of zirconium egg- and soy-phosphatidylcholines, abbreviated to Zr/EPC and Zr/SPC, respectively. Matrices were characterized by transmission electron microscopy and powder X-ray diffractometry studies. Zr/EPC granule at an approximate diameter of 69.43±7.78 nm was a less efficient encapsulator than the granule of Zr/SPC. Interlayer spacing of the matrices encapsulating MAP30 increased from 8.8 and 9.7 Å to 7.4 and 7.9 nm, respectively. In vivo kinetics on degradation and protein release was performed by analyzing the serum sampling of intravenously injected SPF chickens. The first order and biphasic variations were obtained for in vivo kinetics using equilibrium dialysis. Antimicrobial and anti-HIV assays yielded greatly decreased MIC50 and EC50 values of nanoformulated MAP30. An acute toxicity of MAP30 encapsulated in Zr/EPC occurred at a single intravenous dose above 14.24 mg/kg bw in NIH/KM/ICR mice. The folding of MAP30 from Zr/EPC sustained in vivo chickens for more than 8 days in high performance liquid chromatography assays. These matrices could protect MAP30 efficiently with strong structure retention, lowered toxicity and prolonged in vivo life. Copyright © 2015 Elsevier B.V. All rights reserved.

  19. A Servicewide Benthic Mapping Program for National Parks

    USGS Publications Warehouse

    Moses, Christopher S.; Nayegandhi, Amar; Beavers, Rebecca; Brock, John

    2010-01-01

    Classification Standard (CMECS) that is being modified to include all NPS needs, such as lacustrine ecosystems and submerged cultural resources. CMECS Version III (Madden and others, 2010) includes components for water column, biotic cover, surface geology, sub-benthic, and geoform. SBMP Data Archiving. The SBMP calls for the storage of all raw data and final products in common-use data formats. The concept of 'collect once, use often' is essential to efficient use of mapping resources. Data should also be shared with other agencies and the public through various digital clearing houses, such as Geospatial One-Stop (http://gos2.geodata.gov/wps/portal/gos). To be most useful for managing submerged resources, the SBMP advocates the inventory and mapping of the five components of marine ecosystems: surface geology, biotic cover, geoform, sub-benthic, and water column. A complete benthic inventory of a park would include maps of bathymetry and the five components of CMECS. The completion of mapping for any set of components, such as bathymetry and surface geology, or a particular theme (for example, submerged aquatic vegetation) should also include a printed report.

  20. Crystallization efficiencies of inorganic polyphosphate oligomers reacted with magnesium and calcium cations using anion-exchange chromatography with particulate formation-laser scattering detector.

    PubMed

    Ando, Masaki; Imadzu, Sakiyo; Kitagawa, Shinya; Ohtani, Hajime

    2010-08-06

    A particulate formation-laser scattering detector (PFLSD) was developed and used for evaluating the crystallization efficiency of inorganic polyphosphates (PPs) that reacted with either magnesium or calcium cations. As the solutions for reactive crystallization, 0.5 M ammonium buffer (pH 9.6) containing either 0.15 M MgCl(2) or 0.15 M CaCl(2) (MAP: magnesium ammonium phosphate and HAP: hydroxyapatite solution) were used. In the case of mono- and diphosphate (P1 and P2), the significant dependences of the particulate formation efficiency on various types of both P1/P2 and MAP/HAP reaction solutions were observed with the direct sample injection mode. The PFLSD was hyphenated with the anion-exchange chromatography and the dependence of the particulate formation efficiency on the polymerization degree (n(p)) of PP oligomers, separated chromatographically, was evaluated sequentially. The significant suppression of the particulate formation for PP oligomers was clearly confirmed, i.e., the MAP and HAP reaction solutions did not produce the particulates of the PP oligomers having an n(p) value of more than 3 and 5, respectively. As the overall tendency, the particulate formation efficiency in the case of the HAP solution was superior to that in the case of the MAP solution. Copyright 2010 Elsevier B.V. All rights reserved.