Science.gov

Sample records for 38th parallel lineament

  1. Culminating Point and the 38th Parallel

    DTIC Science & Technology

    1994-01-01

    interrelationship among the defense , the campaign plan and the culminating point. The Korean War illustrates his point in almost textbook fashion. The natural... defense . His talk of a "buffer zone" at the Chinese- Korean border was bombastic and superficial. His purpose was to totally destroy the NKPA and unify...the country of Korea . He never did transition to defense . Without imposing restrictions on recrossing the Parallel, Ridgeway’s purpose for the UN

  2. Tectonic study of the extension of the New Madrid fault zone near its intersection with the 38th parallel lineament

    SciTech Connect

    Braile, L.W.; Hinze, W.J.; Sexton, J.L.; Keller, G.R.; Lidiak, E.G.

    1982-06-01

    Gravity, magnetic, geologic, and seismicity data have been combined in a seismotectonic analysis of the New Madrid seismic zone. Previous studies have presented evidence for several rift zones in this area (Upper Mississippi enmbayment), including the Reelfoot rift, a late precambrian-early Paleozoic failed arm which extends north-northeast from the ancient continental margin. We suggest that the northern terminus of the Reelfoot rift forms a rift complex, with arms extending northeast into southwestern Indiana, northwest along the Mississippi River, and east into western Kentucky, which appears to correlate well with the seismicity in the area. This correlation suggests that faults associated with this rift complex are being reactivated in the contemporary stress field (east-northeast compression). If this interpretation is valid, it represents a seismotectonic model which can be used to predict the extent of future seismicity in the New Madrid seismic zone. The proposed rift complex also provides a coherent model for the tectonic development of this region of the North American midcontinent.

  3. 38th Aerospace Mechanisms Symposium

    NASA Technical Reports Server (NTRS)

    Boesiger, Edward A. (Compiler)

    2006-01-01

    The Aerospace Mechanisms Symposium (AMS) provides a unique forum for those active in the design, production and use of aerospace mechanisms. A major focus is the reporting of problems and solutions associated with the development and flight certification of new mechanisms. Organized by the Mechanisms Education Association, the National Aeronautics and Space Administration and Lockheed Martin Space Systems Company (LMSSC) share the responsibility for hosting the AMS. Now in its 38th symposium, the AMS continues to be well attended, attracting participants from both the U.S. and abroad. The 38th AMs, hosted by the NASA Langley Research Center in Williamsburg, Virginia, was held May 17-19, 2006. During these three days, 34 papers were presented. Topics included gimbals, tribology, actuators, aircraft mechanisms, deployment mechanisms, release mechanisms, and test equipment. Hardware displays during the supplier exhibit gave attendees an opportunity to meet with developers of current and future mechanism components.

  4. 38th Annual Maintenance & Operations Cost Study for Schools

    ERIC Educational Resources Information Center

    Agron, Joe

    2009-01-01

    Despite the worst economic environment in generations, spending by K-12 institutions on maintenance and operations (M&O) held its own--defying historical trends that have shown M&O spending among the most affected in times of budget tightening. This article presents data from the 38th annual Maintenance & Operations Cost Study for…

  5. Fracture-correlated lineaments at Great Bay, southeastern New Hampshire

    USGS Publications Warehouse

    Degnan, James R.; Clark, Stewart F.

    2002-01-01

    Analysis by remote-sensing techniques and observations of exposed bedrock structure were preliminary steps taken in a study to locate potential bedrock-fracture zones that may store and transmit ground water near Great Bay, N.H. To help correlate lineaments on the surface with fractures, structural measurements were made at exposed bedrock, largely along the shoreline of the bay, and analyzed to identify fracture trends and fracture characteristics. With these fracture data, lineament-filtering techniques, such as (1) buffer analysis around individual lineaments, (2) discrete-measurement analysis by domain, and (3) spacing-normalized analysis by domain, identified 'fracture-correlated lineaments.' Of the 927 lineaments identified in the study area (180 square kilometers), 406 (44 percent) were evaluated because they either were located within 305 meters of an outcrop with fracture data or intersected one of five 3,300-meter-square grid domain cells that encompassed the fracture data. Of the 406 lineaments, 190 (47 percent) are fracture correlated, although only 15 percent were correlated by more than one filtering technique. The large number of lineaments found in areas of thin glacial overburden and high densities of fractured outcrops suggests that filtering techniques are useful in these areas to selectively identify fracture-correlated lineaments. Fractures parallel to bedding in the Kittery Formation are open locally and often associated with vugs, with up to 1-centimeter aperture, and may provide appreciable secondary porosity in this rock unit. Discrete-measurement analysis by domain identified fracture-correlated lineaments with orientations parallel to these open and vug-filled fractures. Fracture-correlated lineaments related to closely spaced fractures were identified by the spacing-normalized analysis by domain. Analysis results may be used to indicate the potential bedrock pathways for ground-water-discharge points along the shoreline of Great Bay.

  6. Archuleta County CO Lineaments

    DOE Data Explorer

    Zehner, Richard E.

    2012-01-01

    Citation Information: Originator: Earth Science &Observation Center (ESOC), CIRES, University of Colorado at Boulder Originator: Geothermal Development Associates, Reno, Nevada Publication Date: 2012 Title: Archuleta Lineaments Edition: First Publication Information: Publication Place: Reno Nevada Publisher: Geothermal Development Associates, Reno, Nevada Description: This layer traces apparent topographic and air-photo lineaments in the area around Pagosa springs in Archuleta County, Colorado. It was made in order to identify possible fault and fracture systems that might be conduits for geothermal fluids. Geothermal fluids commonly utilize fault and fractures in competent rocks as conduits for fluid flow. Geothermal exploration involves finding areas of high near-surface temperature gradients, along with a suitable “plumbing system” that can provide the necessary permeability. Geothermal power plants can sometimes be built where temperature and flow rates are high. To do this, georeferenced topographic maps and aerial photographs were utilized in an existing GIS, using ESRI ArcMap 10.0 software. The USA_Topo_Maps and World_Imagery map layers were chosen from the GIS Server at server.arcgisonline.com, using a UTM Zone 13 NAD27 projection. This line shapefile was then constructed over that which appeared to be through-going structural lineaments in both the aerial photographs and topographic layers, taking care to avoid manmade features such as roads, fence lines, and right-of-ways. These lineaments may be displaced somewhat from their actual location, due to such factors as shadow effects with low sun angles in the aerial photographs. Note: This shape file was constructed as an aid to geothermal exploration in preparation for a site visit for field checking. We make no claims as to the existence of the lineaments, their location, orientation, and nature. Spatial Domain: Extent: Top: 4132831.990103 m Left: 311979.997741 m Right: 331678.289280 m Bottom: 4116067

  7. Lineaments on Skylab photographs: Detection, mapping, and hydrologic significance in central Tennessee

    NASA Technical Reports Server (NTRS)

    Moore, G. K.

    1976-01-01

    An investigation was carried out to determine the feasibility of mapping lineaments on SKYLAB photographs of central Tennessee and to determine the hydrologic significance of these lineaments, particularly as concerns the occurrence and productivity of ground water. Sixty-nine percent more lineaments were found on SKYLAB photographs by stereo viewing than by projection viewing, but longer lineaments were detected by projection viewing. Most SKYLAB lineaments consisted of topographic depressions and they followed or paralleled the streams. The remainder were found by vegetation alinements and the straight sides of ridges. Test drilling showed that the median yield of wells located on SKYLAB lineaments were about six times the median yield of wells located by random drilling. The best single detection method, in terms of potential savings, was stereo viewing. Larger savings might be achieved by locating wells on lineaments detected by both stereo viewing and projection.

  8. Lineaments and Mineral Occurrences in Pennsylvania

    NASA Technical Reports Server (NTRS)

    Mcmurtry, G. J.; Petersen, G. W. (Principal Investigator); Kowalik, W. S.; Gold, D. P.

    1975-01-01

    The author has identified the following significant results. A conservative lineament map of Pennsylvania interpreted from ERTS-1 channel 7 (infrared) imagery and Skylab photography was compared with the distribution of known metallic mines and mineral occurrences. Of 383 known mineral occurrences, 116 show a geographical association to 1 km wide lineaments, another 24 lie at the intersection of two lineaments, and one lies at the intersection of three lineaments. The Perkiomen Creek lineament in the Triassic Basin is associated with 9 Cu-Fe occurrences. Six Pb-Zn occurrences are associated with the Tyrone-Mount Union lineament. Thirteen other lineaments are associated with 3, 4, or 5 mineral occurrences each.

  9. 38th JANNAF Combustion Subcommittee Meeting. Volume 1

    NASA Technical Reports Server (NTRS)

    Fry, Ronald S. (Editor); Eggleston, Debra S. (Editor); Gannaway, Mary T. (Editor)

    2002-01-01

    This volume, the first of two volumes, is a collection of 55 unclassified/unlimited-distribution papers which were presented at the Joint Army-Navy-NASA-Air Force (JANNAF) 38th Combustion Subcommittee (CS), 26 th Airbreathing Propulsion Subcommittee (APS), 20th Propulsion Systems Hazards Subcommittee (PSHS), and 21 Modeling and Simulation Subcommittee. The meeting was held 8-12 April 2002 at the Bayside Inn at The Sandestin Golf & Beach Resort and Eglin Air Force Base, Destin, Florida. Topics cover five major technology areas including: 1) Combustion - Propellant Combustion, Ingredient Kinetics, Metal Combustion, Decomposition Processes and Material Characterization, Rocket Motor Combustion, and Liquid & Hybrid Combustion; 2) Liquid Rocket Engines - Low Cost Hydrocarbon Liquid Rocket Engines, Liquid Propulsion Turbines, Liquid Propulsion Pumps, and Staged Combustion Injector Technology; 3) Modeling & Simulation - Development of Multi- Disciplinary RBCC Modeling, Gun Modeling, and Computational Modeling for Liquid Propellant Combustion; 4) Guns Gun Propelling Charge Design, and ETC Gun Propulsion; and 5) Airbreathing - Scramjet an Ramjet- S&T Program Overviews.

  10. Manual and Automatic Lineament Mapping: Comparing Results

    NASA Astrophysics Data System (ADS)

    Vaz, D. A.; di Achille, G.; Barata, M. T.; Alves, E. I.

    2008-03-01

    A method for automatic lineament extraction using topographic data is applied on the Thaumasia plateau. A comparison is made between the results that are obtained from the automatic mapping approach and from a traditional tectonic lineament mapping.

  11. Significance of selected lineaments in Alabama

    NASA Technical Reports Server (NTRS)

    Drahovzal, J. A.; Neathery, T. L.; Wielchowsky, C. C.

    1974-01-01

    Four lineaments in the Alabama Appalachians that appear on ERTS-1 imagery have been geologically analysed. Two of the lineaments appear to have regional geologic significance, showing relationships to structural and stratigraphic frameworks, water and mineral resources, geophysical anomalies, and seismicity. The other two lineaments are of local geologic significance, but, nevertheless, have important environmental implications.

  12. Lineaments of the Southern High Plains: possible indicators of recent faulting

    NASA Astrophysics Data System (ADS)

    Harrington, T.; Oluwole, O.; Gurrola, H.

    2011-12-01

    Through the analysis of topographic data, a number of parallel lineaments trending approximately N37°E across the Southern High Plains (SHP) of Texas have been identified. Although some of these lineaments appear to serve as drainage, they trend roughly perpendicular to the down slope direction of the SHP and they cross larger streams in the region. The source of formation of these lineaments is not known, however they appear across the entire Southern High Plains. It seems these features are easily recognized in the Southern High Plains where erosion and deposition is dominated by wind processes. Further north where erosion is more dominated by fluvial processes they are not easily recognized. Analysis of the orientation of brakes in river channels, find a strong orientation parallel to the observed lineaments. A single lineament can be tens of kilometers in length, and can continue well over a hundred kilometers as a set of en echelon lineaments. The length of these features suggests regional faulting in the subsurface to a significant depth. In the southern most part of the high plains there appears to be additional lineaments in the downstream direction perpendicular to the above mentioned lineaments. To determine the source of this linearity, we focus on a portion of two lineaments in Northern Hockley County, Texas that are approximately 2.5 km apart. The particular lineaments appear to have created an almost perpendicular bend in a significant ephemeral stream, causing the drainage and the incision of a deep arroyo to flow nearly perpendicular to the trend of the Southern High Plains. Subsequent gravity profiles show approximately a half a mGal gravity anomaly across each of these lineaments. The short wavelength of these gravity anomalies requires that the source (or at least part of it) be from a shallow structure; however, the small density contrasts common for the shallow sediments in the area cannot account for the entire 0.5 mGal difference in gravity

  13. Problems in the Study of lineaments

    NASA Astrophysics Data System (ADS)

    Anokhin, Vladimir; Kholmyanskii, Michael

    2015-04-01

    The study of linear objects in upper crust, called lineaments, led at one time to a major scientific results - discovery of the planetary regmatic network, the birth of some new tectonic concepts, establishment of new search for signs of mineral deposits. But now lineaments studied not enough for such a promising research direction. Lineament geomorphology has a number of problems. 1.Terminology problems. Lineament theme still has no generally accepted terminology base. Different scientists have different interpretations even for the definition of lineament We offer an expanded definition for it: lineaments - line features of the earth's crust, expressed by linear landforms, geological linear forms, linear anomalies of physical fields may follow each other, associated with faults. The term "lineament" is not identical to the term "fault", but always lineament - reasonable suspicion to fault, and this suspicion is justified in most cases. The structure lineament may include only the objects that are at least presumably can be attributed to the deep processes. Specialists in the lineament theme can overcome terminological problems if together create a common terminology database. 2. Methodological problems. Procedure manual selection lineaments mainly is depiction of straight line segments along the axes of linear morphostructures on some cartographic basis. Reduce the subjective factors of manual selection is possible, following a few simple rules: - The choice of optimal projection, scale and quality of cartographic basis; - Selection of the optimal type of linear objects under study; - The establishment of boundary conditions for the allocation lineament (minimum length, maximum bending, the minimum length to width ratio, etc.); - Allocation of an increasing number of lineaments - for representative sampling and reduce the influence of random errors; - Ranking lineaments: fine lines (rank 3) combined to form larger lineaments rank 2; which, when combined

  14. The spacing calculator software—A Visual Basic program to calculate spatial properties of lineaments

    NASA Astrophysics Data System (ADS)

    Ekneligoda, Thushan C.; Henkel, Herbert

    2006-05-01

    A software tool is presented which calculates the spatial properties azimuth, length, spacing, and frequency of lineaments that are defined by their starting and ending co-ordinates in a two-dimensional (2-D) planar co-ordinate system. A simple graphical interface with five display windows creates a user-friendly interactive environment. All lineaments are considered in the calculations, and no secondary sampling grid is needed for the elaboration of the spatial properties. Several rule-based decisions are made to determine the nearest lineament in the spacing calculation. As a default procedure, the programme defines a window that depends on the mode value of the length distribution of the lineaments in a study area. This makes the results more consistent, compared to the manual method of spacing calculation. Histograms are provided to illustrate and elaborate the distribution of the azimuth, length and spacing. The core of the tool is the spacing calculation between neighbouring parallel lineaments, which gives direct information about the variation of block sizes in a given category of structures. The 2-D lineament frequency is calculated for the actual area that is occupied by the lineaments.

  15. CLustre: semi-automated lineament clustering for palaeo-glacial reconstruction

    NASA Astrophysics Data System (ADS)

    Smith, Mike; Anders, Niels; Keesstra, Saskia

    2016-04-01

    Palaeo glacial reconstructions, or "inversions", using evidence from the palimpsest landscape are increasingly being undertaken with larger and larger databases. Predominant in landform evidence is the lineament (or drumlin) where the biggest datasets number in excess of 50,000 individual forms. One stage in the inversion process requires the identification of lineaments that are generically similar and then their subsequent interpretation in to a coherent chronology of events. Here we present CLustre, a semi-authomated algorithm that clusters lineaments using a locally adaptive, region growing, method. This is initially tested using 1,500 model runs on a synthetic dataset, before application to two case studies (where manual clustering has been undertaken by independent researchers): (1) Dubawnt Lake, Canada and (2) Victoria island, Canada. Results using the synthetic data show that classifications are robust in most scenarios, although specific cases of cross-cutting lineaments may lead to incorrect clusters. Application to the case studies showed a very good match to existing published work, with differences related to limited numbers of unclassified lineaments and parallel cross-cutting lineaments. The value in CLustre comes from the semi-automated, objective, application of a classification method that is repeatable. Once classified, summary statistics of lineament groups can be calculated and then used in the inversion.

  16. Commerce geophysical lineament - Its source, geometry, and relation to the Reelfoot rift and New Madrid seismic zone

    USGS Publications Warehouse

    Langenheim, V.E.; Hildenbrand, T.G.

    1997-01-01

    The Commerce geophysical lineament is a northeast-trending magnetic and gravity feature that extends from central Arkansas to southern Illinois over a distance of ???400 km. It is parallel to the trend of the Reelfoot graben, but offset ???40 km to the northwest of the western margin of the rift floor. Modeling indicates that the source of the aeromagnetic and gravity anomalies is probably a mafic dike swarm. The age of the source of the Commerce geophysical lineament is not known, but the linearity and trend of the anomalies suggest a relationship with the Reelfoot rift, which has undergone episodic igneous activity. The Commerce geophysical lineament coincides with several topographic lineaments, movement on associated faults at least as young as Quaternary, and intrusions of various ages. Several earthquakes (Mb > 3) coincide with the Commerce geophysical lineament, but the diversity of associated focal mechanisms and the variety of surface structural features along the length of the Commerce geophysical lineament obscure its relation to the release of present-day strain. With the available seismicity data, it is difficult to attribute individual earthquakes to a specific structural lineament such as the Commerce geophysical lineament. However, the close correspondence between Quaternary faulting and present-day seismicity along the Commerce geophysical lineament is intriguing and warrants further study.

  17. New global maps of Europa's lineaments

    NASA Astrophysics Data System (ADS)

    Cremonese, Gabriele; Lucchetti, Alice; Simioni, Emanuele

    2016-10-01

    Physical models have been developed to successfully explain the orientations and locations of many fractures observed on Europa's surface. Between the different fractures located on the surface of the icy satellite global-scale lineaments are present. These features are correlated with tidal stress suggesting that they initiated at tensile cracks in response to non-synchronous rotation (Geissler et al., 1998, Geissler et al. 1999). In this work we completed a global map of all type of lineaments presented on the surface of Europa, including also cycloidal lineaments that are interpreted to be tensile cracks that form due to diurnal stresses from Europa's orbital eccentricity (Hoppa et al., 1999).We enhanced the mapping of lineaments in comparison to what previously published, tracking about 5500 lineaments located everywhere on the surface of the icy satellite. We analyze these features in terms of their orientation and location using 2D methods, such as stereo plots and rose diagrams, showing that our preliminary results are in agreement with previous studies (McEwen et al. 1986). In addition, we visualize our results taking into account the 3D information to perform a detailed analysis of lineaments constraining their orientation and behavior. The aim of this work is to characterize the mapped lineaments and investigate the timing of their formation in order to correlate our results with proposed stress pattern models.

  18. Objective procedures for lineament enhancement and extraction

    USGS Publications Warehouse

    Moore, G.K.; Waltz, F.A.

    1983-01-01

    A longterm research goal at EROS Data Center is to develop automated, objective procedures for lineament mapping. In support of this goal, a five-step digital convolution procedure has been used to produce directionally enhanced images, which contain few artifacts and little noise. The main limitation of this procedure is that little enhancement of lineaments occurs in dissected terrain, in shadowed areas, and in flat areas with a uniform land cover. The directional enhancement procedure can be modified to extract edge and line segments from an image. Any of various decision rules can then be used to connect the line segments and to produce a final lineament map. The result is an interpretive map, but one that is based on an objective extraction of lineament components by digital processing. -from Authors

  19. Preliminary analysis of lineaments and engineering properties of bedrock, Penobscot Bay area, Maine

    USGS Publications Warehouse

    Lee, Fitzhugh T.; O'Leary, Dennis W.; Diehl, Sharon F.

    1977-01-01

    A remote sensing study of coastal and near coastal Maine was undertaken to identify bedrock features of possible importance to construction. Major lineaments were identified that separate the region into four distinct terrains. Within each terrain, smaller lineaments and other physiographic features show distinctive and consistent patterns, reflecting similarities in bedrock lithology and structure. The major linear features are given the following provisional geographic names: I. Lewiston line 2. Merrymeeting lineament 3. Dover-Foxcroft line 4. Orland lineament 5. Union lineament These linear or curvilinear trends are caused by lithologic contrasts, joints, faults, or foliation in the middle and lower Paleozoic phyllites, schists, gneisses, and granite intrusives. Initial field and laboratory results from a pilot study of bedrock engineering conditions indicate that variations in rock strength and fracture spacing are controlled by lithology, intensity of structural deformation, and alteration. At several locations in the study area faults and major joints are alined with a major lineament. Strength anisotropy ratios are as high as 10 in the foliated rocks but near 1 in the granites. The higher compressive strength direction commonly is perpendicular to foliation and to lineament trends which, in turn, are parallel to faults, major joints, or foliation. Fracture spacing in rocks in the field study area averages 0.5 m for granite and ranges from 0.2 cm to 0.75 m, with an average of 0.15 m, in metamorphic rocks. Diagnostic indications useful for estimating certain construction conditions are shown by point-load and uniaxial compressive strengths. Values for fresh rock range from less than 30 MN/m 2 for weakly bonded phyllite to more than 350 MN/m 2 for dense andesite and metasiltstone. Rock workability is directly related to compressive strength and spacing of joints.

  20. Workshop on remote sensing/lineament applications for energy extraction

    SciTech Connect

    Howard, J.F.; Komar, C.A.; Cooper, L.M.

    1984-04-01

    The following six papers are presented in this proceedings: (1) lineaments - a look forward and backward; (2) remote sensing/lineament analysis applications workshop - a general summary; (3) geologic controls on lineament systems, Pike County, Kentucky; (4) air-photo lineament domains in a portion of the Appalachian Basin - an application for exploration; (5) subsurface anomalies expressed as lineaments in the Northwest Pennsylvania portion of the Appalachian Plateau; and (6) effect of short lineaments on gas well yield from Devonian shales in Eastern Kentucky. All papers have been processed for inclusion in the Energy Data Base.

  1. Accuracy of lineaments mapping from space

    NASA Technical Reports Server (NTRS)

    Short, Nicholas M.

    1989-01-01

    The use of Landsat and other space imaging systems for lineaments detection is analyzed in terms of their effectiveness in recognizing and mapping fractures and faults, and the results of several studies providing a quantitative assessment of lineaments mapping accuracies are discussed. The cases under investigation include a Landsat image of the surface overlying a part of the Anadarko Basin of Oklahoma, the Landsat images and selected radar imagery of major lineaments systems distributed over much of Canadian Shield, and space imagery covering a part of the East African Rift in Kenya. It is demonstrated that space imagery can detect a significant portion of a region's fracture pattern, however, significant fractions of faults and fractures recorded on a field-produced geological map are missing from the imagery as it is evident in the Kenya case.

  2. A Global Data Base of 433Eros Lineaments

    NASA Astrophysics Data System (ADS)

    Buczkowski, D. L.; Prockter, L.; Barnouin-Jha, O. S.

    2005-12-01

    The Near-Earth Asteroid Rendezvous NEAR-Shoemaker spacecraft orbited the asteroid 433Eros for a year from 2000-2001. The NEAR Multi-Spectral Imager (MSI) collected tens of thousands of high resolution images and as a result Eros is the most comprehensively studied asteroid in the solar system. Previous mapping of lineaments on Eros has supported the suggestion of planes throughout the asteroid. We are creating a global data base of all Eros lineaments to better understand the global distribution of these features. It is particularly challenging to map lineament orientations on a non-spherical body (Eros is the shape of a peanut, measuring 34 km on the long axis). To address this issue we are mapping the lineaments directly on the Eros shapefile using POINTS, developed by J. Joseph at Cornell University. We compare lineament orientation to impact craters to determine if there is a causal relationship between cratering events and lineament formation. For example, preliminary mapping around Narcissus crater indicate at least four groups of lineaments: 1) previously mapped lineaments that may be indicative of planes through Eros, 2) lineaments radial to Narcissus crater, 3) lineaments related to nearby Shoemaker Regio, and 4) a set of lineaments that appear to be conjugate to set 1. Lineament sets 1 and 4 are also evident near Selene crater, as are lineaments radial to the crater. Lineament types include grooves, ridges and pit chains. We identify types of lineaments across the surface using a combination of NEAR Laser Rangefinder (NLR) topographic data and MSI images, and classify them according to region, including areas suggestive of thicker regolith. Further lineament/crater interactions are also examined, to determine the effect that lineaments have on crater shape.

  3. Lineaments of the Apennine Front - Apollo 15 landing site.

    NASA Technical Reports Server (NTRS)

    Wolfe, E. W.; Bailey, N. G.

    1972-01-01

    The well-developed systems of lineaments that suggest regional fracturing or layering of the bedrock in the mountains at the Apollo 15 site may be illusions created by oblique lighting rather than expressions of true geologic structure. Supporting evidence includes: (1) occurrence of similar dominant lineament trends on many of the slopes even though the slope orientations differ widely; (2) presence of a thick mantling regolith that should conceal bedrock structures; (3) approximate symmetry of some of the lineament sets about the sun line; (4) lack of clear-cut topographic expression for many of the lineaments; and (5) general similarity of the lunar lineaments to lineament patterns produced by low oblique lighting on model surfaces mantled by fine sifted powder. Locally, as at Silver Spur, some lineament trends are related to distinct topographic features and are more confidently interpreted to reflect underlying geologic structures.

  4. The 38th Annual Phi Delta Kappa/Gallup Poll of the Public's Attitudes toward the Public Schools

    ERIC Educational Resources Information Center

    Rose, Lowell C.; Gallup, Alec M.

    2006-01-01

    Rose and Gallup report on the results of the 38th Annual PDK/Gallup Poll of the Public's Attitude Toward the Public Schools. This year's survey examined No Child Left Behind and the public's perception of the law, the appropriate role of standardized testing, the achievement gap between white students and black and Hispanic students, the…

  5. Research in Medical Education: Proceedings of the Annual Conference (38th, Washington, DC, October 25-27, 1999).

    ERIC Educational Resources Information Center

    Anderson, M. Brownell, Ed.

    1999-01-01

    The Proceedings of the 38th Annual Conference on Research in Medical Education (Washington, DC, October 25-27, 1999) contain 43 research papers on innovative curricula, diagnostic reasoning, student evaluations of faculty, practicing physicians, prediction, licensing examinations, admissions, faculty development, managed care, technology-enhanced…

  6. 38th Annual Survey Report on State-Sponsored Student Financial Aid, 2006-2007 Academic Year

    ERIC Educational Resources Information Center

    National Association of State Student Grant and Aid Programs, 2007

    2007-01-01

    Each year, the National Association of State Student Grant and Aid Programs (NASSGAP) completes a survey regarding state-funded expenditures for postsecondary student financial aid. This report, the 38th annual survey, represents data from academic year 2006-07. Data highlights of this survey include: (1) In the 2006-2007 academic year, the states…

  7. Cocos plate gravity lineaments due to thermal contraction cracks

    NASA Astrophysics Data System (ADS)

    Balcerak, Ernie

    2011-09-01

    Gravity lineaments, narrow strips of stronger gravity, were first observed on the ocean floor in the south central Pacific several decades ago, but scientists still debate their origin. Because the south central Pacific gravity lineaments align with the tectonic plate motion, some scientists have suggested that they were created through small-scale convection cells below the lithosphere that become elongated as the plate moved. Other proposed explanations have included flow of low-viscosity material along the base of the the lithosphere and cracking of the lithosphere due to thermal contraction. Newly recognized gravity lineaments on the Cocos plate, which lies beneath the Pacific Ocean off the coast of Central America, could help scientists figure out how such lineaments form. Cormier et al. identified the lineaments using satellite altimetry data. They used gravity, bathymetric, and magnetic data to investigate the physical characteristics of the lineaments and how they changed over time. (Geochemistry, Geophysics, Geosystems, doi:10.1029/2011GC003573, 2011)

  8. Structural lineaments in the southern Sierra Nevada, California

    NASA Technical Reports Server (NTRS)

    Liggett, M. A. (Principal Investigator); Childs, J. F.

    1974-01-01

    The author has identified the following significant results. Several lineaments observed in ERTS-1 MSS imagery over the southern Sierra Nevada of California have been studied in the field in an attempt to explain their geologic origins and significance. The lineaments are expressed topographically as alignments of linear valleys, elongate ridges, breaks in slope or combinations of these. Natural outcrop exposures along them are characteristically poor. Two lineaments were found to align with foliated metamorphic roof pendants and screens within granitic country rocks. Along other lineaments, the most consistant correlations were found to be alignments of diabase dikes of Cretaceous age, and younger cataclastic shear zones and minor faults. The location of several Pliocene and Pleistocene volcanic centers at or near lineament intersections suggests that the lineaments may represent zones of crustal weakness which have provided conduits for rising magma.

  9. Florida lineament: Key tectonic element of eastern Gulf of Mexico

    SciTech Connect

    Christenson, G. )

    1990-09-01

    The origin of the Florida lineament, a major basement lineament that strikes northwest-southeast across the West Florida Shelf and southern Florida, is a key to the history of the Gulf of Mexico. Regional magnetic and gravity trends are truncated along the Florida lineament. New geologic data from recent wells on the West Florida Shelf and magnetic anomaly data indicate that pre-Mesozoic basement terranes on opposite sides of the Florida lineament were contiguous prior to Triassic-Jurassic volcanism and exhibit only minimal lateral offset across the Florida lineament at present. The lack of major lateral offset of pre-Mesozoic basement terranes across the Florida lineament and lithologic and geophysical data suggest that the lineament represents a Triassic-Jurassic extensional rift margin. The Florida lineament is interpreted to be the southeastward continuation of the well-documented peripheral fault system, which delineates the rifted continental margin of the northern Gulf basin. The continuation of the peripheral fault system along the Florida lineament suggests that the tectonostratigraphic terranes associated with the Mesozoic producing trends of the northern Gulf basin may extend southeastward along the Florida lineament. The interpretation of the Florida lineament as an extensional rift margin places significant constraints on any tectonic model of the Gulf of Mexico region. A tectonic interpretation consistent with the constraints suggests that the West Florida Shelf and southern Florida region formed as the result of Triassic-Jurassic extension around a pole of rotation in central Florida. The central Florida pole of rotation is intermediate to the poles of rotation counterclockwise of Yucatan out of the northern Gulf basin. This suggests that the region south of the Florida lineament underwent extension synchronous with the rotation of the Yucatan block.

  10. A new fault lineament in Southern California

    NASA Technical Reports Server (NTRS)

    Pease, R. W.; Johnson, C. W.

    1973-01-01

    ERTS-1 imagery clearly shows a 50-mile wide tectonic zone across Southern California oriented about 15 deg to the structures of the Transverse Ranges or with an azimuth of 70 deg. The zone is delineated on the imagery by terrian alignments and vegetational differences. A previously undisclosed tectonic lineament extends across the Mojave Desert and appears as a line of crustal upwarping. Pressure which would have caused this plus the occurrence of many thrust faults with the 70 deg azimuth indicate this to be a zone of crustal compression. Recent earthquake epicenters appear to be related to this compression zone rather than the traditional fault network of Southern California.

  11. From the 38th Parallel to the DMZ: Teaching about the Two Koreas--A Nation Divided (1945-Present).

    ERIC Educational Resources Information Center

    Peters, Richard

    How the Korean peninsula came to be divided in 1945 is discussed, and the stalemated situation that has existed since that time is described. This brief history is followed by lesson plans for grades 6-12. The lesson plans include goals and objectives, activities, and assessment techniques for teaching the following concepts: (1) territorial…

  12. Structural lineament and pattern analysis of Missouri, using LANDSAT imagery

    NASA Technical Reports Server (NTRS)

    Martin, J. A.; Kisvarsanyi, G. (Principal Investigator)

    1977-01-01

    The author has identified the following significant results. Major linear, circular, and arcuate traces were observed on LANDSAT imagery of Missouri. Lineaments plotted within the state boundaries range from 20 to nearly 500 km in length. Several extend into adjoining states. Lineaments plots indicate a distinct pattern and in general reflect structural features of the Precambrian basement of the platform. Coincidence of lineaments traced from the imagery and known structural features in Missouri is high, thus supporting a causative relation between them. The lineament pattern apparently reveals a fundamental style of the deformation of the intracontinental craton. Dozens of heretofore unknown linear features related to epirogenic movements and deformation of this segment of the continental crust were delineated. Lineaments and mineralization are interrelated in a geometrically classifiable pattern.

  13. Shape of lenticulae on Europa and their interaction with lineaments.

    NASA Astrophysics Data System (ADS)

    Culha, Cansu; Manga, Michael

    2015-04-01

    The surface of Europa contains many elliptical features that have been grouped into three classes: (a) positive relief (domes), (b) negative relief (pits), or (c) complex terrain (small chaos). Collectively, these three classes of features are often called "lenticulae". The internal processes that form lenticulae are unknown. However, given that the diameters of all these features are similar, it is parsimonious to ascribe each class of feature to a different stage in the evolution of some process occurring within the ice shell. Proposed models for these features including diapirs (Sotin et al., 2002; Rathbun et al., 1998); melting above diapirs (Schmidt et al., 2011); and sills of water (Michaut and Manga, 2014). The objective of the present study is to first characterize the shape of lecticulae, and then look at the interaction of lenticulae with lineaments, in order to test lenticulae formation mechanisms. Lenticulae and lineaments are mapped and annotated on ArcGIS. We mapped a total of 57 pits and 86 domes. Both pits and domes have similar aspect ratios and orientations. The elliptical similarities of domes and pits suggest that domes and pits are surface expressions of different stages of a common process within the ice shell. The cross cutting relationships between lineaments reveal relative age. Lineaments either lie over or under the lenticulae. All of the lineament segments that appear within pits also appear topographically lower than the rest of the surface. Domes lie over and under lineaments, but unlike pits there are lineaments that lie over domes that do not vary in topography. This suggests that the lineaments that lie above lenticulae and match the lenticulae's topography are older than the lenticulae. Domes have more crossing lineaments. Therefore, on average, they appear to be older than pits. Lineaments also appear on the sides of lenticulae. There are two different ways in which adjacent lineaments appear: 1. they disrupt the shape of the

  14. Mapping of Landsat satellite and gravity lineaments in west Tennessee

    NASA Technical Reports Server (NTRS)

    Argialas, Demetre P.; Stearns, Richard G.; Shahrokhi, Firouz

    1988-01-01

    The analysis of earthquake fault lineament patterns within the alluvial valley of west Tennessee, which is often made difficult by the presence of unconsolidated sediments, is presently undertaken through a synergistic use of Landsat satellite images in conjunction with gravity anomaly data, which were quantitatively analyzed and compared by means of two-dimensional histograms and rose diagrams. The northeastern trend revealed for the lineaments corresponds to faults and is in keeping with reactivation of the Reelfoot rift near the Mississippi River; this suggests that deeper features, perhaps at earthquake focal depth, may extend to the land surface as Landsat-detectable lineaments.

  15. Investigation of lineaments on Skylab and ERTS images of Peninsular Ranges, Southwestern California

    NASA Technical Reports Server (NTRS)

    Merifield, P. M. (Principal Investigator); Lamar, D. L.

    1974-01-01

    The author has identified the following significant results. Northwest trending faults such as the Elsinore and San Jacinto are prominently displayed on Skylab and ERTS images of the Peninsular Ranges, southern California. Northeast, north-south, and west-north-west trending lineaments and faults are also apparent on satellite imagery. Several of the lineaments represent previously unmapped faults. Other lineaments are due to erosion along foliation directions and sharp bends in basement rock contacts rather than faulting. The northeast trending Thing Valley fault appears to be offset by the south branch of the Elsinore fault near Agua Caliente Hot Springs. Larger horizontal displacement along the Elsinore fault further northwest may be distributed along several faults which branch from the Elsinore fault in the Peninsular Ranges. The northeast and west-northwest trending faults are truncated by the major northwest trending faults and appear to be restricted to basement terrane. Limited data on displacement direction suggests that the northeast and west-northwest trending faults formed in response to an earlier period of east-northeast, west-southwest crustal shortening. Such a stress system is consistent with the plate tectonic model of a subduction zone parallel to the continental margin suggested in the late Mesozoic and early Tertiary.

  16. Multiple expressions of plume-ridge interaction at the Galapagos: Volcanic lineaments and ridge jumps

    NASA Astrophysics Data System (ADS)

    Mittelstaedt, E. L.; Soule, S. A.; Harpp, K. S.; Fornari, D. J.

    2011-12-01

    Despite significant evidence for communication between an upwelling mantle plume beneath the Galápagos Archipelago (GA) and the nearby Galápagos Spreading Center (GSC), little is known about the dynamics of the interaction between the ridge and the hotspot. We use new bathymetry, sidescan sonar, magnetic, subbottom seismic, and gravity data from the FLAMINGO cruise (MV1007) to address the mechanism of plume-ridge interaction in the Northern Galápagos Volcanic Province (NGVP), a region centered on the 90° 50'W Galápagos transform fault (GTF). West of the GTF, the Nazca Plate is dominated by numerous seamounts aligned in 3 volcanic lineaments. Volumetrically, the lineaments are composed of ~1500 km3 of erupted lavas. Faulting patterns and seamount elongations suggest that the locations and orientations of the lineaments may be partly controlled by the lithospheric stress field. In contrast, east of the GTF on the Cocos Plate, there is little evidence of constructional volcanism (~69 km3). However, we observe several linear, nearly ridge-parallel, faulted features separating sediment-filled basins, and two large bathymetric highs with up to 1km of relief. Differences in seafloor west and east of the GTF are also observed in the Residual Mantle Bouguer Anomaly (RMBA). On the Nazca Plate, RMBA lows closely contour the volcanic lineaments with minima at the centers of the largest volcanoes along the Wolf-Darwin Lineament. On the Cocos Plate, the RMBA at a given distance from the ridge axis is generally more negative than at similar locations on the Nazca Plate. In addition, two RMBA lows are observed coincident with the observed bathymetric highs, both of which are slightly elongate in a direction sub-parallel to the Eastern GSC, possibly reflecting a period of increased magma flux along a former ridge axis. Magnetic anomalies reveal a complicated history of plate evolution including a series of ridge jumps since ~3 Ma that result in creation of the GTF. We invert

  17. Lineaments that are artifacts of lighting, part G

    NASA Technical Reports Server (NTRS)

    Howard, K. A.; Larsen, B. R.

    1972-01-01

    Apollo 15 orbital photographs, particularly those taken at low sun elevation angles, are examined revealing grid patterns of lineaments. Preliminary results of experiments demonstrate that spurious lineaments and grid patterns can be produced and that the directions are dependent in part upon the position of the light source. The experiments were designed to duplicate the effect of bright sunlight reflecting from a hummocky surface with little or no diffuse light in the shadowed areas.

  18. Some Characteristics of Regular Fracture-lineament Global Network

    NASA Astrophysics Data System (ADS)

    Anokhin, Vladimir; Longinos, Biju

    2013-04-01

    Existence of regular fracture-lineament global network global network (FLGN) (or regmatic network), was known for lands of the Earth in many regions. Authors made more than 20 000 measurements of lineaments and faults azimuths of the lineaments and fractures on geographic, geologic and tectonic maps for number of regions and for all Earth. Later all data files have subjected by the factor analysis. We detect existence FLGN in the Ocean bottom. Statistic relation between fractures and lineaments directions was established. Control of large-scale lineaments by fractures within the competence of the FLGN was based. Predominating strike directions of line elements of FLGN are: 0 - 10˚, 80 - 90˚, 30 - 60˚, 120 - 150˚. FLGN have attribute of fractality. One-direction lines elements of the FLGN alternate with constant step within the competence of defined scale. FLGN was formed under a continuous stress, which exist at least throughout the entire earthcrust thickness and during the time of at least the entire Phanerozoe. This stress was generated by a complex of forces: rotational, pulsating and, possibly, some others in the earthcrust. All of these forces are symmetric to the Earth rotation axis and some of them also to the equator. Rotation and pulsating processes of the Earth are the main factors of these forces and, hence, formation of the fracture- lineament network. FLGN determines the most favorable place for fracturing, formation of fracture-controlled landforms, volcanic and seismic processes (geohazards), fluid flow and ore-formation (minerals).

  19. Lineaments in basement terrane of the Peninsular Ranges, Southern California

    NASA Technical Reports Server (NTRS)

    Merifield, P. M. (Principal Investigator); Lamar, D. L.

    1974-01-01

    The author has identified the following significant results. ERTS and Skylab images reveal a number of prominent lineaments in the basement terrane of the Peninsular Ranges, Southern California. The major, well-known, active, northwest trending, right-slip faults are well displayed; northeast and west to west-northwest trending lineaments are also present. Study of large-scale airphotos followed by field investigations have shown that several of these lineaments represent previously unmapped faults. Pitches of striations on shear surfaces of the northeast and west trending faults indicate oblique slip movement; data are insufficient to determine the net-slip. These faults are restricted to the pre-tertiary basement terrane and are truncated by the major northwest trending faults. They may have been formed in response to an earlier stress system. All lineaments observed in the space photography are not due to faulting, and additional detailed geologic investigations are required to determine the nature of the unstudied lineaments, and the history and net-slip of fault-controlled lineaments.

  20. Recognition of lineaments in Eastern Rhodopes on Landsat multispectral images

    NASA Astrophysics Data System (ADS)

    Borisova, Denitsa; Jelev, Georgi; Atanassov, Valentin; Koprinkova-Hristova, Petia; Alexiev, Kiril

    Lineaments usually appear on the multispectral images as lines (edges) or linear formations as a result of the color variations of the surface structures. Lineaments are line features on earth’s surface which reflect geological structure. The basic geometry of a line is orientation, length and curve. Detection of lineaments is an important operation in the exploration for mineral deposits, in the investigation of active fault patterns, water resources, etc. In this study the integrated approach is applied. It comes together the methods of the visual interpretation of various geological and geographical indications in the satellite images, application of spatial analysis in GIS and automatic processing of Landsat multispectral image by Canny algorithm, Directional Filter and Neural Network. Canny algorithm for extracting edges is series of filters (Gaussian, Sobel, etc.) applied to all bands of the image using the free IDL source (http://www.cis.rit.edu/class/simg782/programs/ canny.pro). Directional Filter is applied to sharpen the image in a specific (preferred) direction. Another method is the Neural Network algorithm for recognizing lineaments. Lineaments are effectively extracted using different methods of automatic. The results from above mentioned methods are compared to results derived from visual interpretation of satellite images and from geological map. The rose-diagrams of distribution of lineaments and maps of their density are completed. Acknowledgments: This study is supported by the project DFNI - I01/8 funded by the Bulgarian Science Fund.

  1. THEMIS Observations of Domes and Associated Lineaments in Arcadia Planitia

    NASA Astrophysics Data System (ADS)

    Milam, K. A.; McSween, H. Y.

    2003-12-01

    The northern plains of Mars contain several high concentrations (Acidalia, Utopia, Elysium, etc.) of small (<10 km diameter) domes, proposed to be volcanic. Recent data sets from the Mars Global Surveyor and Mars Odyssey spacecraft provide new insight into the formation of domes in Arcadia Planitia. Daytime and nighttime thermal infrared (TIR) data from the Thermal Emission Imaging System (THEMIS), visible images from THEMIS and the Mars Orbiter Camera, and elevation data from the Mars Orbiter Laser Altimeter were used to study a 390,000 km 2 area ˜1500 km to the northwest of Elysium Mons. Of interest is a region centered on Tyndall crater and bordered by Phlegras Montes to the west. The area is characterized by gentle, westward-sloping plains, with noticeable slope breaks along several N-S trending wrinkle ridges. Several hundred circular domes dot this area. Domes display features consistent with a volcanic origin. Most are circular to slightly elliptical at their base, with basal diameters ranging from 0.5-6 km. Summits typically rise <300 m above the surrounding plains. Domes have shallow slopes (lacking significant slope breaks) that range from 1-9° . Visible images and TIR-derived temperature data suggest that slopes are composed of finer-grained material (as compared to the coarser-grained summits). Less than 25% of domes appear to have summit depressions and ~ 1% show fractured summit areas. Some domes appear to be randomly distributed, but many are aligned in chains according to wrinkle ridge orientations. Using THEMIS data, we have detected over 165 domes that are aligned with and superimposed upon over 145 lineaments. Most lineaments are <500 m in width and range from 1-66 km in length. Many lineaments do follow N-S trends similar to those of wrinkle ridges, although other orientations are common. Several lineaments can be seen as open fractures, while others appear to be filled with fine-grained sediment. While most domes are superimposed upon

  2. The Cottage Lake Aeromagnetic Lineament: A Possible Onshore Extension of the Southern Whidbey Island Fault, Washington

    USGS Publications Warehouse

    Blakely, Richard J.; Sherrod, Brian L.; Wells, Ray E.; Weaver, Craig S.; McCormack, David H.; Troost, Kathy G.; Haugerud, Ralph A.

    2004-01-01

    topographic and aeromagnetic lineaments cross the tunnel alignment. Some of the disturbance is likely tectonic in origin, although other explanations are possible. Some of the soil disturbance demonstrably predates the 15-13 ka Fraser glaciation of the Puget Lowland; other samples have inconclusive ages and may be younger. Subtle scarps in Pleistocene surfaces are visible on high-resolution lidar topography at a number of locations along the Cottage Lake aeromagnetic lineament. Collectively, the scarps are parallel to the trend of the aeromagnetic lineament and extend a total distance of 18 km. In the field, scarps exhibit 1 to 5 m of north-side-up offset. The scarps provide targets for future paleoseismic trenching studies to test the hypothesis that they have a tectonic origin.

  3. Multiple expressions of plume-ridge interaction in the Galápagos: Volcanic lineaments and ridge jumps

    NASA Astrophysics Data System (ADS)

    Mittelstaedt, E.; Soule, S.; Harpp, K.; Fornari, D.; McKee, C.; Tivey, M.; Geist, D.; Kurz, M. D.; Sinton, C.; Mello, C.

    2012-05-01

    Anomalous volcanism and tectonics between near-ridge mantle plumes and mid-ocean ridges provide important insights into the mechanics of plume-lithosphere interaction. We present new observations and analysis of multibeam, side scan sonar, sub-bottom chirp, and total magnetic field data collected during the R/V Melville FLAMINGO cruise (MV1007; May-June, 2010) to the Northern Galápagos Volcanic Province (NGVP), the region between the Galápagos Archipelago and the Galápagos Spreading Center (GSC) on the Nazca Plate, and to the region east of the Galápagos Transform Fault (GTF) on the Cocos Plate. The NGVP exhibits pervasive off-axis volcanism related to the nearby Galápagos hot spot, which has dominated the tectonic evolution of the region. Observations indicate that ˜94% of the excess volcanism in our survey area occurs on the Nazca Plate in three volcanic lineaments. Identified faults in the NGVP are consistent with normal ridge spreading except for those within a ˜60 km wide swath of transform-oblique faults centered on the GTF. These transform-oblique faults are sub-parallel to the elongation direction of larger lineament volcanoes, suggesting that lineament formation is influenced by the lithospheric stress field. We evaluate current models for lineament formation using existing and new observations as well as numerical models of mantle upwelling and melting. The data support a model where the lithospheric stress field controls the location of volcanism along the lineaments while several processes likely supply melt to these eruptions. Synthetic magnetic models and an inversion for crustal magnetization are used to determine the tectonic history of the study area. Results are consistent with creation of the GTF by two southward ridge jumps, part of a series of jumps that have maintained a plume-ridge separation distance of 145 km to 215 km since ˜5 Ma.

  4. Lineaments in coastal plain sediments as seen in ERTS imagery

    NASA Technical Reports Server (NTRS)

    Withington, C. F.

    1973-01-01

    Examination of satellite imagery over the Atlantic Coastal Plain near Washington, D. C. shows numerous lineaments, which cannot be accounted for by any known cultural or natural features. At least some of these lineaments represent the surface expression of faults, for one of them has been correlated with the outcrop of a fault that had been traced for several miles in southern Prince Georges County, Maryland. If a substantial number of these lineaments do indeed represent fault traces, the fact that they show on the surface suggests that the geologic history of the Coastal Plain is much more complex than has previously been recognized, and that faulting may have occurred in the Holocene, much later than has generally been recognized. The importance that such recent movements could have on future development of the Coastal Plain should be emphasized.

  5. Algorithms for lineaments detection in processing of multispectral images

    NASA Astrophysics Data System (ADS)

    Borisova, D.; Jelev, G.; Atanassov, V.; Koprinkova-Hristova, Petia; Alexiev, K.

    2014-10-01

    Satellite remote sensing is a universal tool to investigate the different areas of Earth and environmental sciences. The advancement of the implementation capabilities of the optoelectronic devices which are long-term-tested in the laboratory and the field and are mounted on-board of the remote sensing platforms further improves the capability of instruments to acquire information about the Earth and its resources in global, regional and local scales. With the start of new high-spatial and spectral resolution satellite and aircraft imagery new applications for large-scale mapping and monitoring becomes possible. The integration with Geographic Information Systems (GIS) allows a synergistic processing of the multi-source spatial and spectral data. Here we present the results of a joint project DFNI I01/8 funded by the Bulgarian Science Fund focused on the algorithms of the preprocessing and the processing spectral data by using the methods of the corrections and of the visual and automatic interpretation. The objects of this study are lineaments. The lineaments are basically the line features on the earth's surface which are a sign of the geological structures. The geological lineaments usually appear on the multispectral images like lines or edges or linear shapes which is the result of the color variations of the surface structures. The basic geometry of a line is orientation, length and curve. The detection of the geological lineaments is an important operation in the exploration for mineral deposits, in the investigation of active fault patterns, in the prospecting of water resources, in the protecting people, etc. In this study the integrated approach for the detecting of the lineaments is applied. It combines together the methods of the visual interpretation of various geological and geographical indications in the multispectral satellite images, the application of the spatial analysis in GIS and the automatic processing of the multispectral images by Canny

  6. Mapping paleostructures from time-related aeromagnetic lineaments and photolineaments

    SciTech Connect

    Peterson, R.M.

    1985-01-01

    A classification of aeromagnetic lineaments has been adapted for use with photolineaments. William Dean Gibbons recently defined sets of structurally-high, aeromagnetic lineaments for each Phanerozoic period; e.g., N19/sup 0/W and N71/sup 0/E for Cambrian time. This model was tested on several Cambrian (Knox) oil fields in Kentucky by mapping dense sets of N34/sup 0/W, N19/sup 0/W, N86/sup 0/E, and N71/sup 0/E lineaments from air photos and interpreting drag folds from lineaments. A comparison of Knox structural contour maps with Cambrian drag fold maps indicates that drag fold maps approximate structural contour maps; clusters of N19/sup 0/W and N71/sup 0/E drag folds correspond to closed highs. Several porosity and permeability trends were identified for Cambrian and Ordovician reservoirs. This procedure for mapping paleohighs has been further tested in the Mid-Continent on structures that range in age from Cambrian to Cretaceous and is being used (with Landsat) to identify locations of paleostructures that range in age from Cambrian to Cretaceous and is being used (with Landsat) to identify locations of paleostructures for seismic surveys. With air photos the method is used for detailed studies on fracture, porosity/permeability trends and delineation of features identified first on Landsat. A current project in the Denver-Julesburg basin, using Landsat RBV images, in the search for new Pennsylvanian and Permian paleohighs.

  7. On the reliability of manually produced bedrock lineament maps

    NASA Astrophysics Data System (ADS)

    Scheiber, Thomas; Viola, Giulio; Fredin, Ola; Jarna, Alexandra; Gasser, Deta; Łapinska-Viola, Renata

    2016-04-01

    Manual extraction of topographic features from digital elevation models (DEMs) is a commonly used technique to produce lineament maps of fractured basement areas. There are, however, several sources of bias which can influence the results. In this study we investigated the influence of the factors (a) scale, (b) illumination azimuth and (c) operator on remote sensing results by using a LiDAR (Light Detection and Ranging) DEM of a fractured bedrock terrain located in SW Norway. Six operators with different backgrounds in Earth sciences and remote sensing techniques mapped the same LiDAR DEM at three different scales and illuminated from three different directions. This resulted in a total of 54 lineament maps which were compared on the basis of number, length and orientation of the drawn lineaments. The maps show considerable output variability depending on the three investigated factors. In detail: (1) at larger scales, the number of lineaments drawn increases, the line lengths generally decrease, and the orientation variability increases; (2) Linear features oriented perpendicular to the source of illumination are preferentially enhanced; (3) The reproducibility among the different operators is generally poor. Each operator has a personal mapping style and his/her own perception of what is a lineament. Consequently, we question the reliability of manually produced bedrock lineament maps drawn by one person only and suggest the following approach: In every lineament mapping study it is important to define clear mapping goals and design the project accordingly. Care should be taken to find the appropriate mapping scale and to establish the ideal illumination azimuths so that important trends are not underrepresented. In a remote sensing project with several persons included, an agreement should be reached on a given common view on the data, which can be achieved by the mapping of a small test area. The operators should be aware of the human perception bias. Finally

  8. Lineaments on Skylab photographs: Detection, mapping, and hydrologic significance in central Tennessee

    NASA Technical Reports Server (NTRS)

    Moore, G. K. (Principal Investigator)

    1976-01-01

    The author has identified the following significant results. Lineaments were detected on Skylab photographs by stereo viewing, projection viewing, and composite viewing. Sixty-nine percent more lineaments were found by stereo viewing than by projection, but segments of projection lineaments are longer; total length of lineaments found by these two methods is nearly the same. Most Skylab lineaments consist of topographic depression: stream channel alinements, straight valley walls, elongated swales, and belts where sinkholes are abundant. Most of the remainder are vegetation alinements. Lineaments are most common in dissected areas having a thin soil cover. Results of test drilling show: (1) the median yield of test wells on Skylab lineaments is about six times the median yield of all existing wells; (2) three out of seven wells on Skylab lineaments yield more than 6.3 1/s (110 gal/min): (3) low yields are possible on lineaments as well as in other favorable locations; and (4) the largest well yields can be obtained at well locations of Skylab lineaments that also are favorably located with respect to topography and geologic structure, and are in the vicinity of wells with large yields.

  9. Landsat and field studies link structural lineaments and mineralization in St. Francois, Missouri

    SciTech Connect

    Robertson, C.E.

    1984-07-01

    Late-stage Precambrian granites in the St. Francois Mountains are among the most uraniferous in North America. The St. Francois province has potential for uranium mineralization of economic importance, especially in the later differentiates. Structural lineaments and circular features displayed on images produced by electronic data processing of Landsat multispectral scanner data may be related to late-stage intrusives with uranium potential. Strong north-south lineaments and associated circular and arcuate features may correspond to major weaknesses in the earth's crust along which fracturing, faulting, and volcanism have occurred. The strike of the lineaments transects the older dominant northwest-southeast and northeast-southwest structural grain of the region. This, and the remarkable preservation of Precambrian structures of volcanic origin, indicate that the lineaments may be related to late-stage, uranium- and thorium-rich intrusives. The Ironton lineament, a major north-south lineament, is closely related spatially to Precambrian iron and manganese deposits. Field work along the Ironton lineament suggests that it is related to a late period of Precambrian volcanism and that structural deformation along the lineament continued into early Paleozoic time. Areas of faulting, shearing, and hydrothermal alteration affecting both Precambrian and Paleozoic rocks have been located. A circular feature along the lineament has been found to be centered by a manganese deposit of possible hydrothermal origin.

  10. Major crustal lineaments and their influence on the geological history of the continental lithosphere

    SciTech Connect

    Reading, H.G.; Watterson, J.; White, S.H.

    1987-01-01

    Thirty-eight contributors provide a synthesis and evaluation of our current understanding of the seismic signatures of major lineaments, their structure and reactivation, their subsequent control of geological processes, and their influence on the location of mineral and hydrocarbon deposits. Topics covered include architecture of the continental lithosphere, earthquakes and lineament infrastructure, crustal shear zones and thrust belts: their geometry and continuity in central Africa, fluid transport in lineaments, and the evolution of the Malvern Lineament. Originally published in Philosphical Transactions of the Royal Society.

  11. Lineament analysis of mineral areas of interest in Afghanistan

    USGS Publications Warehouse

    Hubbard, Bernard E.; Mack, Thomas J.; Thompson, Allyson L.

    2012-01-01

    The purpose of this report and accompanying GIS data is to provide lineament maps that give one indication of areas that warrant further investigation for optimal bedrock water-well placement within 24 target areas for mineral resources (Peters and others, 2011). These data may also support the identification of faults related to modern seismic hazards (for example, Wheeler and others, 2005; Ruleman and others, 2007), as well as support studies attempting to understand the relationship between tectonic and structural controls on hydrothermal fluid flow, subsequent mineralization, and water-quality issues near mined and unmined mineral deposits (for example, Eppinger and others, 2007).

  12. Morphostructure analysis of Sapaya ancient volcanic area based lineament data

    NASA Astrophysics Data System (ADS)

    Massinai, Muhammad Altin; Kadir, Fitrah H.; Ismullah, Muh. Fawzy; Aswad, Sabrianto

    2016-05-01

    Morphostructure of Sapaya ancient volcanic have been analysis by using lineament models. In this models, two methods of retrieval data have been used. First, the field survey of the area, second, the satellite images analysis. The morphostructure of Sapaya ancient volcanic contribute to the crater, caldera, and shown an eroded cone morphology. The directions of eruption from Sapaya ancient volcanic have identified in region of Jeneponto and Takalar, which is had east - west and northeast - southwest structure. These eruptions also give contribution to the character of river in Jenelata watershed, by the presence of tuffs, pillow lava, basalt, andesite, diorite, granodiorite, granite, and gabbro, respectively.

  13. Crack azimuths on Europa: The G1 lineament sequence revisited

    USGS Publications Warehouse

    Sarid, A.R.; Greenberg, R.; Hoppa, G.V.; Brown, D.M.; Geissler, P.

    2005-01-01

    The tectonic sequence in the anti-jovian area covered by regional mapping images from Galileo's orbit E15 is determined from a study of cross-cutting relationships among lineament features. The sequence is used to test earlier results from orbit G1, based on lower resolution images, which appeared to display a progressive change in azimuthal orientation over about 90?? in a clockwise sense. Such a progression is consistent with expected stress variations that would accompany plausible non-synchronous rotation. The more recent data provide a more complete record than the G1 data did. We find that to fit the sequence into a continual clockwise change of orientation would require at least 1000?? (> 5 cycles) of azimuthal rotation. If due to non-synchronous rotation of Europa, this result implies that we are seeing back further into the tectonic record than the G1 results had suggested. The three sets of orientations found by Geissler et al. now appear to have been spaced out over several cycles, not during a fraction of one cycle. While our more complete sequence of lineament formation is consistent with non-synchronous rotation, a statistical test shows that it cannot be construed as independent evidence. Other lines of evidence do support non-synchronous rotation, but azimuths of crack sequences do not show it, probably because only a couple of cracks form in a given region in any given non-synchronous rotation period. ?? 2004 Elsevier Inc. All rights reserved.

  14. Lineaments of Texas - possible surface expressions of deep-seated phenomena. Final report

    SciTech Connect

    Woodruff, C.M. Jr.; Caran, S.C.

    1984-04-01

    Lineaments were identified on 51 Landsat images covering Texas and parts of adjacent states in Mexico and the United States. A method of identifying lineaments was designed so that the findings would be consistent, uncomplicated, objective, and reproducible. Lineaments denoted on the Landsat images were traced onto 1:250,000-scale work maps and then rendered cartographically on maps representing each of the 51 Landsat images at a scale of 1:500,000. At this stage more than 31,000 lineaments were identified. It included significant areas outside of Texas. In preparing the final lineament map of Texas at 1:1,000,000-scale from the 1:500,000-scale maps, all features that lay outside Texas and repetition among features perceived by individual workers were eliminated. Cultural features were checked for before reducing and cartographically fitting the mosaic of 51 individual map sheets to a single map base. Lineaments that were partly colinear but with different end points were modified into a single lineament trace with the combined length of the two or more colinear lineaments. Each lineament was checked to determine its validity according to our definition. The features were edited again to eliminate processing artifacts within the image itself, as well as representations of cultural features (fencelines, roads, and the like) and geomorphic patterns unrelated to bedrock structure. Thus the more than 31,000 lineaments originally perceived were reduced to the approximately 15,000 presented on the 1:1,000,000 map. Interpretations of the lineaments are presented.

  15. GIS based 3D visualization of subsurface and surface lineaments / faults and their geological significance, northern tamil nadu, India

    NASA Astrophysics Data System (ADS)

    Saravanavel, J.; Ramasamy, S. M.

    2014-11-01

    The study area falls in the southern part of the Indian Peninsular comprising hard crystalline rocks of Archaeozoic and Proterozoic Era. In the present study, the GIS based 3D visualizations of gravity, magnetic, resistivity and topographic datasets were made and therefrom the basement lineaments, shallow subsurface lineaments and surface lineaments/faults were interpreted. These lineaments were classified as category-1 i.e. exclusively surface lineaments, category-2 i.e. surface lineaments having connectivity with shallow subsurface lineaments and category-3 i.e. surface lineaments having connectivity with shallow subsurface lineaments and basement lineaments. These three classified lineaments were analyzed in conjunction with known mineral occurrences and historical seismicity of the study area in GIS environment. The study revealed that the category-3 NNE-SSW to NE-SW lineaments have greater control over the mineral occurrences and the N-S, NNE-SSW and NE-SW, faults/lineaments control the seismicities in the study area.

  16. Bootheel lineament: A possible coseismic fault of the great New Madrid earthquakes

    SciTech Connect

    Schweig, E.S. III; Marple, R.T. )

    1991-10-01

    A remote sensing examination of the New Madrid seismic zone has revealed a feature, the Bootheel lineament, that may be the surface expression of one of the coseismic faults of the great New Madrid earthquakes of 1811 and 1812. The lineament extends about 135 km in a north-northeast direction through northeastern Arkansas and southeastern Missouri. The morphology and pattern of the lineament suggest that it reflects a fault with strike-slip displacement. Field data indicate that liquefied sand was injected along the lineament, probably in 1811 and 1812. The Bootheel lineament does not coincide with any of the major arms of New Madrid seismicity, possibly indicating that the current seismicity does not precisely reflect the faults that ruptured in 1811 and 1812.

  17. A tomographic glimpse of the upper mantle source of magmas of the Jemez lineament, New Mexico

    USGS Publications Warehouse

    Spence, W.; Gross, R.S.

    1990-01-01

    To infer spatial distributions of partial melt in the upper mantle source zones for the Rio Grande rift and the Jemez lineament, the lateral variations of P wave velocity in the upper mantle beneath these features has been investigated. Teleseismic P wave delays recorded at a 22-station network were used to perform a damped least squares, three-dimensional inversion for these lateral variations. Results infer that a large magmatic source zone exists beneath the Jemez lineament but not beneath the Rio Grande rift. This implies that the volcanic potential of the Jemez lineaments continues to greatly exceed that of the Rio Grande rift. The magmatic source zones of the Jemez lineament are modeled as due to clockwise rotation of the Colorado Plateau about a pole in northeastern Colorado. This rotation caused extension of the lithosphere beneath the Jemez lineament, permitting concentration there of partially melted rock in the upper mantle. -from Authors

  18. Revealing topographic lineaments through IHS enhancement of DEM data. [Digital Elevation Model

    NASA Technical Reports Server (NTRS)

    Murdock, Gary

    1990-01-01

    Intensity-hue-saturation (IHS) processing of slope (dip), aspect (dip direction), and elevation to reveal subtle topographic lineaments which may not be obvious in the unprocessed data are used to enhance digital elevation model (DEM) data from northwestern Nevada. This IHS method of lineament identification was applied to a mosiac of 12 square degrees using a Cray Y-MP8/864. Square arrays from 3 x 3 to 31 x 31 points were tested as well as several different slope enhancements. When relatively few points are used to fit the plane, lineaments of various lengths are observed and a mechanism for lineament classification is described. An area encompassing the gold deposits of the Carlin trend and including the Rain in the southeast to Midas in the northwest is investigated in greater detail. The orientation and density of lineaments may be determined on the gently sloping pediment surface as well as in the more steeply sloping ranges.

  19. Geological and geophysical signatures of the Jemez lineament: a reactivated Precambrian structure

    SciTech Connect

    Aldrich, M.J. Jr.; Ander, M.E.; Laughlin, A.W.

    1981-01-01

    The Jemez lineament (N52/sup 0/E) is one of several northeast-trending lineaments that traverse the southwestern United States. It is defined by a 500-km-long alignment of late Cenozoic volcanic fields extending southwest from at least the Jemez Mountains in the north-central New Mexico to the San Carlos-Peridot volcanic field in east-central Arizona. Geochronologic data from Precambrian basement rocks indicate that the lineament is approximately coincident with a boundary between Precambrian crustal provinces. Characteristics of the lineament are high heat flow (>104.5 mW/m/sup 2/), an attenuated seismic velocity zone from 25 to 140 km depth, and an upwarp of the crustal electrical conductor inferred from magnetotelluric studies. The high electrical conductivity is probably caused by the presence of interstitial magma in the rocks of the mid-to-upper crust. The average electical strike within the Precambrian basement is N60/sup 0/E, supporting a relationship between the Precambrian structural grain and the Jemez lineament. The geological and geophysical data suggest that the lineament is a structural zone that extends deep into the lithosphere and that its location was controlled by an ancient zone of weakness in the Precambrian basement. Ages of late Cenozoic volcanic rocks along the lineament show no systematic geographic progression, thus indicating that a mantle plume was not responsible for the alignment of the volcanic fields.Most of the faults, dikes, and cinder cone alignments along the lineament trend approximately N25/sup 0/E and N5/sup 0/W. These trends may represent Riedel shears formed by left-lateral transcurrent movement along the structure. Less common trends of cinder cone alignments and dikes are approximately N65/sup 0/W and N85/sup 0/W. The diversity in orientation indicates that the magnitudes of the two horizontal principal stresses within the lineament have been approximately equal for at least the last 5 m.y.

  20. Structural lineaments in the basement rocks of the central Eastern Desert, Egypt

    NASA Astrophysics Data System (ADS)

    Kamel, A. F.

    The Egyptian basement rocks outcrop in Eastern Desert, southern Sinai and southwestern Desert. The rocks belong to Precambrian and consist of igneous and metamorphic rocks which are characterized by crystalline character. Not much work has been done on the tectonics and structure of the basement rocks in Eastern Desert. The present work is a photogeological interpretation of the structural lineaments representing dykes, faults and joints in central Eastern Desert to differentiate between igneous and metamorphic rocks. The photogeological interpretation was carried out using normal aerial photographs scale 1:40 000 and photomosaics scale 50 000. The main trends of lineaments in the studied area are: E-W, ENE-WSW and WNW-ESE, constituting 58.4% of the total length and 54.5% of the total number. Correlating the structural lineaments in igneous rocks of Gebel El Bakriya locality with those in the metamorphic rocks of Gebel Abu Mireiwa shows that there is a marked difference between the two types. Lineaments in igneous rocks are elongated and widely spaced while those in metamorphic rocks are short and closely spaced. The different trends of joints in igneous rocks can be arranged as follows: WNW>E-W>NW>NNW>ENE>NE>N-S>NNE while the different trends of lineaments in metamorphic rocks can be arranged as follows: E-W>ENE>WNW>NW>NNW>NE>N-S>NNE. Comparison between the structural contour maps constructed for the total length of all lineaments and those representing joints in igneous and metamorphic rocks indicates that igneous rocks have lower density of lineaments than metamorphic rocks. The total length of all lineaments in Gebel El Bakriya amounts to 375 km, while lineaments representing joints have a total length of 150 km. In the metamorphic rocks of Gebel Abu Mireiwa, the total length of all lineaments is 425 km and those representing joints have a total length of 175 km. It was found that there is a relationship between the structural lineaments and radioactivity of

  1. Analysis of the Tectonic Lineaments in the Ganiki Planitia (V14) Quadrangle, Venus

    NASA Technical Reports Server (NTRS)

    Venechuk, E. M.; Hurwitz, D. M.; Drury, D. E.; Long, S. M.; Grosfils, E. B.

    2005-01-01

    The Ganiki Planitia quadrangle, located between the Atla Regio highland to the south and the Atalanta Planitia lowland to the north, is deformed by many tectonic lineaments which have been mapped previously but have not yet been assessed in detail. As a result, neither the characteristics of these lineaments nor their relationship to material unit stratigraphy is well constrained. In this study we analyze the orientation of extensional and compressional lineaments in all non-tessera areas in order to begin characterizing the dominant tectonic stresses that have affected the region.

  2. Cleats and their relation to geologic lineaments and coalbed methane potential in Pennsylvanian coals in Indiana

    USGS Publications Warehouse

    Solano-Acosta, W.; Mastalerz, Maria; Schimmelmann, A.

    2007-01-01

    Cleats and fractures in Pennsylvanian coals in southwestern Indiana were described, statistically analyzed, and subsequently interpreted in terms of their origin, relation to geologic lineaments, and significance for coal permeability and coalbed gas generation and storage. These cleats can be interpreted as the result of superimposed endogenic and exogenic processes. Endogenic processes are associated with coalification (i.e., matrix dehydration and shrinkage), while exogenic processes are mainly associated with larger-scale phenomena, such as tectonic stress. At least two distinct generations of cleats were identified on the basis of field reconnaissance and microscopic study: a first generation of cleats that developed early on during coalification and a second generation that cuts through the previous one at an angle that mimics the orientation of the present-day stress field. The observed parallelism between early-formed cleats and mapped lineaments suggests a well-established tectonic control during early cleat formation. Authigenic minerals filling early cleats represent the vestiges of once open hydrologic regimes. The second generation of cleats is characterized by less prominent features (i.e., smaller apertures) with a much less pronounced occurrence of authigenic mineralization. Our findings suggest a multistage development of cleats that resulted from tectonic stress regimes that changed orientation during coalification and basin evolution. The coals studied are characterized by a macrocleat distribution similar to that of well-developed coalbed methane basins (e.g., Black Warrior Basin, Alabama). Scatter plots and regression analyses of meso- and microcleats reveal a power-law distribution between spacing and cleat aperture. The same distribution was observed for fractures at microscopic scale. Our observations suggest that microcleats enhance permeability by providing additional paths for migration of gas out of the coal matrix, in addition to

  3. Lineaments on Skylab photographs: detection, mapping and hydrologic significance in central Tennessee

    USGS Publications Warehouse

    Moore, Gerald K.

    1976-01-01

    Lineaments were detected on SKYLAB photographs by stereo viewing , projection viewing, and composite viewing. Large well yields of 25 gal/min or more can be obtained in the study area by locating future wells on SKYLAB lineaments rather than on lineaments detected on either high-altitude aerial photographs, LANDSAT images, or by random drilling. Larger savings might be achieved by locating wells on lineaments detected by both stereo viewing and projection. The test site is underlain by dense, fractured, flat-lying limestones. Soil cover averages 4 ft thick in the Central Basin and about 40 ft thick on the Eastern Highland rim. Groundwater occurs mostly in horizontal, sheetlike solution cavities, and the trends of these cavities are controlled by joints. (Lardner-ISWS)

  4. Correlation of LANDSAT lineaments with Devonian gas fields in Lawrence County, Ohio

    NASA Technical Reports Server (NTRS)

    Johnson, G. O.

    1981-01-01

    In an effort to locate sources of natural gas in Ohio, the fractures and lineaments in Black Devonian shale were measured by: (1) field mapping of joints, swarms, and fractures; (2) stereophotointerpretation of geomorphic lineaments with precise photoquads; and (3) by interpreting the linear features on LANDSAT images. All results were compiled and graphically represented on 1:250,000 scale maps. The geologic setting of Lawrence County was defined and a field fracture map was generated and plotted as rose patterns at the exposure site. All maps were compared, contrasted, and correlated by superimposing each over the other as a transparency. The LANDSAT lineaments had significant correlation with the limits of oil and gas producing fields. These limits included termination of field production as well as extensions to other fields. The lineaments represent real rock fractures with zones of increased permeability in the near surface bedrock.

  5. Comparison of Skylab and Landsat Lineaments with Joint Orientations in Northcentral Pennsylvania. [on Allegheny Plateau

    NASA Technical Reports Server (NTRS)

    Mcmurtry, G. J.; Petersen, G. W. (Principal Investigator); Kowalik, W. S.

    1975-01-01

    The author has identified the following significant results. The histogram peaks of lineaments mapped from Skylab photograph at a scale of 1:517,000 lie subparallel, within 20 deg, to major shale joints and coal cleats on part of the Allegheny Plateau. The Landsat lineament, mapped at 1:989,000 are biased by illumination and scan line directions. While there is an illumination bias in the Skylab photograph, its direction does not coincide with the main transverse lineament trend, thus providing an independent assessment of the illumination direction bias. The coincidence in direction regardless of scale of the linear features suggests a mechanical relationship between joints, fracture traces, and lineaments which is more consistent with a tensional model than a shear model of origin.

  6. PREFACE: 1st International Workshop on Theoretical and Computational Physics: Condensed Matter, Soft Matter and Materials Physics & 38th National Conference on Theoretical Physics

    NASA Astrophysics Data System (ADS)

    2014-09-01

    This volume contains selected papers presented at the 38th National Conference on Theoretical Physics (NCTP-38) and the 1st International Workshop on Theoretical and Computational Physics: Condensed Matter, Soft Matter and Materials Physics (IWTCP-1). Both the conference and the workshop were held from 29 July to 1 August 2013 in Pullman hotel, Da Nang, Vietnam. The IWTCP-1 was a new activity of the Vietnamese Theoretical Physics Society (VTPS) organized in association with the 38th National Conference on Theoretical Physics (NCTP-38), the most well-known annual scientific forum dedicated to the dissemination of the latest development in the field of theoretical physics within the country. The IWTCP-1 was also an External Activity of the Asia Pacific Center for Theoretical Physics (APCTP). The overriding goal of the IWTCP is to provide an international forum for scientists and engineers from academia to share ideas, problems and solution relating to the recent advances in theoretical physics as well as in computational physics. The main IWTCP motivation is to foster scientific exchanges between the Vietnamese theoretical and computational physics community and world-wide scientists as well as to promote high-standard level of research and education activities for young physicists in the country. About 110 participants coming from 10 countries participated in the conference and the workshop. 4 invited talks, 18 oral contributions and 46 posters were presented at the conference. In the workshop we had one keynote lecture and 9 invited talks presented by international experts in the fields of theoretical and computational physics, together with 14 oral and 33 poster contributions. The proceedings were edited by Nguyen Tri Lan, Trinh Xuan Hoang, and Nguyen Ai Viet. We would like to thank all invited speakers, participants and sponsors for making the conference and the workshop successful. Nguyen Ai Viet Chair of NCTP-38 and IWTCP-1

  7. Geologic evaluation of major Landsat lineaments in Nevada and their relationship to ore districts

    USGS Publications Warehouse

    Rowan, Lawrence C.; Wetlaufer, Pamela Heald

    1979-01-01

    Analysis of diverse geologic, geophysical, and geochemical data shows that eight major lineament systems delineated in Landsat images of Nevada are morphological and tonal expressions of substantially broader structural zones. Southern Nevada is dominated by the 175 km-wide northwest-trending Walker Lane, a 150 km-wide zone of east-trending lineament systems consisting of the Pancake Range, Warm Springs, and Timpahute lineament systems, and a 125 km-wide belt of northeast-trending faults termed the Pahranagat lineament system. Northern Nevada is dominated by the northeast-trending 75-200km wide Midas Trench lineament system, which is marked by northeasterly-oriented faults, broad gravity anomalies, and the Battle Mountain heat flow high; this feature appears to extend into central Montana. The Midas Trench system is transected by the Northern Nevada Rift, a relatively narrow zone of north-northwest-trending basaltic dikes that give rise to a series of prominent aeromagnetic highs. The northwest-trending Rye Patch lineament system, situated at the northeast boundary of the Walker Lane, also intersects the Midas Trench system and is characterized by stratigraphic discontinuities and alignment of aeromagnetic anomalies. Field relationships indicate that all the lineament systems except for the Northern Nevada Rift are conjugate shears formed since mid-Miocene time during extension of the Great Basin. Metallization associated with volcanism was widespread along these systems during the 17-6 m.y. period. However, these zones appear to have been established prior to this period, probably as early as Precambr-an time. These lineament systems are interpreted to be old, fundamental, structural zones that have been reactivated episodically as stress conditions !changed in the western United States. Many metal districts are localized within these zones as magma rose along the pre-existing conduits.

  8. Lineaments and fracture traces, Jennings County and Jefferson Proving Ground, Indiana

    USGS Publications Warehouse

    Greeman, T.K.

    1981-01-01

    Well placement is important in this area, as fractures are a principal source of water to wells. In areas of fractured bedrock, the most productive well locations are at the intersections of two or more mapped lineaments or fracture traces and at the lowest local altitude. Use of the lineament and fracture-trace map will not guarantee a sufficent supply of ground water but will minimize the chance of drilling an inadequate well.

  9. Lineaments derived from analysis of linear features mapped from Landsat images of the Four Corners region of the Southwestern United States

    USGS Publications Warehouse

    Knepper, Daniel H.

    1982-01-01

    Linear features are relatively short, distinct, non-cultural linear elements mappable on Landsat multispectral scanner images (MSS). Most linear features are related to local topographic features, such as cliffs, slope breaks, narrow ridges, and stream valley segments that are interpreted as reflecting directed aspects of local geologic structure including faults, zones of fracturing (joints), and the strike of tilted beds. 6,050 linear features were mapped on computer-enhanced Landsat MSS images of 11 Landsat scenes covering an area from the Rio Grande rift zone on the east to the Grand Canyon on the west and from the San Juan Mountains, Colorado, on the north to the Mogollon Rim on the south. Computer-aided statistical analysis of the linear feature data revealed 5 statistically important trend intervals: 1.) N. 10W.-N.16E., 2.) N.35-72E., 3.) N.33-59W., 4.) N. 74-83W., and 5.) N.89-9-W. and N. 89-90E. Subsequent analysis of the distribution of the linear features indicated that only the first three trend intervals are of regional geologic significance. Computer-generated maps of the linear features in each important trend interval were prepared, as well as contour maps showing the relative concentrations of linear features in each trend interval. These maps were then analyzed for patterns suggestive of possible regional tectonic lines. 20 possible tectonic lines, or lineaments, were interpreted from the maps. One lineament is defined by an obvious change in overall linear feature concentrations along a northwest-trending line cutting across northeastern Arizona. Linear features are abundant northeast of the line and relatively scarce to the southwest. The remaining 19 lineaments represent the axes of clusters of parallel linear features elongated in the direction of the linear feature trends. Most of these lineaments mark previously known structural zones controlled by linear features in the Precambrian basement or show newly recognized relationships to

  10. The Bootheel lineament, the 1811-1812 New Madrid earthquake sequence, and modern seismicity

    SciTech Connect

    Schweig, E.S. Center for Earthquake Research and Information, Memphis, TN ); Ellis, M.A. )

    1992-01-01

    Pedologic, geomorphic, and geochronologic data suggest that liquefaction occurred along the Bootheel lineament of Missouri and Arkansas during the 1811-1812 New Madrid earthquake sequence. The authors propose that the lineament may be the surface trace of a relatively young fault zone consisting of multiple strike-slip flower structures. These structures have been interpreted over a zone at least 5 km wide exhibiting deformed strata at least as young as a regional Eocene/Quaternary unconformity. In physical models, flower structures form in less rigid material in response to low finite displacement across a discrete strike-slip shear zone in a rigid basement. By analogy, the Bootheel lineament may represent the most recent attempt of a strike-slip fault zone of relatively low displacement to propagate through a weak cover. In addition, the Bootheel lineament extends between two well-established, seismically active strike-slip fault zones that current form a restraining step. Restraining steps along strike-slip fault zones are inherently unstable, and thus the Bootheel lineament may be acting to smooth the trace of the New Madrid seismic zone as displacement increases. The current seismic inactivity along the Bootheel lineament may be explained by sequential accommodation of complex strain in which the stress field is highly variable within the source volume. In other words, the current stress field may not represent that which operated during the 1811-1812 sequence. Alternatively, an earthquake on a fault associated with the bootheel lineament may have released sufficient strain energy to temporarily shut down activity.

  11. Semi-automatic extraction of lineaments from remote sensing data and the derivation of groundwater flow-paths

    NASA Astrophysics Data System (ADS)

    Mallast, U.; Gloaguen, R.; Geyer, S.; Rödiger, T.; Siebert, C.

    2011-01-01

    We describe a semi-automatic method to objectively and reproducibly extract lineaments based on the global one arc-second ASTER GDEM. The combined method of linear filtering and object-based classification ensures a high degree of accuracy resulting in a lineament map. Subsequently lineaments are differentiated into geological and morphological lineaments to assign a probable origin and hence a hydro-geological significance. In the western catchment area of the Dead Sea (Israel) the orientation and location of the differentiated lineaments are compared to characteristics of known structural features. The authors demonstrate that a strong correlation between lineaments and structural features exist, being either influenced by the Syrian Arc paleostress field or the Dead Sea stress field or by both. Subsequently, we analyse the distances between lineaments and wells thereby creating an assessment criterion concerning the hydraulic significance of detected lineaments. Derived from this analysis the authors suggest that the statistic analysis of lineaments allows a delineation of flow-paths and thus significant information for groundwater analysis. We validate the flow-path delineation by comparison with existing groundwater model results based on well data.

  12. Topographic and Air-Photo Lineaments in Various Locations Related to Geothermal Exploration in Colorado

    DOE Data Explorer

    Zehner, Richard

    2012-02-01

    Title: Topographic and Air-Photo Lineaments in Various Locations Related to Geothermal Exploration in Colorado Tags: Colorado, lineaments, air-photo, geothermal Summary: These line shapefiles trace apparent topographic and air-photo lineaments in various counties in Colorado. It was made in order to identify possible fault and fracture systems that might be conduits for geothermal fluids, as part of a DOE reconnaissance geothermal exploration program. Description: Geothermal fluids commonly utilize fault and fractures in competent rocks as conduits for fluid flow. Geothermal exploration involves finding areas of high near-surface temperature gradients, along with a suitable “plumbing system” that can provide the necessary permeability. Geothermal power plants can sometimes be built where temperature and flow rates are high. This line shapefile is an attempt to use desktop GIS to delineate possible faults and fracture orientations and locations in highly prospective areas prior to an initial site visit. Geochemical sampling and geologic mapping could then be centered around these possible faults and fractures. To do this, georeferenced topographic maps and aerial photographs were utilized in an existing GIS, using ESRI ArcMap 10.0 software. The USA_Topo_Maps and World_Imagery map layers were chosen from the GIS Server at server.arcgisonline.com, using a UTM Zone 13 NAD27 projection. This line shapefile was then constructed over that which appeared to be through-going structural lineaments in both the aerial photographs and topographic layers, taking care to avoid manmade features such as roads, fence lines, and utility right-of-ways. Still, it is unknown what actual features these lineaments, if they exist, represent. Although the shapefiles are arranged by county, not all areas within any county have been examined for lineaments. Work was focused on either satellite thermal infrared anomalies, known hot springs or wells, or other evidence of geothermal systems

  13. Lineament mapping of vertical fractures of rock outcrops by remote sensing images

    NASA Astrophysics Data System (ADS)

    Matarrese, Raffaella; Masciopinto, Costantino

    2016-04-01

    The monitoring of hydrological processes within the vadose zone is usually difficult, especially in the presence of compact rock subsoil. The possibility of recognizing the trend of the structural lineaments in fractured systems has important fallout in the understanding water infiltration processes, especially when the groundwater flow is strongly affected by the presence of faults and fractures that constitute the preferred ways of water fluxes. This study aims to detect fracture lineaments on fractured rock formations from CASI hyperspectral airborne VNIR images, with a size of 60 cm of the spatial resolution, and collected during November 2014. Lineaments detected with such high resolution have been compared with the fracture lineaments detected by a Landsat 8 image acquired at the same time of the CASI acquisition. The method has processed several remote sensed images at different spatial resolution, and it has produced the visualization of numerous lineament maps, as result of the vertical and sub-vertical fractures of the investigated area. The study has been applied to the fractured limestone outcrop of the Murgia region (Southern Italy). Here the rock formation hosts a deep groundwater, which supplies freshwater for drinking and irrigation purposes. The number of the fractures allowed a rough estimation of the vertical average hydraulic conductivity of the rock outcrop. This value was compared with field saturated rock hydraulic conductivity measurements derived from large ring infiltrometer tests carried out on the same rock outcrop.

  14. Application of satellite photographic and MSS data to selected geologic and natural resource problems in Pennsylvania. 1: Lineaments and mineral occurrences in Pennsylvania. 2: Relation of lineaments to sulfide deposits: Bald Eagle Mountain, Centre County, Pennsylvania. 3: Comparison of Skylab and LANDSAT lineaments with joint orientations in north central Pennsylvania

    NASA Technical Reports Server (NTRS)

    Kowalik, W. S.; Gold, D. P.; Krohn, M. D.

    1975-01-01

    Those metallic mineral occurrences in Pennsylvania are reported which lie near lineaments mapped from LANDSAT-1 satellite imagery and verified from Skylab photography where available. The lineaments were categorized by degree of expression and type of expression; the mineral occurrences were classified by host rock age, mineralization type, and value. The accompanying tables and figure document the mineral occurrences geographically associated with lineaments and serve as a base for a mineral exploration model.

  15. Using Remote Sensing for Lineament Extraction in al Maghrabah Area - Hajjah, Yemen

    NASA Astrophysics Data System (ADS)

    Al-Nahmi, F.; Alami, O. B.; Baidder, L.; Khanbari, K.; Rhinane, H.; Hilali, A.

    2016-10-01

    Remote sensing and GIS data have become more and more important for the study of geology, structural geology and extract of lineament which give us an overview of the tectonic events. The main objective of our study is to design a method for extracting and mapping of lineament from Landsat 8_OLI satellite images of the study area. Different processing techniques were used to achieve our objective : MNF Minimum Noise Fraction bands (4,5,6), Color composites (7,5,3), PCA Principal Component Analysis bands (4,5,7), band ratios (7/5,6/4,4/2), directional filters, and image processing applied on the Landsat 8_OLI composite colours. The results of this study indicate that, the area is really affected by several structural trends: N-S, NW-SE, and NE-SW directions. The result of lineament analysis gives us a good interpretation of the main structural geology and tectonic forces.

  16. Quantification of Geologic Lineaments by Manual and Machine Processing Techniques. [Landsat satellites - mapping/geological faults

    NASA Technical Reports Server (NTRS)

    Podwysocki, M. H.; Moik, J. G.; Shoup, W. C.

    1975-01-01

    The effect of operator variability and subjectivity in lineament mapping and methods to minimize or eliminate these problems by use of several machine preprocessing methods was studied. Mapped lineaments of a test landmass were used and the results were compared statistically. The total number of fractures mapped by the operators and their average lengths varied considerably, although comparison of lineament directions revealed some consensus. A summary map (785 linears) produced by overlaying the maps generated by the four operators shows that only 0.4 percent were recognized by all four operators, 4.7 percent by three, 17.8 percent by two, and 77 percent by one operator. Similar results were obtained in comparing these results with another independent group. This large amount of variability suggests a need for the standardization of mapping techniques, which might be accomplished by a machine aided procedure. Two methods of machine aided mapping were tested, both simulating directional filters.

  17. A new artefacts resistant method for automatic lineament extraction using Multi-Hillshade Hierarchic Clustering (MHHC)

    NASA Astrophysics Data System (ADS)

    Šilhavý, Jakub; Minár, Jozef; Mentlík, Pavel; Sládek, Ján

    2016-07-01

    This paper presents a new method of automatic lineament extraction which includes the removal of the 'artefacts effect' which is associated with the process of raster based analysis. The core of the proposed Multi-Hillshade Hierarchic Clustering (MHHC) method incorporates a set of variously illuminated and rotated hillshades in combination with hierarchic clustering of derived 'protolineaments'. The algorithm also includes classification into positive and negative lineaments. MHHC was tested in two different territories in Bohemian Forest and Central Western Carpathians. The original vector-based algorithm was developed for comparison of the individual lineaments proximity. Its use confirms the compatibility of manual and automatic extraction and their similar relationships to structural data in the study areas.

  18. Wavy Lineaments on Europa: Fracture Propagation into Combined Nonsynchronous and Diurnal Stress Fields

    NASA Technical Reports Server (NTRS)

    Crawford, Zane; Pappalardo, Robert T.; Barr, Amy C.; Gleeson, Damhnait; Mullen, McCall; Nimmo, Francis; Stempel, Michelle M.; Wahr, John

    2005-01-01

    Understanding the processes that have operated on Europa and the manner in which they may have changed through time is fundamental to understanding the satellite's geology and present-day habitability. Previous studies have shown that lineament patterns on Europa can be explained by accumulation of tensile stress from slow nonsynchronous rotation (NSR), while the cycloidal planforms of other Europan lineaments can be explained if fractures propagate through a diurnally changing tensile stress field. We find that fractures propagated into combined diurnal and NSR stress fields can be "wavy" in planform for NSR stress accumulated over 2 to 8 of ice shell rotation and average propagation speeds of approx. 1 to 3 m/s. The variety of Europa's observed lineament planforms from cycloidal, to wavy, to arcuate can be produced by accumulation of NSR stress relative to the diurnal stress field. Varying proportions of these stress mechanisms plausibly may be related to a time-variable (slowing) NSR rate.

  19. Paleostress field analysis of the southeastern part of the Aller Lineament

    NASA Astrophysics Data System (ADS)

    Brandes, Christian; Schmidt, Carolin

    2013-04-01

    The Aller Lineament is more than 200 km long and one of the most important fault zones in the Central European Basin System. It was active during the Mesozoic and there is evidence for Late Quaternary tectonic movements (Veldkamp et al., 2002). The evolution of the central and southeastern part of this fault zone was analysed in detail with seismic surveys (Best, 1996; Lohr et al., 2007), but a comprehensive outcrop based paleostress field analysis stills lacks. We analysed faults and conjugate shear fractures in the upper Aller Valley to derive the deformation history and the paleostress field evolution at the southeastern end of the fault zone. Especially conjugate shear fractures serve as important fabric for the paleostress analysis. Hancock & Kadhi (1978) showed that in such a system σ1 bisects the smallest dihedral angle and σ2 is parallel to the intersection of the conjugate fractures. The great advantage of conjugate shear fractures is that the orientation of the three major stress directions can be accurately constrained. Different paleostress directions were identified in the data from the upper Aller Valley. Sets of NW-SE trending normal faults are developed in Upper Triassic sandstones, which indicate an NE-SW directed extensional phase. This was most likely the initial phase of faulting along the Aller Lineament. It took place in the Late Triassic and matches the results derived from seismic sections presented by Best (1996). Sets of NNE-SSW trending, conjugate fractures with a vertical acute angle in the fracture sets are developed. They indicate a vertical maximum principle normal paleostress, which could have resulted from the formation of relay zones between individual NW-SE trending fault planes. From a second set of conjugate shear fractures, a maximum horizontal compressive paleostress vector was reconstructed that is NNE-SSW to N-S oriented. This fits with the previous results from the Central European Basin System and reflects the NNE

  20. A study of the Tyrone-Mount Union lineament by remote sensing techniques and field methods

    NASA Technical Reports Server (NTRS)

    Gold, D. P. (Principal Investigator)

    1977-01-01

    The author has identified the following significant results. This study has shown that subtle variations in fold axes, fold form, and stratigraphic thickness can be delineated. Many of the conclusions were based on extrapolation in similitude to different scales. A conceptual model was derived for the Tyrone-Mount Union lineament. In this model, the lineament the morphological expression of a zone of fracture concentrations which penetrated basement rocks and may have acted as a curtain to regional stresses or as a domain boundary between uncoupled adjacent crustal blocks.

  1. Application of SEASAT-1 Synthetic Aperture Radar (SAR) data to enhance and detect geological lineaments and to assist LANDSAT landcover classification mapping. [Appalachian Region, West Virginia

    NASA Technical Reports Server (NTRS)

    Sekhon, R.

    1981-01-01

    Digital SEASAT-1 synthetic aperture radar (SAR) data were used to enhance linear features to extract geologically significant lineaments in the Appalachian region. Comparison of Lineaments thus mapped with an existing lineament map based on LANDSAT MSS images shows that appropriately processed SEASAT-1 SAR data can significantly improve the detection of lineaments. Merge MSS and SAR data sets were more useful fo lineament detection and landcover classification than LANDSAT or SEASAT data alone. About 20 percent of the lineaments plotted from the SEASAT SAR image did not appear on the LANDSAT image. About 6 percent of minor lineaments or parts of lineaments present in the LANDSAT map were missing from the SEASAT map. Improvement in the landcover classification (acreage and spatial estimation accuracy) was attained by using MSS-SAR merged data. The aerial estimation of residential/built-up and forest categories was improved. Accuracy in estimating the agricultural and water categories was slightly reduced.

  2. Attempt at correlating Italian long lineaments from LANDSAT-1 satellite images with some geological phenomena. Possible use in geothermal energy research

    NASA Technical Reports Server (NTRS)

    Barbier, E.; Fanelli, M.

    1975-01-01

    By utilizing the images from the LANDSAT-1, in the spectral band 0.8-1.1 microns (near infrared), a photomosaic was obtained of Italian territory. From this mosaic the field of long lineaments was drawn, corresponding to fractures of the earth crust more than 100 km long. The relationship between lineaments, hot springs, volcanic areas, and earthquake epicenters is verified. There is a clear connection between long lineaments and hot springs: 78% of the springs are located on one or more lineaments, and the existence of hot lineaments was observed. A slightly weaker, but still significant, connection exists between the Pliocene-Quaternary volcanic areas and long lineaments. The relationship between earthquakes and long lineaments can only be verified in some cases. The lineaments which can be related to earthquakes have little or no connection with the other phenomena.

  3. Regional Lineament Net of Uzbekistan: RemoteSensing&GIS, mineral deposits and geodynamic structures

    NASA Astrophysics Data System (ADS)

    Sidorova, I. P.

    2005-12-01

    This paper deals with the application of complex study of Remote Sensing images and deep Lithospheric structures to the knowledge of spatial interrelations between regional lineaments and mineral deposits in Uzbekistan. Deciphering of structural units of Uzbekistan territory using space ASTER images allows us to reveal regional, deep-rooted lineament, extending in latitudinal direction over Uzbekistan territory and neighboring countries. Thus lineament could penetrate the Earth up to deep Lithosphere layers, inheriting a position of old fault-lineament systems which origin related to Paleocene tectonic processes. The most extended latitudinal lineament is the Transregional lineament of Central Asia located within 42-44N zone. It stretches for more than 2000km from Sultan-Uvais mountains (Karakalpakstan), through Kyzylkums and Nurata mountains (Uzbekistan), Turkestan-Alay and Atbashi- Inylchek mountains (Kyrgyzstan), to Chinese border with possible extension along the Chinese Tien- Shan. The main objective is to associate the surface indicators as geological, geophysical and tectonic base of data using RS & GIS with the purpose to identify the occurrence special geoobjects of economic interest. Additionally, it will be possible to evaluate geospatial distributions of these altered zones related to morphological structures using Digital Elevation Modelling (DEM) products of ASTER images. RS&GIS methods were used to determine the interrelations of the volcanic and granitic rocks distribution-mineralization-alteration with the faults-lineaments, circular structures. The alteration zones, the tectonic lines and the circular structures related to the cones and calderas determined these methods and checked by group truth studies may be target areas to explore for some new ore deposits. As a result, our investigations envelops more then 10 mineral deposits of such kind bearing Au, Ag, Pt, U, Zn, Cu, Ni, rare earth elements and diamonds. In conclusion, it is necessary to

  4. Hypothesis on the origin of lineaments in the LANDSAT and SLAR images of precambrian soil in the low Contas River Valley (southern Bahia)

    NASA Technical Reports Server (NTRS)

    Liu, C. C. (Principal Investigator); Rodrigues, J. E.

    1984-01-01

    Examination of LANDSAT and SLAR images in southern Bahia reveals numerous linear features, which are grouped in five sets, based on their trends: N65 degrees E, N70 degrees W, N45 degrees E and NS/N15 degrees E. Owing to their topographic expressions, distributive patterns, spacing between individual lineaments and their mutual relationships, the lineament sets of N65 degrees E and N70 degrees W, as well as the sets of N40 degrees E and N45 degrees W, are considered as two groups of conjugate shear fractures and the former is older and is always cut by the latter. Their conjugate shear angles are 45 degrees and 85 degrees and their bisector lines are approximately in east-west and north-south directions, respectively. According to Badgeley's argumentation on the conjugate shear angles, the former conjugate shear fractures would be caused by: (1) vertical movements, and the bisector of their conjugate angle would be parallel to the long axis of horsting or folding, or (2) by a compressive force in the east-west direction and under a condition of low confining pressure and temperature.

  5. Proceedings of the 5th International LISA Symposium and the 38th ESLAB Symposium, ESTEC, Noordwijk, The Netherlands, 12-15 July 2004

    NASA Astrophysics Data System (ADS)

    Jennrich, Oliver

    2005-05-01

    In its 10th year, the International LISA Symposium of 2004 was the first to be organized by ESA. After the first Symposium at the Rutherford Appleton Laboratory in 1996, and subsequent Symposia at Caltech (1998), the Albert Einstein Institute in Golm (2000) and Penn State University (2002), ESA's European Space and Technology Centre (ESTEC) is proud to have hosted the 5th International LISA Symposium in 2004 in conjunction with the 38th ESLAB Symposium. During these 10 years, we have seen the technology required for such ambitious space missions as LISA and LISA Pathfinder developing and maturing to a point where flightlevel hardware can be produced and tested. The technology demonstrator LISA Pathfinder has become a mission well established in the ESA programme and has entered its project implementation phase; the LISA mission is in the formulation phase. At the same time, the preparations for LISA data analysis and first activities on data archives and LISA simulators have begun, indicating that LISA is regarded as more than just a mission far in the future. The astrophysics of the sources of gravitational waves is emerging as a rapidly growing field that will become even more important in the coming years, when the focus of the activities on LISA and LISA Pathfinder will shift from making the missions possible to deriving scientific results from the data. The 5th International LISA Symposium had presentations on all of the above topics, from the technology of LISA and LISA Pathfinder, LISA data analysis, and modelling and simulation, to the astrophysics of the sources. The programme included an overview of the activities at the ground-based detectors, with which LISA shares not only a common technological heritage but also the prospect of detecting gravitational waves in the next 10 years. Such a conference would not be possible without the help of many people, not least the Scientific Organizing Committee, and the local organization, provided by ESA

  6. The Ikom-Mamfe basin, Nigeria: A study of fracture and mineral vein lineament trends and Cretaceous deformations

    NASA Astrophysics Data System (ADS)

    Oden, M. I.; Egeh, E. U.; Amah, E. A.

    2015-01-01

    The Ikom-Mamfe basin is approximately a 130 km long, east-west abutment onto the eastern flank of the lower Benue trough of Nigeria and extends westwards into Cameroon. Two hundred and six fracture lineaments were analyzed in the Nigerian sector of this basin. They vary in length from 0.5 to 23.75 km, with the most frequently occurring fracture length being about 2.25 km. The most prominent fracture sets have NE-SW and NW-SE orientations, while less prominent patterns are in the NNE-SSW and ESE-WNW directions. NW-SE and NNE-SSW fracture sets are interpreted as "ac" extension fractures from two different deformation episodes, while NE-SW and ESE-WNW sets are "bc" tensile fractures parallel to the axes of F1 and F2 folds, respectively. This implies two deformation episodes in this basin, with the earlier one producing the NE-SW (F1) fold axes, exactly as in the Benue trough. Two prominent mineral vein trends in the basin are the NW-SE and NNE-SSW sets, in which minerals are loaded in "ac" extension fractures. The orientations, lengths and frequency of these lineaments should help in differentiating their ages. The less prominent veins are in the NE-SW and ESE-WNW directions, which are in the "bc" tensile fractures. Early Cretaceous sediments are characterized by NW-SE major and NE-SW minor sets of veins, while the late Cretaceous sequence is characterized by NNE-SSW major and ESE-WNW minor, mainly barite, veins. More than 70% of the barite samples tested gave specific gravity values of 4.2 and above, which is the range specified by the American Petroleum Institution (API) as drilling mud additive or weighting agent. Other vein-filling minerals in this basin are lead ore (galena), zinc ore (sphalerite), pyrite and amethyst, which are altogether subsidiary to barite mineralization.

  7. Correlation of lineaments with soil gas anomalies in the Atlantic Coastal Plain

    SciTech Connect

    Wyatt, D.E. |; Richers, D.

    1994-08-23

    Results from a soil gas survey, performed in the Atlantic Coastal Plain and centered on the US Department of Energy Savannah River Site discovered areas of anomalous (greater than one standard deviation above the mean) methane, ethane, propane, hydrogen and carbon dioxide. A lineament study was performed to investigate whether these anomalies may be associated with fractures or faults and therefore be sourced in basement rocks. The lineament study used a regional aeromagnetic map, various scale topographic maps and a Landsat image. The results of the study suggest the following: (1) correcting for barometric pumping effects, the soil gas anomalies have a strong coincidence with lineations, (2) comparing linear features discernible on a variety of sources mapped at different scales allows for a combined data set to be formed that may define a lineation zone, and (3) linear trends compare favorably with suspected structural trends for the coastal plain.

  8. Analysis of pseudocolor transformations of ERTS-1 images of Southern California area. [geological faults and lineaments

    NASA Technical Reports Server (NTRS)

    Merifield, P. M. (Principal Investigator); Lamar, D. L.; Stratton, R. H.; Lamar, J. V.; Gazley, C., Jr.

    1974-01-01

    The author has identified the following significant results. Representative faults and lineaments, natural features on the Mojave Desert, and cultural features of the southern California area were studied on ERTS-1 images. The relative appearances of the features were compared on a band 4 and 5 subtraction image, its pseudocolor transformation, and pseudocolor images of bands 4, 5, and 7. Selected features were also evaluated in a test given students at the University of California, Los Angeles. Observations and the test revealed no significant improvement in the ability to detect and locate faults and lineaments on the pseudocolor transformations. With the exception of dry lake surfaces, no enhancement of the features studied was observed on the bands 4 and 5 subtraction images. Geologic and geographic features characterized by minor tonal differences on relatively flat surfaces were enhanced on some of the pseudocolor images.

  9. MX Siting Investigation. Faults and Lineaments in the MX Siting Region, Nevada and Utah. Volume I.

    DTIC Science & Technology

    1981-11-06

    subduction of the 6 F tgEi a E-TR-54 7 ancient Farallon plate under the North American plate west of the study area (Atwater, 1970; and Atwater and Molnar... Plates Al through All). Prior to the geologic field reconnaissance phase, the locations of faults and lineaments were transferred from aerial pho...160 m). These data were transferred to the 1:250,000-scale maps (Appendix A, Plates Al through All) by photographic reduction. 3.2 GRAVITY ANALYSIS

  10. Lineament Domain of Regional Strike-Slip Corridor: Insight from the Neogene Transtensional De Geer Transform Fault in NW Spitsbergen

    NASA Astrophysics Data System (ADS)

    Cianfarra, P.; Salvini, F.

    2015-05-01

    Lineaments on regional scale images represent controversial features in tectonic studies. Published models explain the presence of the lineament domains in most geodynamic environments as resulting from the enhanced erosion along strikes normal to the upper crustal regional extension. Despite their success in many tectonic frameworks, these models fail to explain the existing lineament domains in the regional strike-slip corridors that separate regional blocks, including the transform faults. The present paper investigates the lineament distribution in such environments, and specifically presents the results from a study along the shear corridor of the De Geer Transform Fault in the North Atlantic, responsible for the separation and drifting away between Northern Greenland and the Svalbard Archipelago since Oligocene times. The study spans from satellite image analysis and outcrop scale investigations to a more regional analysis on a digital bathymetric model of the North Atlantic-Arctic Ocean. Lineaments were automatically detected in the spectral band 8 (0.52-0.9 μm) of a Landsat 7 image (15 m/pixel resolution). A total of 320 image lineaments were extracted from both the regional and the local scale investigations and statistically analyzed. Results from the multi-scalar lineament analyses revealed the existence of a main N-S lineament domain regionally persistent from the De Geer corridor to the western margin of northern Spitsbergen where it relates to the youngest, post-Oligocene, tectonics observed onshore. This is confirmed by field observations showing that the N-S faults represent the youngest brittle deformation system and systematically cut the deformations associated with the building of the Tertiary West Spitsbergen fold and thrust belt. The N-S lineament domain is the result of the activity of a larger, regional scale tectonic feature, NW-SE oriented and responsible for the localized extension within its deformation corridor, the De Geer Transform

  11. Seismo-Lineament Analysis Method (SLAM) Applied to the South Napa Earthquake

    NASA Astrophysics Data System (ADS)

    Worrell, V. E.; Cronin, V. S.

    2014-12-01

    We used the seismo-lineament analysis method (SLAM; http://bearspace.baylor.edu/Vince_Cronin/www/SLAM/) to "predict" the location of the fault that produced the M 6.0 South Napa earthquake of 24 August 2014, using hypocenter and focal mechanism data from NCEDC (http://www.ncedc.org/ncedc/catalog-search.html) and a digital elevation model from the USGS National Elevation Dataset (http://viewer.nationalmap.gov/viewer/). The ground-surface trace of the causative fault (i.e., the Browns Valley strand of the West Napa fault zone; Bryant, 2000, 1982) and virtually all of the ground-rupture sites reported by the USGS and California Geological Survey (http://www.eqclearinghouse.org/2014-08-24-south-napa/) were located within the north-striking seismo-lineament. We also used moment tensors published online by the USGS and GCMT (http://comcat.cr.usgs.gov/earthquakes/eventpage/nc72282711#scientific_moment-tensor) as inputs to SLAM and found that their northwest-striking seismo-lineaments correlated spatially with the causative fault. We concluded that SLAM could have been used as soon as these mechanism solutions were available to help direct the search for the trace of the causative fault and possible rupture-related damage. We then considered whether the seismogenic fault could have been identified using SLAM prior to the 24 August event, based on the focal mechanisms of smaller prior earthquakes reported by the NCEDC or ISC (http://www.isc.ac.uk). Seismo-lineaments from three M~3.5 events from 1990 and 2012, located in the Vallejo-Crockett area, correlate spatially with the Napa County Airport strand of the West Napa fault and extend along strike toward the Browns Valley strand (Bryant, 2000, 1982). Hence, we might have used focal mechanisms from smaller earthquakes to establish that the West Napa fault is likely seismogenic prior to the South Napa earthquake. Early recognition that a fault with a mapped ground-surface trace is seismogenic, based on smaller earthquakes

  12. FORTRAN 4 programs for summarization and analysis of fracture trace and lineament patterns

    NASA Technical Reports Server (NTRS)

    Podwysocki, M. H.; Lowman, P. D., Jr.

    1974-01-01

    Systematic and detailed analysis of lineament and fracture trace patterns has long been neglected because of the large number of observations involved in such an analysis. Three FORTRAN 4 programs were written to facilitate this manipulation. TRANSFORM converts the initial fracture map data into a format compatible with AZMAP, whose options allow repetitive manipulation of the data for optimization of the analysis. ROSE creates rose diagrams of the fracture patterns suitable for map overlays and tectonic interpretation. Examples are given and further analysis techniques using output from these programs are discussed.

  13. Ancient to Modern History of the Jemez Lineament of North Central New Mexico

    NASA Astrophysics Data System (ADS)

    Magnani, M.; Miller, K. C.; Levander, A.; Karlstrom, K. E.

    2002-12-01

    The continental lithosphere of the southwestern US was derived from mantle sources and accreted to the proto-North American continent, more specifically the Archean-age Wyoming province, during a succession of island arc collisions between 1.8-1.6 Ga. Many of the assembly structures, that is, the sutures between accreted island arcs and oceanic fragments, have been difficult to identify from surface geology. Likewise the tectonic significance of major lineaments in today's lithosphere remains uncertain. The Jemez Lineament (JL), originally defined as an alignment of Tertiary-Quaternary volcanic centers, is a NE trending zone characterized by active uplift, low seismic velocity in the mantle, and repeated reactivation. It also coincides with the southern edge of a 300 km wide transition zone between the Yavapai (1.8-1.7 Ga) and Mazatzal (1.7-1.6 Ga) Proterozoic provinces. This study presents new deep crustal seismic reflection results across the Jemez Lineament of NM, and the Proterozoic Yavapai-Mazatzal terrane boundary. The crust is strongly reflective from the sedimentary column to its base at ~39-42 km. The seismic data show large-scale structures that we interpret as a Proterozoic bi-vergent orogen that extends for at least 170 km laterally and roots into the mantle at the southern edge of the Jemez lineament. A significant portion of this orogen is a 20km thick, south-vergent, crustal duplex occupying at least 50% of the crust south of the JL. North of the JL the depth migrated seismic images show a set of upper crustal north verging recumbent folds and related thrusts in the Proterozoic rocks as well as a south dipping reflectivity in the lower crust that we interpret as one of the north-verging ramps associated with the orogen. Subhorizontal high amplitude reflections across the JL at about 7-15km depth are interpreted to be extensive magmatic intrusions of uncertain age. Based on the seismic and geologic data, we argue that the JL represents both a

  14. Wavy and Cycloidal Lineament Formation on Europa from Combined Diurnal and Nonsynchronous Stresses

    NASA Technical Reports Server (NTRS)

    Gleeson, Damhnait; Crawford, Zane; Barr, Amy C.; Mullen, McCall; Pappalardo, Robert T.; Prockter, Louise M.; Stempel, Michelle M.; Wahr, John

    2005-01-01

    In a companion abstract, we show that fractures propagated into combined diurnal and nonsynchronous rotation (NSR) stress fields can be cycloidal, "wavy," or arcuate in planform as the relative proportion of NSR stress in increased. These transitions occur as NSR stress accumulates over approx. 0 to 10 deg of ice shell rotation, for average fracture propagation speeds of approx. 1 to 3 m/s. Here we consider the NSR speed parameter space for these morphological transitions, and explore the effects on cycloids of adding NSR to diurnal stress. Fitting individual Europan lineaments can constrain the combined NSR plus diurnal stress field at the time of formation.

  15. Deformation and Quaternary Faulting in Southeast Missouri across the Commerce Geophysical Lineament

    USGS Publications Warehouse

    Stephenson, W.J.; Odum, J.K.; Williams, R.A.; Pratt, T.L.; Harrison, R.W.; Hoffman, D.

    1999-01-01

    High-resolution seismic-reflection data acquired at three sites along the surface projection of the Commerce geophysical lineament in southeast Missouri reveal a complex history of post-Cretaceous faulting that has continued into the Quaternary. Near Qulin, Missouri, approximately 20 m of apparent vertical fault displacement has occurred in the Quaternary. Reflection data collected at Idalia Hill, about 45 km to the northeast, reveal a series of reverse and possibly right-lateral strike-slip faults with Quaternary displacement. In the Benton Hills, 45 km northeast of Idalia Hill, seismic data image a complicated series of anticlinal and synclinal fault-bounded blocks immediately north of the Commerce fault. We infer that most of the deformation imaged in the upper 400 m of these three data sets occurred since post-Cretaceous time, and a significant portion of it occurred during Quaternary time. Collectively, these seismic data along with geomorphic and surface-geologic evidence suggest (1) the existence of at least one potential seismogenic structure in southeastern Missouri outside the main zones of New Madrid seismicity, and (2) these structures have been active during the Quaternary. The geographic location of the imaged deformation suggests it is related to structures along with the Commerce geophysical lineament.

  16. Imaging of subsurface lineaments in the southwestern part of the Thrace Basin from gravity data

    NASA Astrophysics Data System (ADS)

    Aydogan, D.; Pinar, A.; Elmas, A.; Bal, O. Tarhan; Yuksel, S.

    2013-04-01

    Linear anomalies, as an indicator of the structural features of some geological bodies, are very important for the interpretation of gravity and magnetic data. In this study, an image processing technique known as the Hough transform (HT) algorithm is described for determining invisible boundaries and extensions in gravity anomaly maps. The Hough function implements the Hough transform used to extract straight lines or circles within two-dimensional potential field images. It is defined as image and Hough space. In the Hough domain, this function transforms each nonzero point in the parameter domain to a sinusoid. In the image space, each point in the Hough space is transformed to a straight line or circle. Lineaments are depicted from these straight lines which are transformed in the image domain. An application of the Hough transform to the Bouguer anomaly map of the southwestern part of the Thrace Basin, NW Turkey, shows the effectiveness of the proposed approach. Based on geological data and gravity data, the structural features in the southwestern part of the Thrace Basin are investigated by applying the proposed approach and the Blakely and Simpson method. Lineaments identified by these approaches are generally in good accordance with previously-mapped surface faults.

  17. Relation of lineaments to sulfide deposits: Bald Eagle Mountain, Centre County, Pennsylvania

    NASA Technical Reports Server (NTRS)

    Mcmurtry, G. J.; Petersen, G. W. (Principal Investigator); Krohn, M. D.; Gold, D. P.

    1975-01-01

    The author has identified the following significant results. Discrete areas of finely-fractured and brecciated sandstone float are present along the crest of Bald Mountain and are commonly sites of sulfide mineralization, as evidenced by the presence of barite and limonite gossans. The frequency distributions of the brecciated float as the negative binomial distribution supports the interpretation of a separate population of intensely fractured material. Such zones of concentrated breccia float have an average width of one kilometer with a range from 0.4 to 1.6 kilometers and were observed in a quarry face to have subvertical dips. Direct spatial correlation of the Landsat-derived lineaments to the fractured areas on the ridge is low; however, the mineralized and fracture zones are commonly assymetrical to the lineament positions. Such a systematic dislocation might result from an inherent bias in the float population or could be the product of the relative erosional resistance of the silicified material in the mineralized areas in relation to the erosionally weak material at the stream gaps.

  18. Lineaments from airborne SAR images and the 1988 Saguenay earthquake, Quebec, Canada

    SciTech Connect

    Roy, D.W.; Schmitt, L.; Woussen, G.; Duberger, R. )

    1993-08-01

    Airborne SAR images provided essential clues to the tectonic setting of (1) the MbLg 6.5 Saguenay earthquake of 25 November 1988, (2) the Charlevoix-Kamouraska seismic source zone, and (3) some of the low *eve* seismic activity in the Eastern seismic background zone of Canada. The event occurred in the southeastern part of the Canadian Shield in an area where the boundary between the Saguenay graben and the Jacques Cartier horst is not well defined. These two tectonic blocks are both associated with the Iapetan St-Lawrence rift. These blocks exhibit several important structural breaks and distinct domains defined by the lineament orientations, densities, and habits. Outcrop observations confirm that several lineament sets correspond to Precambrian ductile shear zones reactivated as brittle faults during the Phanerozoic. In addition, the northeast and southwest limits of recent seismic activity in the Charlevoix-Kamouraska zone correspond to major elements of the fracture pattern identified on the SAR images. These fractures appear to be related to the interaction of the Charlevoix astrobleme with the tectonic features of the area. 20 refs.

  19. Raton-Clayton Volcanic Field magmatism in the context of the Jemez Lineament

    NASA Astrophysics Data System (ADS)

    Schrader, C. M.; Pontbriand, A.

    2013-12-01

    The Raton-Clayton Volcanic Field (RCVF) was active from 9 Ma to approximately 50 Ka and stretches from Raton, New Mexico in the west to Clayton, New Mexico in the east. The field occurs in the Great Plains at the northeastern end of the Jemez Lineament, a major crustal feature and focus of volcanism that extends southwest to the Colorado Plateau in Arizona and encompasses five other major volcanic fields. Jemez Lineament magmatism is temporally related to Rio Grande Rift magmatism, though it extends NE and SW from the rift itself, and it has been suggested that it represents an ancient crustal suture that serves as a conduit for magmatism occurring beneath the larger region of north and central New Mexico (Magnani et al., 2004, GEOL SOC AM BULL, 116:7/8, pp. 1-6). This study extends our work into the RCVF from prior and ongoing work in the Mount Taylor Volcanic Field, where we identified different mantle sources with varying degrees of subduction alteration and we determined some of the crustal processes that contribute to the diversity of magma chemistry and eruptive styles there (e.g., AGU Fall Meeting, abst. #V43D-2884 and #V43D-2883). In the RCVF, we are analyzing multiple phases by electron microprobe and plagioclase phenocrysts and glomerocrysts by LA-ICPMS for Sr isotopes and trace elements. We are undertaking this investigation with the following goals: (1) to evaluate previous magma mixing and crustal assimilation models for Sierra Grande andesites (Zhu, 1995, unpublished Ph.D. dissertation, Rice University; Hesse, 1999, unpublished M.S. thesis, Northern Arizona University); (2) to evaluate subduction-modified mantle as the source for RCVF basanites (specifically those at Little Grande); and (3) to assess the possible role of deep crustal cumulates in buffering transitional basalts. In the larger context, these data will be used to evaluate the varying degree of subduction-modification and the effect of crustal thickness on magmatism along the Jemez

  20. A structural fabric defined by topographic lineaments: Correlation with Tertiary deformation of Ellesmere and Axel Heiberg Islands, Canadian Arctic

    NASA Technical Reports Server (NTRS)

    Oakey, Gordon

    1994-01-01

    Digital topographic contours from four 1:250000 scale maps have been gridded to produce a digital elevation model for part of Ellesmere and Axel Heiberg islands in the Canadian Arctic Islands. Gradient calculations were used to define both east and west dipping slopes defining a pattern of lineaments that have been compared with mapped geological structures. In ice-covered areas, where geological mapping was not possible, well-defined topographic lineaments have been identified and are correlated to extensions of major structural features. The northeast-southwest patterns of both topographic lineaments and mapped structures are strongly unimodal and support a single compressive event oriented at 67 deg west of north. This orientation is coincidental with the convergence direction calculated from the kinematic poles of rotation for Greenland relative to North America between 56 and 35 Ma. A minor secondary peak at 70 east of north is observed for thrust and normal fault solutions and is not directly related to the predicted convergence direction. Whether this represents a unique phase of deformation or is a subcomponent of a single event is not known. The agreement of structural components, lineament orientations, and convergence direction suggests an overwhelming over print of Eurekan deformation on any preexisting structural fabric. This study confirms, for the first time, an excellent compatibility between geological and geophysical constraints for the timing and geometry of the Eurekan orogeny.

  1. Wolf-Darwin lineament and plume-ridge interaction in northern Galápagos

    NASA Astrophysics Data System (ADS)

    Harpp, Karen; Geist, Dennis

    2002-11-01

    The Wolf-Darwin Lineament (WDL), located in the northwestern sector of the Galápagos Archipelago, lies between the focus of the Galápagos hot spot and the Galápagos Spreading Center. Consequently, most researchers have attributed its origin to the interaction between the plume and the adjacent ridge. We propose that the WDL is caused only partially by the plume-ridge interaction, and instead that it is primarily the result of tensional stresses emanating from the inside corner of the transform fault at 91°W. An additional factor that amplifies the tension in this region is the oblique orientation of the major transform fault with respect to the Nazca plate's spreading direction. This setting creates a transtensional zone whereby strain is partitioned into strike-slip motion along the transform and extension throughout the inside corner of the ridge-transform system. The area under tension is magmatic owing to the overlapping effects of the ridge and the Galápagos plume. The extensional model predicts no age-progressive volcanism, which is supported by observed age relationships. The WDL volcanoes define two distinct chemical groups: lavas erupted south of Wolf Island have compositions similar to those produced along the GSC west of 93°W, while those from the northern WDL resemble GSC lavas from the segment directly north of the lineament. This geographic distribution implies that the WDL is supplied by the same type of plume-affected mantle as the segment of the GSC that produced the lithosphere underlying the volcanoes. The observed WDL geochemical gradients are consistent with the extension model; the region under tension simply taps hybrid products of mixing at the margins of the subridge convection system and the periphery of the plume. Essentially, the stress field around the transform fault, normally not observable in a typical midocean ridge setting, is illuminated by the presence of melt from the adjacent hot spot.

  2. A comparison of lineament and fracture trace extraction from LANDSAT ETM+ panchromatic band and panchromatic aerial photograph in Gunungsewu karst area, Java-Indonesia

    NASA Astrophysics Data System (ADS)

    Haryono, E.; Widartono, B. S.; Lukito, H.; Kusumayuda, S. B.

    2016-11-01

    This paper aims at exploring interpretability of the panchromatic band of Landsat Enhanced Thematic Mapper Plus (ETM+) compared to the panchromatic aerial photograph in lineament and fracture trace extraction. Special interest is addressed to karst terrain where lineaments and fracture traces are expressed by aligned valleys and closed depressions. The study area is an single aerial photographic coverage in the Gunungsewu Karst, Java-Indonesia which is characterized by a lineament-controlled cone karst and labyrinth-cone karst. The result shows that the recording time of the Landsat ETM+ with respect to the shadow resulting from the sun illumination angle is the key factor in the performance of lineament and fracture traces extraction. Directional filtering and slicing techniques significantly enhance the lineament interpretability of the panchromatic band of Landsat ETM+. The two methods result in more lineaments and fracture traces which T-test affirm in 0.001 and 0.004 significant levels. Length based lineament analysis attains a better result compared to frequency based analysis.

  3. Parallel rendering

    NASA Technical Reports Server (NTRS)

    Crockett, Thomas W.

    1995-01-01

    This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.

  4. Faults, lineaments, and earthquake epicenters digital map of the Pahute Mesa 30' x 60' Quadrangle, Nevada

    USGS Publications Warehouse

    Minor, S.A.; Vick, G.S.; Carr, M.D.; Wahl, R.R.

    1996-01-01

    This map database, identified as Faults, lineaments, and earthquake epicenters digital map of the Pahute Mesa 30' X 60' quadrangle, Nevada, has been approved for release and publication by the Director of the USGS. Although this database has been subjected to rigorous review and is substantially complete, the USGS reserves the right to revise the data pursuant to further analysis and review. Furthermore, it is released on condition that neither the USGS nor the United States Government may be held liable for any damages resulting from its authorized or unauthorized use. This digital map compilation incorporates fault, air photo lineament, and earthquake epicenter data from within the Pahute Mesa 30' by 60' quadrangle, southern Nye County, Nevada (fig. 1). The compilation contributes to the U.S. Department of Energy's Yucca Mountain Project, established to determine whether or not the Yucca Mountain site is suitable for the disposal of high-level nuclear waste. Studies of local and regional faulting and earthquake activity, including the features depicted in this compilation, are carried out to help characterize seismic hazards and tectonic processes that may be relevant to the future stability of Yucca Mountain. The Yucca Mountain site is located in the central part of the Beatty 30' by 60' quadrangle approximately 15 km south of the south edge of the Pahute Mesa quadrangle (fig. 1). The U.S. Geological Survey participates in studies of the Yucca Mountain site under Interagency Agreement DE-AI08-78ET44802. The map compilation is only available on line as a digital database in ARC/INFO ASCII (Generate) and export formats. The database can be downloaded via 'anonymous ftp' from a USGS system named greenwood.cr.usgs.gov (136.177.48.5). The files are located in a directory named /pub/open-file-reports/ofr-96-0262. This directory contains a text document named 'README.1 ST' that contains database technical and explanatory documentation, including instructions for

  5. New improvements for lineaments study of gravity data with improved Euler inversion and phase congruency of the field data

    NASA Astrophysics Data System (ADS)

    Wang, Jun; Meng, Xiaohong; Li, Fang

    2017-01-01

    Lineaments study is a routine practice for the interpretation of gravity data. It can be utilized to visualize features of geologic structures and is important for a wide range of geological tasks. Over the past years, a variety of methods based on the use of derivatives of gravity data have been proposed to determine the lineaments. However, interpretation outcomes of gravity data are found to be highly variable among interpreters and lack consistency even within an individual. This paper presents new ways to improve the confidence of structural interpretation by improved Euler inversion and phase congruency of the field data. The improved Euler inversion is independent of the structural index. Meanwhile, the principal component analysis (PCA) technique is incorporated to improve computational stability. Moreover, we find the feature strengths of ridges, valleys and edges within data by analyzing their local frequencies. Then, all of the lineaments detected by these two methods are analyzed comprehensively to select the appropriate solutions. With the proposed computation scheme, lineaments in the study area can be divided into different types with different reliabilities. This division makes it easier for geologists to give more detailed interpretations. Tests on synthetic data showed that the proposed techniques can be utilized to compute Euler inversions correctly and the selected solutions are in good accordance with the true positon of theoretical model. We also applied the proposed methods to real gravity data recorded from a metallic mineral deposit area in the middle-lower reaches of the Yangtze River in China. The results obtained with the proposed methods showed good agreement with geological interpretation.

  6. The tectonic and volcanic history of Mercury as inferred from studies of scarps, ridges, troughs, and other lineaments

    NASA Technical Reports Server (NTRS)

    Dzurisin, D.

    1978-01-01

    Tectonic and volcanic modification of the Mercurian surface is discussed, and nine landform classes are defined. An evolutionary chronology, based on reported interpretations of scarp, ridge, trough, and other lineament features, is presented, and the roles (in chronological order) of accretion and differentiation, tidal spindown, plains volcanism, heavy bombardment, cooling/contraction, Caloris impact, intense surface modification, basin subsidence, local plains volcanism, isostatic basin uplift, and light cratering are considered. The observational data were obtained by Mariner 10.

  7. Statistical analysis of surface lineaments and fractures for characterizing naturally fractured reservoirs

    SciTech Connect

    Guo, Genliang; George, S.A.; Lindsey, R.P.

    1997-08-01

    Thirty-six sets of surface lineaments and fractures mapped from satellite images and/or aerial photos from parts of the Mid-continent and Colorado Plateau regions were collected, digitized, and statistically analyzed in order to obtain the probability distribution functions of natural fractures for characterizing naturally fractured reservoirs. The orientations and lengths of the surface linear features were calculated using the digitized coordinates of the two end points of each individual linear feature. The spacing data of the surface linear features within an individual set were, obtained using a new analytical sampling technique. Statistical analyses were then performed to find the best-fit probability distribution functions for the orientation, length, and spacing of each data set. Twenty-five hypothesized probability distribution functions were used to fit each data set. A chi-square goodness-of-fit test was used to rank the significance of each fit. A distribution which provides the lowest chi-square goodness-of-fit value was considered the best-fit distribution. The orientations of surface linear features were best-fitted by triangular, normal, or logistic distributions; the lengths were best-fitted by PearsonVI, PearsonV, lognormal2, or extreme-value distributions; and the spacing data were best-fitted by lognormal2, PearsonVI, or lognormal distributions. These probability functions can be used to stochastically characterize naturally fractured reservoirs.

  8. Applicability of ERTS-1 to lineament and photogeologic mapping in Montana: Preliminary report

    NASA Technical Reports Server (NTRS)

    Weidman, R. M.; Alt, D. D.; Flood, R. E.; Hawley, K. T.; Wackwitz, L. K.; Berg, R. B.; Johns, W. M.

    1973-01-01

    A lineament map prepared from a mosaic of western Montana shows about 85 lines not represented on the state geologic map, including elements of a northeast-trending set through central western Montana which merit ground truth checking and consideration in regional structural analysis. Experimental fold annotation resulted in a significant local correction to the state geologic map. Photogeologic mapping studies produced only limited success in identification of rock types, but they did result in the precise delineation of a late Cretaceous or early Tertiary volcanic field (Adel Mountain field) and the mapping of a connection between two granitic bodies shown on the state map. Imagery was used successfully to map clay pans associated with bentonite beds in gently dipping Bearpaw Shale. It is already apparent that ERTS imagery should be used to facilitate preparation of a much needed statewide tectonic map and that satellite imagery mapping, aided by ground calibration, provides and economical means to discover and correct errors in the state geologic map.

  9. Enhancement of Landsat images for lineament analysis in the area of the Salina Basin, New York and Pennsylvania

    USGS Publications Warehouse

    Krohn, M. Dennis

    1979-01-01

    Digital image processing of Landsat images of New York and Pennsylvania was undertaken to provide optimum images for lineament analysis in the area of the Salina Basin. Preliminary examination of Landsat images from photographic prints indicated sufficient differences between the spectral bands of the Landsat Multispectral Scanner (MSS) to warrant digital processing of both MSS band 7 and MSS band 5. Selective contrast stretching based on analysis of the Landsat MSS histograms proved to be the most important factor affecting the appearance of the images. A three-point linear stretch using the two end points and a middle point to the Landsat frequency distribution was most successful. The flexibility of the REMAPP image processing system was helpful in creating such custom-tailored stretches. An edge enhancement was tested on the MSS band 5 image, but was not used. Stereoscopic Landsat images acquired from adjacent orbits aided recognition of topographic features; the area of stereoscopic coverage could be increased by utilizing the precession of Landsat-1?s orbit . Improvements in the digitally processed scenes did affect the analysis of lineaments for the New York area; on the enhanced MSS band 5 image, an ENE trending set of lineaments is visible, which was not recognized from other images.

  10. An interdisciplinary analysis of multispectral satellite data for selected cover types in the Colorado Mountains, using automatic data processing techniques. [geological lineaments and mineral exploration

    NASA Technical Reports Server (NTRS)

    Hoffer, R. M. (Principal Investigator)

    1974-01-01

    The author has identified the following significant results. One capability which has been recognized by many geologists working with space photography is the ability to see linear features and alinements which were previously not apparent. To the exploration geologist, major lineaments seen on satellite images are of particular interest. A portion of ERTS-1 frame 1407-17193 (3 Sept. 1973) was used for mapping lineaments and producing an iso-lineament intersection map. Skylab photography over the area of prime area was not useable due to snow cover. Once the lineaments were mapped, a grid with 2.5 km spacing was overlayed on the map and the lineament intersections occurring within each grid square were counted and the number plotted in the center of the grid square. These numbers were then contoured producing a contour map of equal lineament intersection. It is believed that the areas of high intersection concentration would be the most favorable area for ore mineralization if favorable host rocks are also present. These highly fractured areas would act as conduits for carrying the ore forming solutions to the site of deposition in a favorable host rock. Two of the six areas of high intersection concentration are over areas of present or past mining camps and small claims are known to exist near the others. These would be prime target areas for future mineral exploration.

  11. Active faulting on the Wallula fault within the Olympic-Wallowa Lineament (OWL), eastern Washington State

    NASA Astrophysics Data System (ADS)

    Sherrod, B. L.; Lasher, J. P.; Barnett, E. A.

    2013-12-01

    Several studies over the last 40 years focused on a segment of the Wallula fault exposed in a quarry at Finley, Washington. The Wallula fault is important because it is part of the Olympic-Wallowa lineament (OWL), a ~500-km-long topographic and structural lineament extending from Vancouver Island, British Columbia to Walla Walla, Washington that accommodates Basin and Range extension. The origin and nature of the OWL is of interest because it contains potentially active faults that are within 50 km of high-level nuclear waste facilities at the Hanford Site. Mapping in the 1970's and 1980's suggested the Wallula fault did not offset Holocene and late Pleistocene deposits and is therefore inactive. New exposures of the Finley quarry wall studied here suggest otherwise. We map three main packages of rocks and sediments in a ~10 m high quarry exposure. The oldest rocks are very fine grained basalts of the Columbia River Basalt Group (~13.5 Ma). The next youngest deposits include a thin layer of vesicular basalt, white volcaniclastic deposits, colluvium containing clasts of vesicular basalt, and indurated paleosols. A distinct angular unconformity separates these vesicular basalt-bearing units from overlying late Pleistocene flood deposits, two colluvium layers containing angular clasts of basalt, and Holocene tephra-bearing loess. A tephra within the loess likely correlates to nearby outcrops of Mazama ash. We recognize three styles of faults: 1) a near vertical master reverse or oblique fault juxtaposing very fine grained basalt against late Tertiary-Holocene deposits, and marked by a thick (~40 cm) vertical seam of carbonate cemented breccia; 2) subvertical faults that flatten upwards and displace late Tertiary(?) to Quaternary(?) soils, colluvium, and volcaniclastic deposits; and 3) flexural slip faults along bedding planes in folded deposits in the footwall. We infer at least two Holocene earthquakes from the quarry exposure. The first Holocene earthquake deformed

  12. Parallel machines: Parallel machine languages

    SciTech Connect

    Iannucci, R.A. )

    1990-01-01

    This book presents a framework for understanding the tradeoffs between the conventional view and the dataflow view with the objective of discovering the critical hardware structures which must be present in any scalable, general-purpose parallel computer to effectively tolerate latency and synchronization costs. The author presents an approach to scalable general purpose parallel computation. Linguistic Concerns, Compiling Issues, Intermediate Language Issues, and hardware/technological constraints are presented as a combined approach to architectural Develoement. This book presents the notion of a parallel machine language.

  13. Parallel pipelining

    SciTech Connect

    Joseph, D.D.; Bai, R.; Liao, T.Y.; Huang, A.; Hu, H.H.

    1995-09-01

    In this paper the authors introduce the idea of parallel pipelining for water lubricated transportation of oil (or other viscous material). A parallel system can have major advantages over a single pipe with respect to the cost of maintenance and continuous operation of the system, to the pressure gradients required to restart a stopped system and to the reduction and even elimination of the fouling of pipe walls in continuous operation. The authors show that the action of capillarity in small pipes is more favorable for restart than in large pipes. In a parallel pipeline system, they estimate the number of small pipes needed to deliver the same oil flux as in one larger pipe as N = (R/r){sup {alpha}}, where r and R are the radii of the small and large pipes, respectively, and {alpha} = 4 or 19/7 when the lubricating water flow is laminar or turbulent.

  14. A multi-directional and multi-scale roughness filter to detect lineament segments on digital elevation models - analyzing spatial objects in R

    NASA Astrophysics Data System (ADS)

    Baumann, Sebastian; Robl, Jörg; Wendt, Lorenz; Willingshofer, Ernst; Hilberg, Sylke

    2016-04-01

    Automated lineament analysis on remotely sensed data requires two general process steps: The identification of neighboring pixels showing high contrast and the conversion of these domains into lines. The target output is the lineaments' position, extent and orientation. We developed a lineament extraction tool programmed in R using digital elevation models as input data to generate morphological lineaments defined as follows: A morphological lineament represents a zone of high relief roughness, whose length significantly exceeds the width. As relief roughness any deviation from a flat plane, defined by a roughness threshold, is considered. In our novel approach a multi-directional and multi-scale roughness filter uses moving windows of different neighborhood sizes to identify threshold limited rough domains on digital elevation models. Surface roughness is calculated as the vertical elevation difference between the center cell and the different orientated straight lines connecting two edge cells of a neighborhood, divided by the horizontal distance of the edge cells. Thus multiple roughness values depending on the neighborhood sizes and orientations of the edge connecting lines are generated for each cell and their maximum and minimum values are extracted. Thereby negative signs of the roughness parameter represent concave relief structures as valleys, positive signs convex relief structures as ridges. A threshold defines domains of high relief roughness. These domains are thinned to a representative point pattern by a 3x3 neighborhood filter, highlighting maximum and minimum roughness peaks, and representing the center points of lineament segments. The orientation and extent of the lineament segments are calculated within the roughness domains, generating a straight line segment in the direction of least roughness differences. We tested our algorithm on digital elevation models of multiple sources and scales and compared the results visually with shaded relief map

  15. Reconstructing the 3D fracture distribution model from core (10 cm) to outcrop (10 m) and lineament (10 km) scales

    NASA Astrophysics Data System (ADS)

    Darcel, C.; Davy, P.; Bour, O.; de Dreuzy, J.

    2006-12-01

    Considering the role of fractures in hydraulic flow, the knowledge of the 3D spatial distribution of fractures is a basic concern for any hydrogeology-related study (potential leakages in waste repository, aquifer management, ?). Unfortunately geophysical imagery is quite blind with regard to fractures, and only the largest ones are generally detected, if they are. Actually most of the information has to be derived from statistical models whose parameters are defined from a few sparse sampling areas, such as wells, outcrops, or lineament maps. How these observations obtained at different scales can be linked to each other is a critical point, which directly addresses the issue of fracture scaling. In this study, we use one of the most important datasets that have ever been collected for characterizing fracture networks. It was collected by the Swedish company SKB for their research program on deep repository for radioactive waste, and consists of large-scale lineament maps covering about 100 km2, several outcrops of several hundreds of m2 mapped with a fracture trace length resolution down to 0.50 m, and a series of 1000m-deep cored boreholes where both fracture orientations and fracture intensities were carefully recorded. Boreholes are an essential complement to surface outcrops as they allow the sampling of horizontal fracture planes that, generally, are severely undersampled in subhorizontal outcrops. Outcrops, on the other hand, provide information on fracture sizes which is not possible to address from core information alone. However linking outcrops and boreholes is not straightforward: the sampling scale is obviously different and some scaling rules have to be applied to relate both fracture distributions; outcrops are 2D planes while boreholes are mostly 1D records; outcrops can be affected by superficial fracturing processes that are not representative of the fracturing at depth. We present here the stereology methods for calculating the 3D distribution

  16. Interpretation of high resolution aeromagnetic data for lineaments study and occurrence of Banded Iron Formation in Ogbomoso area, Southwestern Nigeria

    NASA Astrophysics Data System (ADS)

    Oladunjoye, Michael Adeyinka; Olayinka, Abel Idowu; Alaba, Mustapha; Adabanija, Moruffdeen Adedapo

    2016-02-01

    The quest for solid mineral resource as an alternative for oil income in Nigeria presents opportunity to diversify the resource base of the country. To fill some information gap on the long abandoned Ajase and Gbede Banded Iron Formations (BIF) in Ogbomoso area, Southwestern Nigeria, high resolution aeromagnetic data of Ogbomoso - Sheet 222 was interpreted; to provide a better understanding of the mode of occurrence of the iron ore and associated structural features and geologic model. These were accomplished by subjecting reduced-to-pole (RTP) residual aeromagnetic intensity map to various data filtering and processing involving total horizontal derivative, vertical derivative, Upward Continuation (UC), Downward Continuation (DC), Euler Deconvolution at different Spectral Indices (SI), and Analytical signal using Geosoft Oasis Montaj 6.4.2 (HJ) data processing and analysis software. The resultants maps were overlain, compared and or plotted on RTP residual aeromagnetic intensity map and or geological map and interpreted in relation to the surface geological map. Positive magnetic anomalies observed on the RTP residual aeromagnetic intensity map ranged from 2.1 to 94.0 nT and associated with contrasting basement rocks, Ajase and Gbede BIF; while negative magnetic anomalies varied between -54.7 nT and -2.8 nT and are associated with intrusive bodies. Interpreted lineaments obtained from total horizontal derivative map were separated into two categories namely ductile and brittle based on their character vis-à-vis magnetic anomalies on RTP intensity map. Whilst the brittle lineaments were interpreted as fracture or faults; the ductile lineaments were interpreted as folds or representing the internal fabric of the rock units. In addition prominent magnetic faults mainly due to offset of similar magnetic domain/gradient were also interpreted. The iron ore mineralization is distributed within the eastern portion of the study area with Ajase BIF at relatively greater

  17. Study of the structure changes caused by earthquakes in Chile applying the lineament analysis to the Aster (Terra) satellite data.

    NASA Astrophysics Data System (ADS)

    Arellano-Baeza, A.; Zverev, A.; Malinnikov, V.

    Chile is one of the most seismically and volcanically active regions in the South America due to a constant subdiction of the South American plate, converging with the Nazca plate in the extreme North of Chile. Four events, namely: the Ovalle earthquake of Juny 18, 2003, M=6.3, with epicenter localized at (-30:49:33, -71:18:53), the Calama earthquake of Junly 19, 2001, M=5.2, (-30:29:38,-68:33:18), the Pica earthquake of April 10, 2003, M=5.1, (-21:03:20,-68:47:10) and the La Ligua earthquake of May 6, 2001, M=5.1, (-32:35:31,-71:07:58:) were analysed using the 15 m resolution satellite images, provided by the ASTER/VNIR instrument. The Lineament Extraction and Stripes Statistic Analysis (LESSA) software package was used to examine changes in the lineament features caused by sismic activity. Lack of vegetation facilitates the study of the changes in the topography common to all events and makes it possible to evaluate the sismic risk in this region for the future.

  18. Automatic Extraction of tectonic lineaments from high and medium resolution remote sensing data in the Hindukush-Pamir.

    NASA Astrophysics Data System (ADS)

    Rahnama, Mehdi; Gloaguen, Richard

    2013-04-01

    We propose to investigate neotectonic activity and deformation in north-east of the Afghanistan and south-east Tajikistan using remote sensing. We analyse geomorphic features such as stream profiles and drainage patterns and combine it with lineament analysis from high and medium resolution satellite data. For this purpose, we developed a new toolbox for the Automatic Lineaments Feature Extraction . Considering both classical and recently-reported edge detection methods, we chose 4 filter types: Sobel, LoG, Canny and Prewitt and then used the Hough transform for the identification of linear features. We finally merge close line segments with similar directions using polynomial curve fitting. The modern deformation, the fault movements" and the induced earthquakes in Afghanistan and Tajikistan are driven by the collision between the northward-moving Indian subcontinent and Eurasia. The general patterns and orientations of faults and the styles of deformation that we interpret from the imagery are consistent with the styles of faulting determined from focal mechanisms of historical earthquakes. With these techniques we are able to assess the activity of faults otherwise inaccessible. We show that the SW-Pamir is largely controlled by the Chaman-Herat fault system and, to a lesser extent by the Darvaz fault zone.

  19. Automatic derivation of natural and artificial lineaments from ALS point clouds in floodplains

    NASA Astrophysics Data System (ADS)

    Mandlburger, G.; Briese, C.

    2009-04-01

    Water flow is one of the most important driving forces in geomorphology and river systems have ever since formed our landscapes. With increasing urbanisation fertile flood plains were more and more cultivated and the defence of valuable settlement areas by dikes and dams became an important issue. Today, we are dealing with landscapes built up by natural as well as man-made artificial forces. In either case the general shape of the terrain can be portrayed by lineaments representing discontinuities of the terrain slope. Our contribution, therefore, presents an automatic method for delineating natural and artificial structure lines based on randomly distributed point data with high density of more than one point/m2. Preferably, the last echoes of airborne laser scanning (ALS) point clouds are used, since the laser signal is able to penetrate vegetation through small gaps in the foliage. Alternatively, point clouds from (multi) image matching can be employed, but poor ground point coverage in vegetated areas is often the limiting factor. Our approach is divided into three main steps: First, potential 2D start segments are detected by analyzing the surface curvature in the vicinity of each data point, second, the detailed 3D progression of each structure line is modelled patch-wise by intersecting surface pairs (e.g. planar patch pairs) based on the detected start segments and by performing line growing and, finally, post-processing like line cleaning, smoothing and networking is carried out in a last step. For the initial detection of start segments a best fitting two dimensional polynomial surface (quadric) is computed in each data point based on a set of neighbouring points, from which the minimum and maximum curvature is derived. Patches showing high maximum and low minimum curvatures indicate linear discontinuities in the surface slope and serve as start segments for the subsequent 3D modelling. Based on the 2D location and orientation of the start segments

  20. Tectonic lineaments in the cenozoic volcanics of southern Guatemala: Evidence for a broad continental plate boundary zone

    NASA Technical Reports Server (NTRS)

    Baltuck, M.; Dixon, T. H.

    1984-01-01

    The northern Caribbean plate boundary has been undergoing left lateral strike slip motion since middle Tertiary time. The western part of the boundary occurs in a complex tectonic zone in the continental crust of Guatemala and southernmost Mexico, along the Chixoy-Polochic, Motogua and possibly Jocotan-Chamelecon faults. Prominent lineaments visible in radar imagery in the Neogene volcanic belt of southern Guatemala and western El Salvador were mapped and interpreted to suggest southwest extensions of this already broad plate boundary zone. Because these extensions can be traced beneath Quaternary volcanic cover, it is thought that this newly mapped fault zone is active and is accommodating some of the strain related to motion between the North American and Caribbean plates. Onshore exposures of the Motoqua-Polochic fault systems are characterized by abundant, tectonically emplaced ultramafic rocks. A similar mode of emplacement for these off shore ultramafics, is suggested.

  1. Preliminary Use of the Seismo-Lineament Analysis Method (SLAM) to Investigate Seismogenic Faulting in the Grand Canyon Area, Northern Arizona

    NASA Astrophysics Data System (ADS)

    Cronin, V. S.; Cleveland, D. M.; Prochnow, S. J.

    2007-12-01

    This is a progress report on our application of the Seismo-Lineament Analysis Method (SLAM) to the eastern Grand Canyon area of northern Arizona. SLAM is a new integrated method for identifying potentially seismogenic faults using earthquake focal-mechanism solutions, geomorphic analysis and field work. There are two nodal planes associated with any double-couple focal-mechanism solution, one of which is thought to coincide with the fault that produced the earthquake; the slip vector is normal to the other (auxiliary) plane. When no uncertainty in the orientation of the fault-plane solution is reported, we use the reported vertical and horizontal uncertainties in the focal location to define a tabular uncertainty volume whose orientation coincides with that of the fault-plane solution. The intersection of the uncertainty volume and the ground surface (represented by the DEM) is termed a seismo-lineament. An image of the DEM surface is illuminated perpendicular to the strike of the seismo- lineament to accentuate geomorphic features within the seismo-lineament that may be related to seismogenic faulting. This evaluation of structural geomorphology is repeated for several different azimuths and elevations of illumination. A map is compiled that includes possible geomorphic indicators of faulting as well as previously mapped faults within each seismo-lineament, constituting a set of hypotheses for the possible location of seismogenic fault segments that must be evaluated through fieldwork. A fault observed in the field that is located within a seismo-lineament, and that has an orientation and slip characteristics that are statistically compatible with the fault-plane solution, is considered potentially seismogenic. We compiled a digital elevation model (DEM) of the Grand Canyon area from published data sets. We used earthquake focal-mechanism solutions produced by David Brumbaugh (2005, BSSA, v. 95, p. 1561-1566) for five M > 3.5 events reported between 1989 and 1995

  2. Parallel pivoting combined with parallel reduction

    NASA Technical Reports Server (NTRS)

    Alaghband, Gita

    1987-01-01

    Parallel algorithms for triangularization of large, sparse, and unsymmetric matrices are presented. The method combines the parallel reduction with a new parallel pivoting technique, control over generations of fill-ins and a check for numerical stability, all done in parallel with the work being distributed over the active processes. The parallel technique uses the compatibility relation between pivots to identify parallel pivot candidates and uses the Markowitz number of pivots to minimize fill-in. This technique is not a preordering of the sparse matrix and is applied dynamically as the decomposition proceeds.

  3. Comparative lineament analysis using SIR-C radar, Landsat TM and stereo LFC photographs for assessing groundwater resources, Red Sea Hills area, Sudan

    SciTech Connect

    Koch, M.; Mather, P.M.; Leason, A.

    1996-08-01

    This paper describes preliminary results from a comparative investigation of lineament mapping from stereoscopic LFC and from SIR-C L and C band synthetic aperture radar data. The lineament patterns are used together with other spatial data sets describing lithology and geomorphic characteristics in order to test a model of groundwater flow in the semi-arid Red Sea Hills area of Sudan described by Koch (1993). Initial results show that the LFC imagery is most useful for mapping detailed fracture patterns while the combination of L and C bands (total power) of the SIR-C synthetic aperture radar is helpful in the location of major deep-seated fracture zones. L-band SAR data together with a false-colour Landsat TM composite show the presence of subsurface moisture and vegetation respectively. These results are discussed within the context of a hydrogeological model.

  4. Active tectonics on Deception Island (West-Antarctica): A new approach by using the fractal anisotropy of lineaments, fault slip measurements and the caldera collapse shape

    USGS Publications Warehouse

    Pérez-López, R.; Giner-Robles, J.L.; Martínez-Díaz, J.J.; Rodríguez-Pascua, M.A.; Bejar, M.; Paredes, C.; González-Casado, J.M.

    2007-01-01

    The tectonic field on Deception Island (South Shetlands, West Antarctica) is determined from structural and fractal analyses. Three different analyses are applied to the study of the strain and stress fields in the area: (1) field measurements of faults (strain analysis), (2) fractal geometry of the spatial distribution of lineaments and (3) the caldera shape (stress analyses). In this work, the identified strain field is extensional with the maximum horizontal shortening trending NE-SW and NW-SE. The fractal technique applied to the spatial distribution of lineaments indicates a stress field with SHMAX oriented NE-SW. The elliptical caldera of Deception Island, determined from field mapping, satellite imagery, vents and fissure eruptions, has an elongate shape and a stress field with SHMAX trending NE-SW.

  5. Mariner 10 data analysis. 1: Scarps, ridges, troughs, and other lineaments on Mercury. 2: Geologic significance of photometric variations on Mercury. Ph.D. Thesis

    NASA Technical Reports Server (NTRS)

    Dzurisin, D.

    1977-01-01

    Volcanic and tectonic implications of the surface morphology of Mercury are discussed. Mercurian scarps, ridges, troughs, and other lineaments are described and classified as planimetrically linear, arcuate, lobate, or irregular. A global pattern of lineaments is interpreted to reflect modification of linear crustal joints formed in response to stresses induced by tidal spindown. Large arcuate scarps on Mercury most likely record a period of compressional tectonism near the end of heavy bombardment. Shrinkage owing to planetary cooling is the mechanism preferred for their production. Measurements of local normal albedo are combined with computer-generated photometric maps of Mercury to provide constraints on the nature of surface materials and processes. If the mercurian surface obeys the average lunar photometric function, its normal albedo at 554 nm is .16 + or - .03.

  6. Geobotanical and lineament analysis of sandsat satellite imagery for hydrocarbon microseeps

    SciTech Connect

    Warner, T.A.

    1997-10-01

    Both geobotanical and structural interpretations of remotely sensed data tend to be plagued by random associations. However, a combination of these methods has the potential to provide a methodology for excluding many false associations. To test this approach, a test site in West Virginia has been studied using remotely sensed and field data. The historic Volcano Oil Field, in Wood, Pleasants and Ritchie Counties was known as an area of hydrocarbon seeps in the last century. Although pressures in the reservoir are much reduced today, hydrocarbons remain in the reservoir. An examination of a multi-seasonal Landsat Thematic Mapper imagery has shown little difference between the forests overlying the hydrocarbon reservoirs compared to the background areas, with the exception of an image in the very early fall. This image has been enhanced using an nPDF spectral transformation that maximizes the contrast between the anomalous and background areas. A field survey of soil gas chemistry showed that hydrocarbon concentration is generally higher over the anomalous region. In addition, soil gas hydrocarbon concentration increases with proximity to linear features that cross the strike of the overall structure of the reservoir. Linear features that parallel the strike, however, do not have any discernible influence on gas concentration. Field spectral measurements were made periodically through the summer and early fall to investigate the origin of the spectral reflectance anomaly. Measurements were made with a full-range spectro-radiometer (400 nm to 2500 nm) on a number of different species, both on and off the spectral anomaly. The results lend support to the finding that in the early fall spectral reflectance increases in the near infrared and mid infrared in the spectrally anomalous regions.

  7. Special parallel processing workshop

    SciTech Connect

    1994-12-01

    This report contains viewgraphs from the Special Parallel Processing Workshop. These viewgraphs deal with topics such as parallel processing performance, message passing, queue structure, and other basic concept detailing with parallel processing.

  8. Interpretation of gravity data by the continuous wavelet transform: The case of the Chad lineament (North-Central Africa)

    NASA Astrophysics Data System (ADS)

    Li, Yuanyuan; Braitenberg, Carla; Yang, Yushan

    2013-03-01

    A slightly bended gravity high along the Chad lineament in Central North Africa is analyzed and interpreted by the continuous wavelet transform (CWT) method. We use scale normalization on the continuous wavelet transform, allowing analysis of the gravity field in order to determine the sources at different depths. By focusing on homogenous standard sources, such as sphere or cube, horizontal cylinder or prism, sheet and infinite step, we derive the relationships between the source depth and pseudo-wavenumber. Then the source depth can be recovered from tracing the maximal values of the modulus of the complex wavelet coefficients in the CWT-based scalograms that are function of the pseudo-wavenumber. The studied area includes a central gravity high up to 75 km wide, and a secondary high that occurs at the southern part of the anomaly. The interpretation of the depth slices and vertical sections of the modulus maxima of the complex wavelet coefficients allows recognition of a relatively dense terrane located at middle crustal levels (10-25 km depth). A reasonable geological model derived from the 2.5D gravity forward modelling indicates the presence of high density bodies, probably linked to a buried suture, which were thrusted up into the mid-crust during the Neo-Proterozoic terrane collisions between the Saharan metacraton and the Arabian-Nubian shield. We conclude that the Chad line delineates a first order geological boundary, missing on the geologic maps.

  9. The Alegre Lineament and its role over the tectonic evolution of the Campos Basin and adjacent continental margin, Southeastern Brazil

    NASA Astrophysics Data System (ADS)

    Calegari, Salomão Silva; Neves, Mirna Aparecida; Guadagnin, Felipe; França, George Sand; Vincentelli, Maria Gabriela Castillo

    2016-08-01

    The structural framework and tectonic evolution of the sedimentary basins along the eastern margin of the South American continent are closely associated with the tectonic framework and crustal heterogeneities inherited from the Precambrian basement. However, the role of NW-SE and NNW-SSE structures observed at the outcropping basement in Southeastern Brazil and its impact over the development of those basins have not been closely investigated. In the continental region adjacent to the Campos Basin, we described a geological feature with NNW-SSE orientation, named in this paper as the Alegre Fracture Zone (AFZ), which is observed in the onshore basement and can be projected to the offshore basin. The main goal of this work was to study this structural lineament and its influence on the tectonic evolution of the central portion of the Campos Basin and adjacent mainland. The onshore area was investigated through remote sensing data joint with field observations, and the offshore area was studied through the interpretation of 2-D seismic data calibrated by geophysical well logs. We concluded that the AFZ occurs in both onshore and offshore as a brittle deformation zone formed by multiple sets of fractures that originated in the Cambrian and were reactivated mainly as normal faults during the rift phase and in the Cenozoic. In the Campos Basin, the AFZ delimitates the western side of the Corvina-Parati Low, composing a complex fault system with the NE-SW faults and the NW-SE transfer faults.

  10. Parallel rendering techniques for massively parallel visualization

    SciTech Connect

    Hansen, C.; Krogh, M.; Painter, J.

    1995-07-01

    As the resolution of simulation models increases, scientific visualization algorithms which take advantage of the large memory. and parallelism of Massively Parallel Processors (MPPs) are becoming increasingly important. For large applications rendering on the MPP tends to be preferable to rendering on a graphics workstation due to the MPP`s abundant resources: memory, disk, and numerous processors. The challenge becomes developing algorithms that can exploit these resources while minimizing overhead, typically communication costs. This paper will describe recent efforts in parallel rendering for polygonal primitives as well as parallel volumetric techniques. This paper presents rendering algorithms, developed for massively parallel processors (MPPs), for polygonal, spheres, and volumetric data. The polygon algorithm uses a data parallel approach whereas the sphere and volume render use a MIMD approach. Implementations for these algorithms are presented for the Thinking Ma.chines Corporation CM-5 MPP.

  11. Variable near-surface deformation along the Commerce segment of the Commerce geophysical lineament, southeast Missouri to southern Illinois, USA

    USGS Publications Warehouse

    Odum, J.K.; Stephenson, W.J.; Williams, R.A.

    2003-01-01

    Recent studies have demonstrated a plausible link between surface and near-surface tectonic features and the vertical projection of the Commerce geophysical lineament (CGL). The CGL is a 5- to 10-km-wide zone of basement magnetic and gravity anomalies traceable for more than 600 km, extending from Arkansas through southeast Missouri and southern Illinois and into Indiana. Twelve kilometers of high-resolution seismic reflection data, collected at four sites along a 175-km segment of the CGL projection, are interpreted to show varying amounts of deformation involving Tertiary and some Quaternary sediments. Some of the locally anomalous geomorphic features in the northern Mississippi embayment region (i.e., paleoliquefaction features, anomalous directional changes in stream channels, and areas of linear bluff escarpments) overlying the CGL can be correlated with specific faults and/or narrow zones of deformed (faulted and folded) strata that are imaged on high-resolution seismic reflection data. There is an observable change in near-surface deformation style and complexity progressing from the southwest to the northeast along the trace of the CGL. The seismic reflection data collaborate mapping evidence which suggests that this region has undergone a complex history of deformation, some of which is documented to be as young as Quaternary, during multiple episodes of reactivation under varying stress fields. This work, along with that of other studies presented in this volume, points to the existence of at least one major crustal feature outside the currently defined zone of seismic activity (New Madrid Seismic Zone) that should be considered as a significant potential source zone for seismogenic activity within the midcontinent region of the United States. ?? 2003 Elsevier B.V. All rights reserved.

  12. Variable near-surface deformation along the Commerce segment of the Commerce geophysical lineament, southeast Missouri to southern Illinois, USA

    NASA Astrophysics Data System (ADS)

    Odum, Jack K.; Stephenson, William J.; Williams, Robert A.

    2003-06-01

    Recent studies have demonstrated a plausible link between surface and near-surface tectonic features and the vertical projection of the Commerce geophysical lineament (CGL). The CGL is a 5- to 10-km-wide zone of basement magnetic and gravity anomalies traceable for more than 600 km, extending from Arkansas through southeast Missouri and southern Illinois and into Indiana. Twelve kilometers of high-resolution seismic reflection data, collected at four sites along a 175-km segment of the CGL projection, are interpreted to show varying amounts of deformation involving Tertiary and some Quaternary sediments. Some of the locally anomalous geomorphic features in the northern Mississippi embayment region (i.e., paleoliquefaction features, anomalous directional changes in stream channels, and areas of linear bluff escarpments) overlying the CGL can be correlated with specific faults and/or narrow zones of deformed (faulted and folded) strata that are imaged on high-resolution seismic reflection data. There is an observable change in near-surface deformation style and complexity progressing from the southwest to the northeast along the trace of the CGL. The seismic reflection data collaborate mapping evidence which suggests that this region has undergone a complex history of deformation, some of which is documented to be as young as Quaternary, during multiple episodes of reactivation under varying stress fields. This work, along with that of other studies presented in this volume, points to the existence of at least one major crustal feature outside the currently defined zone of seismic activity (New Madrid Seismic Zone) that should be considered as a significant potential source zone for seismogenic activity within the midcontinent region of the United States.

  13. Early Tertiary rupture of the Pacific plate: 1700 km of dextral offset along the Emperor trough-Line Islands lineament

    NASA Astrophysics Data System (ADS)

    Farrar, Edward; Dixon, John M.

    1981-05-01

    Between 67 and ˜40 Ma ago a northwest-southeast-trending fracture system over 8000 km long split the Pacific plate and accumulated at least 1700 km of dextral offset between the east and west portions. This system, here named the Emperor fracture zone (EFZ) system, consisted of several segments, one along the present trace of the Emperor trough and another along the Line Islands, joined by short spreading ridges. The EFZ terminated at its northern end against the Kula-Pacific ridge, and at its southern end in a ridge-transform system, called the Emperor spreading system, which extended to the west, north of Australia. The finite angular velocity vector describing the relative motion between the East and West Pacific plates is ˜0.6°/Ma about a pole at 36°N, 70°W. This vector, added to the known Early Tertiary motion of the Pacific plate with respect to the global hotspot reference frame, accounts in large part for the NNW trend of the Emperor seamount chain relative to the WNW Hawaiian trend, without violation of the integrity of the Antarctic plate. The Meiji-Emperor and Emperor-Hawaiian bends date, respectively, the initiation (˜67 Ma ago) and cessation (˜40 Ma ago) of seafloor spreading on the Emperor spreading system. The postulated Early Tertiary relative motion along the EFZ between the East and West Pacific plates explains (1) the present misalignment of the two sets of magnetic bights of the Pacific, (2) the abrupt truncation of eastern Pacific bathymetric lineaments against the Emperor trough and Line Islands, (3) the contrast in paleolatitude between the eastern and western Pacific as indicated by paleomagnetic and sedimentologic studies, and (4) the anomalous gravity signature of the central Hawaiian ridge that indicates that the ridge loaded thin hot lithosphere.

  14. Analysis of the influence of tectonics on the evolution valley network based on the SRTM DEM and the relationship of automatically extracted lineaments and the tectonic faults, Jemma River basin, Ethiopia

    NASA Astrophysics Data System (ADS)

    Kusák, Michal

    2016-04-01

    The Ethiopian Highland is good example of high plateau landscape formed by combination of tectonic uplift and episodic volcanism (Kazmin, 1975; Pik et al., 2003; Gani et al., 2009). Deeply incised gorges indicate active fluvial erosion which leads to instabilities of over-steepened slopes. In this study we focus on Jemma River basin which is a left tributary of Abay - Blue Nile to assess the influence of neotectonics on the evolution of its river and valley network. Tectonic lineaments, shape of valley networks, direction of river courses and intensity of fluvial erosion were compared in six subregions which were delineate beforehand by means of morphometric analysis. The influence of tectonics on the valley network is low in the older deep and wide canyons and in the and on the high plateau covered with Tertiary lava flows while younger upper part of the canyons it is high. Furthermore, the coincidence of the valley network with the tectonic lineaments differs in the subregions. The fluvial erosion along the main tectonic zones (NE-SW) direction made the way for backward erosion possible to reach far distant areas in E for the fluvial erosion. This tectonic zone also separates older areas in the W from the youngest landscape evolution subregions in the E, next to the Rift Valley. We studied the functions that can automatically extract lineaments in programs ArcGIS 10.1 and PCI Geomatica. The values of input parameters and their influence of the final shape and number of lineaments. A map of automated extracted lineaments was created and compared with 1) the tectonic faults by Geology Survey of Ethiopia (1996); and 2) the lineaments based on visual interpretation of by the author. The comparation of lineaments by automated visualization in GIS and visual interpretation of lineaments by the author proves that both sets of lineaments are in the same azimuth (NE-SW) - the same direction as the orientation of the rift. But it the mapping of lineaments by automated

  15. Parallel flow diffusion battery

    DOEpatents

    Yeh, Hsu-Chi; Cheng, Yung-Sung

    1984-08-07

    A parallel flow diffusion battery for determining the mass distribution of an aerosol has a plurality of diffusion cells mounted in parallel to an aerosol stream, each diffusion cell including a stack of mesh wire screens of different density.

  16. Parallel flow diffusion battery

    DOEpatents

    Yeh, H.C.; Cheng, Y.S.

    1984-01-01

    A parallel flow diffusion battery for determining the mass distribution of an aerosol has a plurality of diffusion cells mounted in parallel to an aerosol stream, each diffusion cell including a stack of mesh wire screens of different density.

  17. The West Beverly Hills Lineament and Beverly Hills High School: Ethical Issues in Geo-Hazard Communication

    NASA Astrophysics Data System (ADS)

    Gath, Eldon; Gonzalez, Tania; Roe, Joe; Buchiarelli, Philip; Kenny, Miles

    2014-05-01

    Results of geotechnical studies for the Westside Subway were disclosed in a public hearing on Oct. 19, 2011, showing new "active faults" of the Santa Monica fault and the West Beverly Hills Lineament (WBHL), identified as a northern extension of the Newport-Inglewood fault. Presentations made spoke of the danger posed by these faults, the possibility of killing people, and how it was good news that these faults had been discovered now instead of later. The presentations were live and are now memorialized as YouTube videos, (http://www.youtube.com/watch?v=Omx2BTIpzAk and others). No faults had been physically exposed or observed by the study; the faults were all interpreted from cone penetrometer probes, supplemented by core borings and geophysical transects. Several of the WBHL faults traversed buildings of the Beverly Hills High School (BHHS), triggering the school district to geologically map and characterize these faults for future planning efforts, and to quantify risk to the students in the 1920's high school building. 5 exploratory trenches were excavated within the high school property, 12 cone penetrometers were pushed, and 26-cored borings were drilled. Geologic logging of the trenches and borings and interpretation of the CPT data failed to confirm the presence of the mapped WBHL faults, instead showing an unfaulted, 3° NE dipping sequence of mid-Pleistocene alluvial fan deposits conformably overlying an ~1 Ma marine sand. Using 14C, OSL, and soil pedology for stratigraphic dating, the BHHS site was cleared from fault rupture hazards and the WBHL was shown to be an erosional margin of Benedict Canyon, partially buttressed by 40-200 ka alluvial deposits from Benedict Wash. The consequence of the Westside Subway's active fault maps has been the unexpected expenditure of millions of dollars for emergency fault investigations at BHHS and several other private properties within a densely developed urban highrise environment. None of these studies have found

  18. Parallel processing ITS

    SciTech Connect

    Fan, W.C.; Halbleib, J.A. Sr.

    1996-09-01

    This report provides a users` guide for parallel processing ITS on a UNIX workstation network, a shared-memory multiprocessor or a massively-parallel processor. The parallelized version of ITS is based on a master/slave model with message passing. Parallel issues such as random number generation, load balancing, and communication software are briefly discussed. Timing results for example problems are presented for demonstration purposes.

  19. Research in parallel computing

    NASA Technical Reports Server (NTRS)

    Ortega, James M.; Henderson, Charles

    1994-01-01

    This report summarizes work on parallel computations for NASA Grant NAG-1-1529 for the period 1 Jan. - 30 June 1994. Short summaries on highly parallel preconditioners, target-specific parallel reductions, and simulation of delta-cache protocols are provided.

  20. Parallel simulation today

    NASA Technical Reports Server (NTRS)

    Nicol, David; Fujimoto, Richard

    1992-01-01

    This paper surveys topics that presently define the state of the art in parallel simulation. Included in the tutorial are discussions on new protocols, mathematical performance analysis, time parallelism, hardware support for parallel simulation, load balancing algorithms, and dynamic memory management for optimistic synchronization.

  1. Parallel algorithm development

    SciTech Connect

    Adams, T.F.

    1996-06-01

    Rapid changes in parallel computing technology are causing significant changes in the strategies being used for parallel algorithm development. One approach is simply to write computer code in a standard language like FORTRAN 77 or with the expectation that the compiler will produce executable code that will run in parallel. The alternatives are: (1) to build explicit message passing directly into the source code; or (2) to write source code without explicit reference to message passing or parallelism, but use a general communications library to provide efficient parallel execution. Application of these strategies is illustrated with examples of codes currently under development.

  2. Parallel Atomistic Simulations

    SciTech Connect

    HEFFELFINGER,GRANT S.

    2000-01-18

    Algorithms developed to enable the use of atomistic molecular simulation methods with parallel computers are reviewed. Methods appropriate for bonded as well as non-bonded (and charged) interactions are included. While strategies for obtaining parallel molecular simulations have been developed for the full variety of atomistic simulation methods, molecular dynamics and Monte Carlo have received the most attention. Three main types of parallel molecular dynamics simulations have been developed, the replicated data decomposition, the spatial decomposition, and the force decomposition. For Monte Carlo simulations, parallel algorithms have been developed which can be divided into two categories, those which require a modified Markov chain and those which do not. Parallel algorithms developed for other simulation methods such as Gibbs ensemble Monte Carlo, grand canonical molecular dynamics, and Monte Carlo methods for protein structure determination are also reviewed and issues such as how to measure parallel efficiency, especially in the case of parallel Monte Carlo algorithms with modified Markov chains are discussed.

  3. Proceedings of the 38th Lunar and Planetary Science Conference

    NASA Technical Reports Server (NTRS)

    2007-01-01

    The sessions in the conference include: Titan, Mars Volcanism, Mars Polar Layered Deposits, Early Solar System Isotopes, SPECIAL SESSION: Mars Reconnaissance Orbiter: New Ways of Studying the Red Planet, Achondrites: Exploring Oxygen Isotopes and Parent-Body Processes, Solar System Formation and Evolution, SPECIAL SESSION: SMART-1, . Impact Cratering: Observations and Experiments, SPECIAL SESSION: Volcanism and Tectonism on Saturnian Satellites, Solar Nebula Composition, Mars Fluvial Geomorphology, Asteroid Observations: Spectra, Mostly, Mars Sediments and Geochemistry: View from the Surface, Mars Tectonics and Crustal Dichotomy, Stardust: Wild-2 Revealed, Impact Cratering from Observations and Interpretations, Mars Sediments and Geochemistry: The Map View, Chondrules and Their Formation, Enceladus, Asteroids and Deep Impact: Structure, Dynamics, and Experiments, Mars Surface Process and Evolution, Martian Meteorites: Nakhlites, Experiments, and the Great Shergottite Age Debate, Stardust: Mainly Mineralogy, Astrobiology, Wind-Surface Interactions on Mars and Earth, Icy Satellite Surfaces, Venus, Lunar Remote Sensing, Space Weathering, and Impact Effects, Interplanetary Dust/Genesis, Mars Cratering: Counts and Catastrophes?, Chondrites: Secondary Processes, Mars Sediments and Geochemistry: Atmosphere, Soils, Brines, and Minerals, Lunar Interior and Differentiation, Mars Magnetics and Atmosphere: Core to Ionosphere, Metal-rich Chondrites, Organics in Chondrites, Lunar Impacts and Meteorites, Presolar/Solar Grains, Topics for Print Only papers are: Outer Planets/Satellites, Early Solar System, Interplanetary Dust, Comets and Kuiper Belt Objects, Asteroids and Meteoroids, Chondrites, Achondrites, Meteorite Related, Mars Reconnaissance Orbiter, Mars, Astrobiology, Planetary Differentiation, Impacts, Mercury, Lunar Samples and Modeling, Venus, Missions and Instruments, Global Warming, Education and Public Outreach, Poster sessions are: Asteroids/Kuiper Belt Objects, Galilean Satellites: Geology and Mapping, Titan, Volcanism and Tectonism on Saturnian Satellites, Early Solar System, Achondrite Hodgepodge, Ordinary Chondrites, Carbonaceous Chondrites, Impact Cratering from Observations and Interpretations, Impact Cratering from Experiments and Modeling, SMART-1, Planetary Differentiation, Mars Geology, Mars Volcanism, Mars Tectonics, Mars: Polar, Glacial, and Near-Surface Ice, Mars Valley Networks, Mars Gullies, Mars Outflow Channels, Mars Sediments and Geochemistry: Spirit and Opportunity, Mars Reconnaissance Orbiter: New Ways of Studying the Red Planet, Mars Reconnaissance Orbiter: Geology, Layers, and Landforms, Oh, My!, Mars Reconnaissance Orbiter: Viewing Mars Through Multicolored Glasses; Mars Science Laboratory, Phoenix, and ExoMars: Science, Instruments, and Landing Sites; Planetary Analogs: Chemical and Mineral, Planetary Analogs: Physical, Planetary Analogs: Operations, Future Mission Concepts, Planetary Data, Imaging, and Cartography, Outer Solar System, Presolar/Solar Grains, Stardust Mission; Interplanetary Dust, Genesis, Asteroids and Comets: Models, Dynamics, and Experiments, Venus, Mercury, Laboratory Instruments, Methods, and Techniques to Support Planetary Exploration; Instruments, Techniques, and Enabling Techologies for Planetary Exploration; Lunar Missions and Instruments, Living and Working on the Moon, Meteoroid Impacts on the Moon, Lunar Remote Sensing, Lunar Samples and Experiments, Lunar Atmosphere, Moon: Soils, Poles, and Volatiles, Lunar Topography and Geophysics, Lunar Meteorites, Chondrites: Secondary Processes, Chondrites, Martian Meteorites, Mars Cratering, Mars Surface Processes and Evolution, Mars Sediments and Geochemistry: Regolith, Spectroscopy, and Imaging, Mars Sediments and Geochemistry: Analogs and Mineralogy, Mars: Magnetics and Atmosphere, Mars Aeolian Geomorphology, Mars Data Processing and Analyses, Astrobiology, Engaging Student Educators and the Public in Planetary Science,

  4. Parallel digital forensics infrastructure.

    SciTech Connect

    Liebrock, Lorie M.; Duggan, David Patrick

    2009-10-01

    This report documents the architecture and implementation of a Parallel Digital Forensics infrastructure. This infrastructure is necessary for supporting the design, implementation, and testing of new classes of parallel digital forensics tools. Digital Forensics has become extremely difficult with data sets of one terabyte and larger. The only way to overcome the processing time of these large sets is to identify and develop new parallel algorithms for performing the analysis. To support algorithm research, a flexible base infrastructure is required. A candidate architecture for this base infrastructure was designed, instantiated, and tested by this project, in collaboration with New Mexico Tech. Previous infrastructures were not designed and built specifically for the development and testing of parallel algorithms. With the size of forensics data sets only expected to increase significantly, this type of infrastructure support is necessary for continued research in parallel digital forensics. This report documents the implementation of the parallel digital forensics (PDF) infrastructure architecture and implementation.

  5. Introduction to Parallel Computing

    DTIC Science & Technology

    1992-05-01

    Topology C, Ada, C++, Data-parallel FORTRAN, 2D mesh of node boards, each node FORTRAN-90 (late 1992) board has 1 application processor Devopment Tools ...parallel machines become the wave of the present, tools are increasingly needed to assist programmers in creating parallel tasks and coordinating...their activities. Linda was designed to be such a tool . Linda was designed with three important goals in mind: to be portable, efficient, and easy to use

  6. Parallel Wolff Cluster Algorithms

    NASA Astrophysics Data System (ADS)

    Bae, S.; Ko, S. H.; Coddington, P. D.

    The Wolff single-cluster algorithm is the most efficient method known for Monte Carlo simulation of many spin models. Due to the irregular size, shape and position of the Wolff clusters, this method does not easily lend itself to efficient parallel implementation, so that simulations using this method have thus far been confined to workstations and vector machines. Here we present two parallel implementations of this algorithm, and show that one gives fairly good performance on a MIMD parallel computer.

  7. PCLIPS: Parallel CLIPS

    NASA Technical Reports Server (NTRS)

    Hall, Lawrence O.; Bennett, Bonnie H.; Tello, Ivan

    1994-01-01

    A parallel version of CLIPS 5.1 has been developed to run on Intel Hypercubes. The user interface is the same as that for CLIPS with some added commands to allow for parallel calls. A complete version of CLIPS runs on each node of the hypercube. The system has been instrumented to display the time spent in the match, recognize, and act cycles on each node. Only rule-level parallelism is supported. Parallel commands enable the assertion and retraction of facts to/from remote nodes working memory. Parallel CLIPS was used to implement a knowledge-based command, control, communications, and intelligence (C(sup 3)I) system to demonstrate the fusion of high-level, disparate sources. We discuss the nature of the information fusion problem, our approach, and implementation. Parallel CLIPS has also be used to run several benchmark parallel knowledge bases such as one to set up a cafeteria. Results show from running Parallel CLIPS with parallel knowledge base partitions indicate that significant speed increases, including superlinear in some cases, are possible.

  8. Application Portable Parallel Library

    NASA Technical Reports Server (NTRS)

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  9. Parallel Algorithms and Patterns

    SciTech Connect

    Robey, Robert W.

    2016-06-16

    This is a powerpoint presentation on parallel algorithms and patterns. A parallel algorithm is a well-defined, step-by-step computational procedure that emphasizes concurrency to solve a problem. Examples of problems include: Sorting, searching, optimization, matrix operations. A parallel pattern is a computational step in a sequence of independent, potentially concurrent operations that occurs in diverse scenarios with some frequency. Examples are: Reductions, prefix scans, ghost cell updates. We only touch on parallel patterns in this presentation. It really deserves its own detailed discussion which Gabe Rockefeller would like to develop.

  10. Candidate-penetrative-fracture mapping of the Grand Canyon area, Arizona, from spatial correlation of deep geophysical features and surficial lineaments

    USGS Publications Warehouse

    Gettings, Mark E.; Bultman, Mark W.

    2005-01-01

    Some aquifers of the southwestern Colorado Plateaus Province are deeply buried and overlain by several impermeable shale layers, and so recharge to the aquifer probably is mainly by seepage down penetrative-fracture systems. The purpose of this 2-year study, sponsored by the U.S. National Park Service, was to map candidate deep penetrative fractures over a 120,000-km2 area, using gravity and aeromagnetic-anomaly data together with surficial-fracture data. The study area was on the Colorado Plateau south of the Grand Canyon and west of Black Mesa; mapping was carried out at a scale of 1:250,000. The resulting database constitutes a spatially registered estimate of deep-fracture locations. Candidate penetrative fractures were located by spatial correlation of horizontal- gradient and analytic-signal maximums of gravity and magnetic anomalies with major surficial lineaments obtained from geologic, topographic, side-looking-airborne-radar, and satellite imagery. The maps define a subset of candidate penetrative fractures because of limitations in the data coverage and the analytical technique. In particular, the data and analytical technique used cannot predict whether the fractures are open or closed. Correlations were carried out by using image-processing software, such that every pixel on the resulting images was coded to uniquely identify which datasets are correlated. The technique correctly identified known and many new deep fracture systems. The resulting penetrative-fracture-distribution maps constitute an objectively obtained, repeatable dataset and a benchmark from which additional studies can begin. The maps also define in detail the tectonic fabrics of the southwestern Colorado Plateaus Province. Overlaying the correlated lineaments on the normalized-density-of-vegetation-index image reveals that many of these lineaments correlate with the boundaries of vegetation zones in drainages and canyons and so may be controlling near-surface water availability in

  11. Magnetic Data Interpretation for the Source-Edge Locations in Parts of the Tectonically Active Transition Zone of the Narmada-Son Lineament in Central India

    NASA Astrophysics Data System (ADS)

    Ghosh, G. K.

    2016-02-01

    The study has been carried out in the transition zone of the Narmada-Son lineament (NSL) which is seismically active with various geological complexities, upwarp movement of the mantle material into the crust through fault, fractures lamination and upwelling. NSL is one of the most prominent lineaments in central India after the Himalaya in the Indian geology. The area of investigation extends from longitude 80.25°E to 81.50°E and latitude 23.50°N to 24.37°N in the central part of the Indian continent. Different types of subsurface geological formations viz. alluvial, Gondwana, Deccan traps, Vindhyan, Mahakoshal, Granite and Gneisses groups exist in this area with varying geological ages. In this study area tectonic movement and crustal variation have been taken place during the past time and which might be reason for the variation of magnetic field. Magnetic anomaly suggests that the area has been highly disturbed which causes the Narmada-Son lineament trending in the ENE-WSW direction. Magnetic anomaly variation has been taken place due to the lithological variations subject to the changes in the geological contacts like thrusts and faults in this area. Shallow and deeper sources have been distinguished using frequency domain analysis by applying different filters. To enhance the magnetic data, various types of derivatives to identify the source-edge locations of the causative source bodies. The present study carried out the interpretation using total horizontal derivative, tilt angle derivative, horizontal tilt angle derivative and Cos (θ) derivative map to get source-edge locations. The results derived from various derivatives of magnetic data have been compared with the basement depth solutions calculated from 3D Euler deconvolution. It is suggested that total horizontal derivative, tilt angle derivative and Cos (θ) derivative are the most useful tools for identifying the multiple source edge locations of the causative bodies in this tectonically active

  12. The occurrence of a complete continental rift type of volcanic rocks suite along the Yerer-Tullu Wellel Volcano Tectonic Lineament, Central Ethiopia

    NASA Astrophysics Data System (ADS)

    Abebe Adhana, Tsegaye

    2014-11-01

    The Yerer-Tullu Wellel Volcano-tectonic Lineament (YTVL) is an E-W trending fault system or aborted rift that intercepts the Main Ethiopian Rift (MER) at Debre Zeyt (Bishoftu)/Yerer, in the eastern periphery of Addis Ababa. The structure is in correspondence with the westward extension of the southern margin of the Gulf of Aden rift. The YTVL extends for more than 500 km with a very clear northern fault margin, between Addis Ababa and Ambo known as the “Ambo Fault”. The southern margin is indicated by an E-W trending segmented lineaments at the latitude of about N 8°30‧, the Bedele-Metu being the most clear segment. In between these limits there are several evolved central volcanoes and cinder cones. The central volcanoes range in age from 12 to 7 Ma in the western most (Tullu Wellel) and gradually the upper limit get younger towards East to less than 1 Ma in the Wenchi and Debre Zeyt (Bishoftu) areas. These volcanic products cover the whole spectrum of a continental rift volcanic rocks suite: (1) in the eastern zone (Yerer-Bishoftu) the suite is silica over-saturated, ranging in composition from transitional basalt to peralkaline rhyolite, (2) moving westwards, between Wechacha and Wenchi, the rocks suite is silica saturated ranging in composition from alkali basalt to trachyte, (3) further West between Ijaji-Konchi and Nekemt the rocks suite is silica under-saturated ranging in composition from basanite to phonolite. Crossing the Dedessa lineament, the Tullu Wellel rocks appear to be silica saturated. Within a single suite fractional crystallization is the predominant evolutional process even in the silica over-saturated suite. The westwards progressive silica under-saturation and increase in alkalinity (except for the Tullu Wellel volcanic centers) is interpreted by the gradual deepening of an anomalous mantle where partial fusion took place. Therefore, as distance increases from the MER junction to the West, the amount of melt on the upper mantle was

  13. Geochemical survey of lower Pennsylvanian Corbin Sandstone outcrop belt in eastern Kentucky

    SciTech Connect

    Richers, D.M.

    1981-09-01

    Geochemical anomalies that may constitute further evidence for the existence of the east-west trending 38th Parallel lineament have been discovered in Wolfe, Powell, and Menifee Counties, Kentucky. Stream-water, stream-sediment, and outcrop samples collected along the northeast-southwest-trending Corbin Sandstone outcrop belt show anomalous concentrations of U, Th, Zn, Cu, and Ni only in parts of the belt that skirt the 38th Parallel lineament. Landsat studies also show that anomalies are closely associated with the intersections of the four major linear trends present in eastern Kentucky. This association, in part, suggests that the anomalies resulted from ascending fluids which utilize these lineaments as conduits. The presence of slightly uraniferous rock at Bell Branch in Menifee County (together with the presence of geochemical anomalies in the stream-water and sediment samples in Powell, Wolfe, and Menifee Counties) is encouraging even though commercial quantities of uranium or base metals have not been discovered. An anomalously uraniferous kimberlite pipe in Elliot County warrants additional study for this part of eastern Kentucky.

  14. Parallel Lisp simulator

    SciTech Connect

    Weening, J.S.

    1988-05-01

    CSIM is a simulator for parallel Lisp, based on a continuation passing interpreter. It models a shared-memory multiprocessor executing programs written in Common Lisp, extended with several primitives for creating and controlling processes. This paper describes the structure of the simulator, measures its performance, and gives an example of its use with a parallel Lisp program.

  15. Parallel and Distributed Computing.

    DTIC Science & Technology

    1986-12-12

    program was devoted to parallel and distributed computing . Support for this part of the program was obtained from the present Army contract and a...Umesh Vazirani. A workshop on parallel and distributed computing was held from May 19 to May 23, 1986 and drew 141 participants. Keywords: Mathematical programming; Protocols; Randomized algorithms. (Author)

  16. Massively parallel mathematical sieves

    SciTech Connect

    Montry, G.R.

    1989-01-01

    The Sieve of Eratosthenes is a well-known algorithm for finding all prime numbers in a given subset of integers. A parallel version of the Sieve is described that produces computational speedups over 800 on a hypercube with 1,024 processing elements for problems of fixed size. Computational speedups as high as 980 are achieved when the problem size per processor is fixed. The method of parallelization generalizes to other sieves and will be efficient on any ensemble architecture. We investigate two highly parallel sieves using scattered decomposition and compare their performance on a hypercube multiprocessor. A comparison of different parallelization techniques for the sieve illustrates the trade-offs necessary in the design and implementation of massively parallel algorithms for large ensemble computers.

  17. Totally parallel multilevel algorithms

    NASA Technical Reports Server (NTRS)

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  18. Parallel computing works

    SciTech Connect

    Not Available

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  19. Tectonic architecture through Landsat-7 ETM+/SRTM DEM-derived lineaments and relationship to the hydrogeologic setting in Siwa region, NW Egypt

    NASA Astrophysics Data System (ADS)

    Masoud, Alaa; Koike, Katsuaki

    2006-08-01

    Fracture zones on the Earth's surface are important elements in the understanding of plate motion forces, the dynamics of the subsurface fluid flow, and earthquake distributions. However, good exposures of these features are always lacking in arid regions, characterized by flat topography and where sand dunes extensively cover the terrain. During field surveys these conditions, in many cases, hinder the proper characterization of such features. Therefore, an approach that identifies the regional fractures as lineaments on remotely-sensed images or shaded digital terrain models, with its large scale synoptic coverage, could be promising. In the present work, a segment tracing algorithm (STA), for lineament detection from Landsat-7 Enhanced Thematic Mapper Plus (ETM+) imagery, and the data from the Shuttle Radar Topographic Mission (SRTM) 30 m digital elevation model (DEM), has been applied in the Siwa region, located in the northwest of the Western Desert of Egypt. The objectives are to analyze the spatial variation in orientation of the detected linear features and its relation to the hydrogeologic setting in the area and the underlying geology, and to evaluate the performance of the algorithm applied to the ETM+ and the DEM data. Detailed structural analysis and better understanding of the tectonic evolution of the area could provide useful tools for hydrologists for reliable groundwater management and development planning. The results obtained have been evaluated by the structural analysis of the area and field observations. Four major vertical fracture zones were detected corresponding to two conjugate sets of strike-slip faults that governed the surface, and subsurface environments of the lakes in the region, and these correlate well with the regional tectonics.

  20. The NAS parallel benchmarks

    NASA Technical Reports Server (NTRS)

    Bailey, David (Editor); Barton, John (Editor); Lasinski, Thomas (Editor); Simon, Horst (Editor)

    1993-01-01

    A new set of benchmarks was developed for the performance evaluation of highly parallel supercomputers. These benchmarks consist of a set of kernels, the 'Parallel Kernels,' and a simulated application benchmark. Together they mimic the computation and data movement characteristics of large scale computational fluid dynamics (CFD) applications. The principal distinguishing feature of these benchmarks is their 'pencil and paper' specification - all details of these benchmarks are specified only algorithmically. In this way many of the difficulties associated with conventional benchmarking approaches on highly parallel systems are avoided.

  1. Parallel programming with PCN

    SciTech Connect

    Foster, I.; Tuecke, S.

    1991-12-01

    PCN is a system for developing and executing parallel programs. It comprises a high-level programming language, tools for developing and debugging programs in this language, and interfaces to Fortran and C that allow the reuse of existing code in multilingual parallel programs. Programs developed using PCN are portable across many different workstations, networks, and parallel computers. This document provides all the information required to develop parallel programs with the PCN programming system. In includes both tutorial and reference material. It also presents the basic concepts that underly PCN, particularly where these are likely to be unfamiliar to the reader, and provides pointers to other documentation on the PCN language, programming techniques, and tools. PCN is in the public domain. The latest version of both the software and this manual can be obtained by anonymous FTP from Argonne National Laboratory in the directory pub/pcn at info.mcs.anl.gov (c.f. Appendix A).

  2. Parallel programming with PCN

    SciTech Connect

    Foster, I.; Tuecke, S.

    1991-09-01

    PCN is a system for developing and executing parallel programs. It comprises a high-level programming language, a set of tools for developing and debugging programs in this language, and interfaces to Fortran and C that allow the reuse of existing code in multilingual parallel programs. Programs developed using PCN are portable across many different workstations, networks, and parallel computers. This document provides all the information required to develop parallel programs with the PCN programming system. It includes both tutorial and reference material. It also presents the basic concepts that underlie PCN, particularly where these are likely to be unfamiliar to the reader, and provides pointers to other documentation on the PCN language, programming techniques, and tools. PCN is in the public domain. The latest version of both the software and this manual can be obtained by anonymous FTP from Argonne National Laboratory at info.mcs.anl.gov.

  3. The Parallel Axiom

    ERIC Educational Resources Information Center

    Rogers, Pat

    1972-01-01

    Criteria for a reasonable axiomatic system are discussed. A discussion of the historical attempts to prove the independence of Euclids parallel postulate introduces non-Euclidean geometries. Poincare's model for a non-Euclidean geometry is defined and analyzed. (LS)

  4. Scalable parallel communications

    NASA Technical Reports Server (NTRS)

    Maly, K.; Khanna, S.; Overstreet, C. M.; Mukkamala, R.; Zubair, M.; Sekhar, Y. S.; Foudriat, E. C.

    1992-01-01

    Coarse-grain parallelism in networking (that is, the use of multiple protocol processors running replicated software sending over several physical channels) can be used to provide gigabit communications for a single application. Since parallel network performance is highly dependent on real issues such as hardware properties (e.g., memory speeds and cache hit rates), operating system overhead (e.g., interrupt handling), and protocol performance (e.g., effect of timeouts), we have performed detailed simulations studies of both a bus-based multiprocessor workstation node (based on the Sun Galaxy MP multiprocessor) and a distributed-memory parallel computer node (based on the Touchstone DELTA) to evaluate the behavior of coarse-grain parallelism. Our results indicate: (1) coarse-grain parallelism can deliver multiple 100 Mbps with currently available hardware platforms and existing networking protocols (such as Transmission Control Protocol/Internet Protocol (TCP/IP) and parallel Fiber Distributed Data Interface (FDDI) rings); (2) scale-up is near linear in n, the number of protocol processors, and channels (for small n and up to a few hundred Mbps); and (3) since these results are based on existing hardware without specialized devices (except perhaps for some simple modifications of the FDDI boards), this is a low cost solution to providing multiple 100 Mbps on current machines. In addition, from both the performance analysis and the properties of these architectures, we conclude: (1) multiple processors providing identical services and the use of space division multiplexing for the physical channels can provide better reliability than monolithic approaches (it also provides graceful degradation and low-cost load balancing); (2) coarse-grain parallelism supports running several transport protocols in parallel to provide different types of service (for example, one TCP handles small messages for many users, other TCP's running in parallel provide high bandwidth

  5. Parallel image compression

    NASA Technical Reports Server (NTRS)

    Reif, John H.

    1987-01-01

    A parallel compression algorithm for the 16,384 processor MPP machine was developed. The serial version of the algorithm can be viewed as a combination of on-line dynamic lossless test compression techniques (which employ simple learning strategies) and vector quantization. These concepts are described. How these concepts are combined to form a new strategy for performing dynamic on-line lossy compression is discussed. Finally, the implementation of this algorithm in a massively parallel fashion on the MPP is discussed.

  6. Revisiting and parallelizing SHAKE

    NASA Astrophysics Data System (ADS)

    Weinbach, Yael; Elber, Ron

    2005-10-01

    An algorithm is presented for running SHAKE in parallel. SHAKE is a widely used approach to compute molecular dynamics trajectories with constraints. An essential step in SHAKE is the solution of a sparse linear problem of the type Ax = b, where x is a vector of unknowns. Conjugate gradient minimization (that can be done in parallel) replaces the widely used iteration process that is inherently serial. Numerical examples present good load balancing and are limited only by communication time.

  7. Parallel architectures for vision

    SciTech Connect

    Maresca, M. ); Lavin, M.A. ); Li, H. )

    1988-08-01

    Vision computing involves the execution of a large number of operations on large sets of structured data. Sequential computers cannot achieve the speed required by most of the current applications and therefore parallel architectural solutions have to be explored. In this paper the authors examine the options that drive the design of a vision oriented computer, starting with the analysis of the basic vision computation and communication requirements. They briefly review the classical taxonomy for parallel computers, based on the multiplicity of the instruction and data stream, and apply a recently proposed criterion, the degree of autonomy of each processor, to further classify fine-grain SIMD massively parallel computers. They identify three types of processor autonomy, namely operation autonomy, addressing autonomy, and connection autonomy. For each type they give the basic definitions and show some examples. They focus on the concept of connection autonomy, which they believe is a key point in the development of massively parallel architectures for vision. They show two examples of parallel computers featuring different types of connection autonomy - the Connection Machine and the Polymorphic-Torus - and compare their cost and benefit.

  8. Sublattice parallel replica dynamics

    NASA Astrophysics Data System (ADS)

    Martínez, Enrique; Uberuaga, Blas P.; Voter, Arthur F.

    2014-06-01

    Exascale computing presents a challenge for the scientific community as new algorithms must be developed to take full advantage of the new computing paradigm. Atomistic simulation methods that offer full fidelity to the underlying potential, i.e., molecular dynamics (MD) and parallel replica dynamics, fail to use the whole machine speedup, leaving a region in time and sample size space that is unattainable with current algorithms. In this paper, we present an extension of the parallel replica dynamics algorithm [A. F. Voter, Phys. Rev. B 57, R13985 (1998), 10.1103/PhysRevB.57.R13985] by combining it with the synchronous sublattice approach of Shim and Amar [Y. Shim and J. G. Amar, Phys. Rev. B 71, 125432 (2005), 10.1103/PhysRevB.71.125432], thereby exploiting event locality to improve the algorithm scalability. This algorithm is based on a domain decomposition in which events happen independently in different regions in the sample. We develop an analytical expression for the speedup given by this sublattice parallel replica dynamics algorithm and compare it with parallel MD and traditional parallel replica dynamics. We demonstrate how this algorithm, which introduces a slight additional approximation of event locality, enables the study of physical systems unreachable with traditional methodologies and promises to better utilize the resources of current high performance and future exascale computers.

  9. Parallel optical sampler

    DOEpatents

    Tauke-Pedretti, Anna; Skogen, Erik J; Vawter, Gregory A

    2014-05-20

    An optical sampler includes a first and second 1.times.n optical beam splitters splitting an input optical sampling signal and an optical analog input signal into n parallel channels, respectively, a plurality of optical delay elements providing n parallel delayed input optical sampling signals, n photodiodes converting the n parallel optical analog input signals into n respective electrical output signals, and n optical modulators modulating the input optical sampling signal or the optical analog input signal by the respective electrical output signals, and providing n successive optical samples of the optical analog input signal. A plurality of output photodiodes and eADCs convert the n successive optical samples to n successive digital samples. The optical modulator may be a photodiode interconnected Mach-Zehnder Modulator. A method of sampling the optical analog input signal is disclosed.

  10. CRUNCH_PARALLEL

    SciTech Connect

    Shumaker, Dana E.; Steefel, Carl I.

    2016-06-21

    The code CRUNCH_PARALLEL is a parallel version of the CRUNCH code. CRUNCH code version 2.0 was previously released by LLNL, (UCRL-CODE-200063). Crunch is a general purpose reactive transport code developed by Carl Steefel and Yabusake (Steefel Yabsaki 1996). The code handles non-isothermal transport and reaction in one, two, and three dimensions. The reaction algorithm is generic in form, handling an arbitrary number of aqueous and surface complexation as well as mineral dissolution/precipitation. A standardized database is used containing thermodynamic and kinetic data. The code includes advective, dispersive, and diffusive transport.

  11. The NAS Parallel Benchmarks

    SciTech Connect

    Bailey, David H.

    2009-11-15

    The NAS Parallel Benchmarks (NPB) are a suite of parallel computer performance benchmarks. They were originally developed at the NASA Ames Research Center in 1991 to assess high-end parallel supercomputers. Although they are no longer used as widely as they once were for comparing high-end system performance, they continue to be studied and analyzed a great deal in the high-performance computing community. The acronym 'NAS' originally stood for the Numerical Aeronautical Simulation Program at NASA Ames. The name of this organization was subsequently changed to the Numerical Aerospace Simulation Program, and more recently to the NASA Advanced Supercomputing Center, although the acronym remains 'NAS.' The developers of the original NPB suite were David H. Bailey, Eric Barszcz, John Barton, David Browning, Russell Carter, LeoDagum, Rod Fatoohi, Samuel Fineberg, Paul Frederickson, Thomas Lasinski, Rob Schreiber, Horst Simon, V. Venkatakrishnan and Sisira Weeratunga. The original NAS Parallel Benchmarks consisted of eight individual benchmark problems, each of which focused on some aspect of scientific computing. The principal focus was in computational aerophysics, although most of these benchmarks have much broader relevance, since in a much larger sense they are typical of many real-world scientific computing applications. The NPB suite grew out of the need for a more rational procedure to select new supercomputers for acquisition by NASA. The emergence of commercially available highly parallel computer systems in the late 1980s offered an attractive alternative to parallel vector supercomputers that had been the mainstay of high-end scientific computing. However, the introduction of highly parallel systems was accompanied by a regrettable level of hype, not only on the part of the commercial vendors but even, in some cases, by scientists using the systems. As a result, it was difficult to discern whether the new systems offered any fundamental performance advantage

  12. Highly parallel computation

    NASA Technical Reports Server (NTRS)

    Denning, Peter J.; Tichy, Walter F.

    1990-01-01

    Among the highly parallel computing architectures required for advanced scientific computation, those designated 'MIMD' and 'SIMD' have yielded the best results to date. The present development status evaluation of such architectures shown neither to have attained a decisive advantage in most near-homogeneous problems' treatment; in the cases of problems involving numerous dissimilar parts, however, such currently speculative architectures as 'neural networks' or 'data flow' machines may be entailed. Data flow computers are the most practical form of MIMD fine-grained parallel computers yet conceived; they automatically solve the problem of assigning virtual processors to the real processors in the machine.

  13. Adaptive parallel logic networks

    NASA Technical Reports Server (NTRS)

    Martinez, Tony R.; Vidal, Jacques J.

    1988-01-01

    Adaptive, self-organizing concurrent systems (ASOCS) that combine self-organization with massive parallelism for such applications as adaptive logic devices, robotics, process control, and system malfunction management, are presently discussed. In ASOCS, an adaptive network composed of many simple computing elements operating in combinational and asynchronous fashion is used and problems are specified by presenting if-then rules to the system in the form of Boolean conjunctions. During data processing, which is a different operational phase from adaptation, the network acts as a parallel hardware circuit.

  14. Dynamic topography of the western Great Plains: landscape evidence for mantle-driven uplift associated with the Jemez lineament of NE New Mexico and SE Colorado

    NASA Astrophysics Data System (ADS)

    Nereson, A. L.; Karlstrom, K. E.; McIntosh, W. C.; Heizler, M. T.; Kelley, S. A.; Brown, S. W.

    2011-12-01

    Dynamic topography results when viscous stresses created by flow within the mantle are transmitted through the lithosphere and interact with, and deform, the Earth's surface. Because dynamic topography is characterized by low amplitudes and long wavelengths, its subtle effects may be best recorded in low-relief areas such as the Great Plains of the USA where they can be readily observed and measured. We apply this concept to a unique region of the western Great Plains in New Mexico and Colorado where basalt flows of the Jemez lineament (Raton-Clayton and Ocate fields) form mesas (inverted topography) that record the evolution of the Great Plains surface through time. This study uses multiple datasets to evaluate the mechanisms which have driven the evolution of this landscape. Normalized channel steepness index (ksn) analysis identifies anomalously steep river gradients across broad (50-100 km) convexities within a NE- trending zone of differential river incision where higher downstream incision rates in the last 1.5 Ma suggest headwater uplift. At 2-8 Ma timescales, 40Ar/39Ar ages of basalt-capped paleosurfaces in the Raton-Clayton and Ocate volcanic fields indicate that rates of denudation increase systematically towards the NW from a NE-trending zone of approximately zero denudation (that approximately coincides with the high ksn zone), also suggestive of regional warping above the Jemez lineament. Onset of more rapid denudation is observed in the Raton-Clayton field beginning at ca. 3.6 Ma. Furthermore, two 300-400-m-high NE-trending erosional escarpments impart a staircase-like topographic profile to the region. Tomographic images from the EarthScope experiment show that NE-trending topographic features of this region correspond to an ~8 % P-wave velocity gradient of similar trend at the margin of the low-velocity Jemez mantle anomaly. We propose that the erosional landscapes of this unique area are, in large part, the surface expression of dynamic mantle

  15. Parallel Coordinate Axes.

    ERIC Educational Resources Information Center

    Friedlander, Alex; And Others

    1982-01-01

    Several methods of numerical mappings other than the usual cartesian coordinate system are considered. Some examples using parallel axes representation, which are seen to lead to aesthetically pleasing or interesting configurations, are presented. Exercises with alternative representations can stimulate pupil imagination and exploration in…

  16. Parallel programming with PCN

    SciTech Connect

    Foster, I.; Tuecke, S.

    1993-01-01

    PCN is a system for developing and executing parallel programs. It comprises a high-level programming language, tools for developing and debugging programs in this language, and interfaces to Fortran and Cthat allow the reuse of existing code in multilingual parallel programs. Programs developed using PCN are portable across many different workstations, networks, and parallel computers. This document provides all the information required to develop parallel programs with the PCN programming system. It includes both tutorial and reference material. It also presents the basic concepts that underlie PCN, particularly where these are likely to be unfamiliar to the reader, and provides pointers to other documentation on the PCN language, programming techniques, and tools. PCN is in the public domain. The latest version of both the software and this manual can be obtained by anonymous ftp from Argonne National Laboratory in the directory pub/pcn at info.mcs. ani.gov (cf. Appendix A). This version of this document describes PCN version 2.0, a major revision of the PCN programming system. It supersedes earlier versions of this report.

  17. Parallel Dislocation Simulator

    SciTech Connect

    2006-10-30

    ParaDiS is software capable of simulating the motion, evolution, and interaction of dislocation networks in single crystals using massively parallel computer architectures. The software is capable of outputting the stress-strain response of a single crystal whose plastic deformation is controlled by the dislocation processes.

  18. Massively parallel processor computer

    NASA Technical Reports Server (NTRS)

    Fung, L. W. (Inventor)

    1983-01-01

    An apparatus for processing multidimensional data with strong spatial characteristics, such as raw image data, characterized by a large number of parallel data streams in an ordered array is described. It comprises a large number (e.g., 16,384 in a 128 x 128 array) of parallel processing elements operating simultaneously and independently on single bit slices of a corresponding array of incoming data streams under control of a single set of instructions. Each of the processing elements comprises a bidirectional data bus in communication with a register for storing single bit slices together with a random access memory unit and associated circuitry, including a binary counter/shift register device, for performing logical and arithmetical computations on the bit slices, and an I/O unit for interfacing the bidirectional data bus with the data stream source. The massively parallel processor architecture enables very high speed processing of large amounts of ordered parallel data, including spatial translation by shifting or sliding of bits vertically or horizontally to neighboring processing elements.

  19. High performance parallel architectures

    SciTech Connect

    Anderson, R.E. )

    1989-09-01

    In this paper the author describes current high performance parallel computer architectures. A taxonomy is presented to show computer architecture from the user programmer's point-of-view. The effects of the taxonomy upon the programming model are described. Some current architectures are described with respect to the taxonomy. Finally, some predictions about future systems are presented. 5 refs., 1 fig.

  20. Parallel fast gauss transform

    SciTech Connect

    Sampath, Rahul S; Sundar, Hari; Veerapaneni, Shravan

    2010-01-01

    We present fast adaptive parallel algorithms to compute the sum of N Gaussians at N points. Direct sequential computation of this sum would take O(N{sup 2}) time. The parallel time complexity estimates for our algorithms are O(N/n{sub p}) for uniform point distributions and O( (N/n{sub p}) log (N/n{sub p}) + n{sub p}log n{sub p}) for non-uniform distributions using n{sub p} CPUs. We incorporate a plane-wave representation of the Gaussian kernel which permits 'diagonal translation'. We use parallel octrees and a new scheme for translating the plane-waves to efficiently handle non-uniform distributions. Computing the transform to six-digit accuracy at 120 billion points took approximately 140 seconds using 4096 cores on the Jaguar supercomputer. Our implementation is 'kernel-independent' and can handle other 'Gaussian-type' kernels even when explicit analytic expression for the kernel is not known. These algorithms form a new class of core computational machinery for solving parabolic PDEs on massively parallel architectures.

  1. Parallel hierarchical radiosity rendering

    SciTech Connect

    Carter, M.

    1993-07-01

    In this dissertation, the step-by-step development of a scalable parallel hierarchical radiosity renderer is documented. First, a new look is taken at the traditional radiosity equation, and a new form is presented in which the matrix of linear system coefficients is transformed into a symmetric matrix, thereby simplifying the problem and enabling a new solution technique to be applied. Next, the state-of-the-art hierarchical radiosity methods are examined for their suitability to parallel implementation, and scalability. Significant enhancements are also discovered which both improve their theoretical foundations and improve the images they generate. The resultant hierarchical radiosity algorithm is then examined for sources of parallelism, and for an architectural mapping. Several architectural mappings are discussed. A few key algorithmic changes are suggested during the process of making the algorithm parallel. Next, the performance, efficiency, and scalability of the algorithm are analyzed. The dissertation closes with a discussion of several ideas which have the potential to further enhance the hierarchical radiosity method, or provide an entirely new forum for the application of hierarchical methods.

  2. Parallel hierarchical global illumination

    SciTech Connect

    Snell, Quinn O.

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  3. Parallel Multigrid Equation Solver

    SciTech Connect

    Adams, Mark

    2001-09-07

    Prometheus is a fully parallel multigrid equation solver for matrices that arise in unstructured grid finite element applications. It includes a geometric and an algebraic multigrid method and has solved problems of up to 76 mullion degrees of feedom, problems in linear elasticity on the ASCI blue pacific and ASCI red machines.

  4. Photogeologic and kinematic analysis of lineaments at Yucca Mountain, Nevada: Implications for strike-slip faulting and oroclinal bending

    SciTech Connect

    O`Neill, J.M.; Whitney, J.W.; Hudson, M.R.

    1992-12-31

    The main structural grain at Yucca Mountain, as seen from aerial photographs, is a pronounced north-trending linear fabric defined by parallel east-tilted fault-block ridges. The ridges are bounded on the west by normal faults that are easily recognizable on aerial photographs, mainly as isolated, colinear scarps in alluvium and as offset bedrock units. AH ridge-bounding to adjacent faults, most commonly by short northwest-trending fault splays. The generally north-trending high-angle faults primarily display down-to-the-west normal offset, but also have an auxiliary component of left-lateral slip. Left-lateral slip is indicated by offset stream channels, slickenlines, and en echelon fault splays that are structurally linked, commonly by pull-apart grabens. These grabens, best seen on low-sun angle aerial photographs, rangefrom tens of meters to more than 3 kilometers wide. The smallest pull-apart zones are well developed along the Windy Wash and Solitario Canyon faults on the west side of Yucca Mountain; the largest of these features is interpreted to structurally link the Bow Ridge and Solitario Canyon faults in the north-central part of Yucca Mountain; the pronounced northwest-trending drainage system in this part of Yucca Mountain appears to be controlled by tension fractures related to left-lateral strike-slip movement on these north-trending faults. Midway Valley, directly east of this pull-apart graben, may also owe its origin, in part, to a pull-apart mechanism.

  5. Neoarchean and Paleoproterozoic granitoids marginal to the Jeceaba-Bom Sucesso lineament (SE border of the southern São Francisco craton): Genesis and tectonic evolution

    NASA Astrophysics Data System (ADS)

    Campos, José Carlos Sales; Carneiro, Maurício Antônio

    2008-12-01

    The sialic crust of the southern São Francisco craton along the Jeceaba-Bom Sucesso lineament, central-southern part of Minas Gerais (Brazil), encompasses, among other rock types, Neoarchean and Paleoproterozoic granitoids. These granitoids, according to their petrographic, lithogeochemical and geochronologic characteristics, were grouped into two Neoarchean suites (Samambaia-Bom Sucesso and Salto Paraopeba-Babilônia) and three Paleoproterozoic suites (Cassiterita-Tabuões, Ritápolis and São Tiago). Varied processes and tectonic environments were involved in the genesis of these suites. In particular, the lithogeochemistry of the (Archean and Paleoproterozoic) TTG-type granitoids indicates an origin by partial melting of hydrated basaltic crust in a subduction environment. In the Neoarchean, between 2780 and 2703 Ma, a dominant TTG granitoid genesis related to an active continental margin was followed by another granite genesis related to crustal anatexis processes at 2612-2550 Ma. In the Paleoproterozoic, the generation of TTG and granites s.s. occurred at three distinct times: 2162, 2127 and 1887 Ma. This fact, plus the rock-type diversity produced by this granite genesis, indicates that the continental margin of the southern portion of the São Francisco craton was affected by more than one consumption episode of oceanic crust, involving different island arc segments, and the late Neoarchean consolidate continent. A Paleoproterozoic tectonic evolution in three stages is proposed in this work.

  6. Identification of a new fault and associated lineament features in Oregon's Summer Lake Valley using high-resolution LiDAR data

    NASA Astrophysics Data System (ADS)

    Bennett, L.; Madin, I.

    2012-12-01

    In 2012, the Oregon Department of Geology and Mineral Industries (DOGAMI) contracted WSI to co-acquire airborne Light Detecting and Ranging (LiDAR) and Thermal Infrared Imagery (TIR) data within the region surrounding Summer Lake, Oregon. The objective of this project was to detect surficial expressions of geothermal activity and associated geologic features. An analysis of the LiDAR data revealed one newly identified fault and several accompanying lineaments that strike northwest, similar to the trend of the Ana River, Brothers, and Eugene-Denio Fault Zones in Central Oregon. The age of the Ana River Fault Zone and Summer Lake bed is known to be within the Holocene epoch. Apparent scarp height observed from the LiDAR is up to 8 meters. While detailed analysis is ongoing, the data illustrated the effectiveness of using high resolution remote sensing data for surficial analysis of geologic displacement. This presentation will focus on direct visual detection of features in the Summer Lake, Oregon landscape using LiDAR data.

  7. Parallel grid population

    DOEpatents

    Wald, Ingo; Ize, Santiago

    2015-07-28

    Parallel population of a grid with a plurality of objects using a plurality of processors. One example embodiment is a method for parallel population of a grid with a plurality of objects using a plurality of processors. The method includes a first act of dividing a grid into n distinct grid portions, where n is the number of processors available for populating the grid. The method also includes acts of dividing a plurality of objects into n distinct sets of objects, assigning a distinct set of objects to each processor such that each processor determines by which distinct grid portion(s) each object in its distinct set of objects is at least partially bounded, and assigning a distinct grid portion to each processor such that each processor populates its distinct grid portion with any objects that were previously determined to be at least partially bounded by its distinct grid portion.

  8. Parallel Anisotropic Tetrahedral Adaptation

    NASA Technical Reports Server (NTRS)

    Park, Michael A.; Darmofal, David L.

    2008-01-01

    An adaptive method that robustly produces high aspect ratio tetrahedra to a general 3D metric specification without introducing hybrid semi-structured regions is presented. The elemental operators and higher-level logic is described with their respective domain-decomposed parallelizations. An anisotropic tetrahedral grid adaptation scheme is demonstrated for 1000-1 stretching for a simple cube geometry. This form of adaptation is applicable to more complex domain boundaries via a cut-cell approach as demonstrated by a parallel 3D supersonic simulation of a complex fighter aircraft. To avoid the assumptions and approximations required to form a metric to specify adaptation, an approach is introduced that directly evaluates interpolation error. The grid is adapted to reduce and equidistribute this interpolation error calculation without the use of an intervening anisotropic metric. Direct interpolation error adaptation is illustrated for 1D and 3D domains.

  9. Parallel Subconvolution Filtering Architectures

    NASA Technical Reports Server (NTRS)

    Gray, Andrew A.

    2003-01-01

    These architectures are based on methods of vector processing and the discrete-Fourier-transform/inverse-discrete- Fourier-transform (DFT-IDFT) overlap-and-save method, combined with time-block separation of digital filters into frequency-domain subfilters implemented by use of sub-convolutions. The parallel-processing method implemented in these architectures enables the use of relatively small DFT-IDFT pairs, while filter tap lengths are theoretically unlimited. The size of a DFT-IDFT pair is determined by the desired reduction in processing rate, rather than on the order of the filter that one seeks to implement. The emphasis in this report is on those aspects of the underlying theory and design rules that promote computational efficiency, parallel processing at reduced data rates, and simplification of the designs of very-large-scale integrated (VLSI) circuits needed to implement high-order filters and correlators.

  10. Parallel multilevel preconditioners

    SciTech Connect

    Bramble, J.H.; Pasciak, J.E.; Xu, Jinchao.

    1989-01-01

    In this paper, we shall report on some techniques for the development of preconditioners for the discrete systems which arise in the approximation of solutions to elliptic boundary value problems. Here we shall only state the resulting theorems. It has been demonstrated that preconditioned iteration techniques often lead to the most computationally effective algorithms for the solution of the large algebraic systems corresponding to boundary value problems in two and three dimensional Euclidean space. The use of preconditioned iteration will become even more important on computers with parallel architecture. This paper discusses an approach for developing completely parallel multilevel preconditioners. In order to illustrate the resulting algorithms, we shall describe the simplest application of the technique to a model elliptic problem.

  11. Ultrascalable petaflop parallel supercomputer

    DOEpatents

    Blumrich, Matthias A.; Chen, Dong; Chiu, George; Cipolla, Thomas M.; Coteus, Paul W.; Gara, Alan G.; Giampapa, Mark E.; Hall, Shawn; Haring, Rudolf A.; Heidelberger, Philip; Kopcsay, Gerard V.; Ohmacht, Martin; Salapura, Valentina; Sugavanam, Krishnan; Takken, Todd

    2010-07-20

    A massively parallel supercomputer of petaOPS-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC) having up to four processing elements. The ASIC nodes are interconnected by multiple independent networks that optimally maximize the throughput of packet communications between nodes with minimal latency. The multiple networks may include three high-speed networks for parallel algorithm message passing including a Torus, collective network, and a Global Asynchronous network that provides global barrier and notification functions. These multiple independent networks may be collaboratively or independently utilized according to the needs or phases of an algorithm for optimizing algorithm processing performance. The use of a DMA engine is provided to facilitate message passing among the nodes without the expenditure of processing resources at the node.

  12. Homology, convergence and parallelism

    PubMed Central

    Ghiselin, Michael T.

    2016-01-01

    Homology is a relation of correspondence between parts of parts of larger wholes. It is used when tracking objects of interest through space and time and in the context of explanatory historical narratives. Homologues can be traced through a genealogical nexus back to a common ancestral precursor. Homology being a transitive relation, homologues remain homologous however much they may come to differ. Analogy is a relationship of correspondence between parts of members of classes having no relationship of common ancestry. Although homology is often treated as an alternative to convergence, the latter is not a kind of correspondence: rather, it is one of a class of processes that also includes divergence and parallelism. These often give rise to misleading appearances (homoplasies). Parallelism can be particularly hard to detect, especially when not accompanied by divergences in some parts of the body. PMID:26598721

  13. Parallel unstructured grid generation

    NASA Technical Reports Server (NTRS)

    Loehner, Rainald; Camberos, Jose; Merriam, Marshal

    1991-01-01

    A parallel unstructured grid generation algorithm is presented and implemented on the Hypercube. Different processor hierarchies are discussed, and the appropraite hierarchies for mesh generation and mesh smoothing are selected. A domain-splitting algorithm for unstructured grids which tries to minimize the surface-to-volume ratio of each subdomain is described. This splitting algorithm is employed both for grid generation and grid smoothing. Results obtained on the Hypercube demonstrate the effectiveness of the algorithms developed.

  14. Development of Parallel GSSHA

    DTIC Science & Technology

    2013-09-01

    C en te r Paul R. Eller , Jing-Ru C. Cheng, Aaron R. Byrd, Charles W. Downer, and Nawa Pradhan September 2013 Approved for public release...Program ERDC TR-13-8 September 2013 Development of Parallel GSSHA Paul R. Eller and Jing-Ru C. Cheng Information Technology Laboratory US Army Engineer...5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) Paul Eller , Ruth Cheng, Aaron Byrd, Chuck Downer, and Nawa Pradhan 5d. PROJECT NUMBER

  15. Xyce parallel electronic simulator.

    SciTech Connect

    Keiter, Eric R; Mei, Ting; Russo, Thomas V.; Rankin, Eric Lamont; Schiek, Richard Louis; Thornquist, Heidi K.; Fixel, Deborah A.; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide.

  16. Massively Parallel Genetics.

    PubMed

    Shendure, Jay; Fields, Stanley

    2016-06-01

    Human genetics has historically depended on the identification of individuals whose natural genetic variation underlies an observable trait or disease risk. Here we argue that new technologies now augment this historical approach by allowing the use of massively parallel assays in model systems to measure the functional effects of genetic variation in many human genes. These studies will help establish the disease risk of both observed and potential genetic variants and to overcome the problem of "variants of uncertain significance."

  17. Parallel sphere rendering

    SciTech Connect

    Krogh, M.; Painter, J.; Hansen, C.

    1996-10-01

    Sphere rendering is an important method for visualizing molecular dynamics data. This paper presents a parallel algorithm that is almost 90 times faster than current graphics workstations. To render extremely large data sets and large images, the algorithm uses the MIMD features of the supercomputers to divide up the data, render independent partial images, and then finally composite the multiple partial images using an optimal method. The algorithm and performance results are presented for the CM-5 and the M.

  18. Implementation of Parallel Algorithms

    DTIC Science & Technology

    1993-06-30

    their socia ’ relations or to achieve some goals. For example, we define a pair-wise force law of i epulsion and attraction for a group of identical...quantization based compression schemes. Photo-refractive crystals, which provide high density recording in real time, are used as our holographic media . The...of Parallel Algorithms (J. Reif, ed.). Kluwer Academic Pu’ ishers, 1993. (4) "A Dynamic Separator Algorithm", D. Armon and J. Reif. To appear in

  19. Trajectory optimization using parallel shooting method on parallel computer

    SciTech Connect

    Wirthman, D.J.; Park, S.Y.; Vadali, S.R.

    1995-03-01

    The efficiency of a parallel shooting method on a parallel computer for solving a variety of optimal control guidance problems is studied. Several examples are considered to demonstrate that a speedup of nearly 7 to 1 is achieved with the use of 16 processors. It is suggested that further improvements in performance can be achieved by parallelizing in the state domain. 10 refs.

  20. The Galley Parallel File System

    NASA Technical Reports Server (NTRS)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    As the I/O needs of parallel scientific applications increase, file systems for multiprocessors are being designed to provide applications with parallel access to multiple disks. Many parallel file systems present applications with a conventional Unix-like interface that allows the application to access multiple disks transparently. The interface conceals the parallelism within the file system, which increases the ease of programmability, but makes it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. Furthermore, most current parallel file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic parallel workloads. We discuss Galley's file structure and application interface, as well as an application that has been implemented using that interface.

  1. Asynchronous interpretation of parallel microprograms

    SciTech Connect

    Bandman, O.L.

    1984-03-01

    In this article, the authors demonstrate how to pass from a given synchronous interpretation of a parallel microprogram to an equivalent asynchronous interpretation, and investigate the cost associated with the rejection of external synchronization in parallel microprogram structures.

  2. Status of TRANSP Parallel Services

    NASA Astrophysics Data System (ADS)

    Indireshkumar, K.; Andre, Robert; McCune, Douglas; Randerson, Lewis

    2006-10-01

    The PPPL TRANSP code suite has been used successfully over many years to carry out time dependent simulations of tokamak plasmas. However, accurately modeling certain phenomena such as RF heating and fast ion behavior using TRANSP requires extensive computational power and will benefit from parallelization. Parallelizing all of TRANSP is not required and parts will run sequentially while other parts run parallelized. To efficiently use a site's parallel services, the parallelized TRANSP modules are deployed to a shared ``parallel service'' on a separate cluster. The PPPL Monte Carlo fast ion module NUBEAM and the MIT RF module TORIC are the first TRANSP modules to be so deployed. This poster will show the performance scaling of these modules within the parallel server. Communications between the serial client and the parallel server will be described in detail, and measurements of startup and communications overhead will be shown. Physics modeling benefits for TRANSP users will be assessed.

  3. Resistor Combinations for Parallel Circuits.

    ERIC Educational Resources Information Center

    McTernan, James P.

    1978-01-01

    To help simplify both teaching and learning of parallel circuits, a high school electricity/electronics teacher presents and illustrates the use of tables of values for parallel resistive circuits in which total resistances are whole numbers. (MF)

  4. Parallel Debugging Using Graphical Views

    DTIC Science & Technology

    1988-03-01

    Voyeur , a prototype system for creating graphical views of parallel programs, provid(s a cost-effective way to construct such views for any parallel...programming system. We illustrate Voyeur by discussing four views created for debugging Poker programs. One is a vteneral trace facility for any Poker...Graphical views are essential for debugging parallel programs because of the large quan- tity of state information contained in parallel programs. Voyeur

  5. Parallel Pascal - An extended Pascal for parallel computers

    NASA Technical Reports Server (NTRS)

    Reeves, A. P.

    1984-01-01

    Parallel Pascal is an extended version of the conventional serial Pascal programming language which includes a convenient syntax for specifying array operations. It is upward compatible with standard Pascal and involves only a small number of carefully chosen new features. Parallel Pascal was developed to reduce the semantic gap between standard Pascal and a large range of highly parallel computers. Two important design goals of Parallel Pascal were efficiency and portability. Portability is particularly difficult to achieve since different parallel computers frequently have very different capabilities.

  6. Pleistocene terrace deposition related to tectonically controlled surface uplift: An example of the Kyrenia Range lineament in the northern part of Cyprus

    NASA Astrophysics Data System (ADS)

    Palamakumbura, Romesh N.; Robertson, Alastair H. F.

    2016-06-01

    In this study, we consider how surface uplift of a narrow mountain range has interacted with glacial-related sea-level cyclicity and climatic change to produce a series of marine and non-marine terrace systems. The terrace deposits of the Kyrenia Range record rapid surface uplift of a long-lived tectonic lineament during the early Pleistocene, followed by continued surface uplift at a reduced rate during mid-late Pleistocene. Six terrace depositional systems are distinguished and correlated along the northern and southern flanks of the range, termed K0 to K5. The oldest and highest (K0 terrace system) is present only within the central part of the range. The K2-K5 terrace systems formed later, at sequentially lower levels away from the range. The earliest stage of surface uplift (K0 terrace system) comprises lacustrine carbonates interbedded with mass-flow facies (early Pleistocene?). The subsequent terrace system (K1) is made up of colluvial conglomerate and aeolian dune facies on both flanks of the range. The later terrace systems (K2 to K5) each begin with a basal marine deposit, interpreted as a marine transgression. Deltaic conglomerates prograded during inferred global interglacial stages. Overlying aeolian dune facies represent marine regressions, probably related to global glacial stages. Each terrace depositional system was uplifted and preserved, followed by subsequent deposits at progressively lower topographic levels. Climatic variation during interglacial-glacial cycles and autocyclic processes also exerted an influence on deposition, particularly on short-period fluvial and aeolian deposition.

  7. CSM parallel structural methods research

    NASA Technical Reports Server (NTRS)

    Storaasli, Olaf O.

    1989-01-01

    Parallel structural methods, research team activities, advanced architecture computers for parallel computational structural mechanics (CSM) research, the FLEX/32 multicomputer, a parallel structural analyses testbed, blade-stiffened aluminum panel with a circular cutout and the dynamic characteristics of a 60 meter, 54-bay, 3-longeron deployable truss beam are among the topics discussed.

  8. Roo: A parallel theorem prover

    SciTech Connect

    Lusk, E.L.; McCune, W.W.; Slaney, J.K.

    1991-11-01

    We describe a parallel theorem prover based on the Argonne theorem-proving system OTTER. The parallel system, called Roo, runs on shared-memory multiprocessors such as the Sequent Symmetry. We explain the parallel algorithm used and give performance results that demonstrate near-linear speedups on large problems.

  9. Parallel Eclipse Project Checkout

    NASA Technical Reports Server (NTRS)

    Crockett, Thomas M.; Joswig, Joseph C.; Shams, Khawaja S.; Powell, Mark W.; Bachmann, Andrew G.

    2011-01-01

    Parallel Eclipse Project Checkout (PEPC) is a program written to leverage parallelism and to automate the checkout process of plug-ins created in Eclipse RCP (Rich Client Platform). Eclipse plug-ins can be aggregated in a feature project. This innovation digests a feature description (xml file) and automatically checks out all of the plug-ins listed in the feature. This resolves the issue of manually checking out each plug-in required to work on the project. To minimize the amount of time necessary to checkout the plug-ins, this program makes the plug-in checkouts parallel. After parsing the feature, a request to checkout for each plug-in in the feature has been inserted. These requests are handled by a thread pool with a configurable number of threads. By checking out the plug-ins in parallel, the checkout process is streamlined before getting started on the project. For instance, projects that took 30 minutes to checkout now take less than 5 minutes. The effect is especially clear on a Mac, which has a network monitor displaying the bandwidth use. When running the client from a developer s home, the checkout process now saturates the bandwidth in order to get all the plug-ins checked out as fast as possible. For comparison, a checkout process that ranged from 8-200 Kbps from a developer s home is now able to saturate a pipe of 1.3 Mbps, resulting in significantly faster checkouts. Eclipse IDE (integrated development environment) tries to build a project as soon as it is downloaded. As part of another optimization, this innovation programmatically tells Eclipse to stop building while checkouts are happening, which dramatically reduces lock contention and enables plug-ins to continue downloading until all of them finish. Furthermore, the software re-enables automatic building, and forces Eclipse to do a clean build once it finishes checking out all of the plug-ins. This software is fully generic and does not contain any NASA-specific code. It can be applied to any

  10. Parallel sphere rendering

    SciTech Connect

    Krogh, M.; Hansen, C.; Painter, J.; de Verdiere, G.C.

    1995-05-01

    Sphere rendering is an important method for visualizing molecular dynamics data. This paper presents a parallel divide-and-conquer algorithm that is almost 90 times faster than current graphics workstations. To render extremely large data sets and large images, the algorithm uses the MIMD features of the supercomputers to divide up the data, render independent partial images, and then finally composite the multiple partial images using an optimal method. The algorithm and performance results are presented for the CM-5 and the T3D.

  11. Parallelized direct execution simulation of message-passing parallel programs

    NASA Technical Reports Server (NTRS)

    Dickens, Phillip M.; Heidelberger, Philip; Nicol, David M.

    1994-01-01

    As massively parallel computers proliferate, there is growing interest in findings ways by which performance of massively parallel codes can be efficiently predicted. This problem arises in diverse contexts such as parallelizing computers, parallel performance monitoring, and parallel algorithm development. In this paper we describe one solution where one directly executes the application code, but uses a discrete-event simulator to model details of the presumed parallel machine such as operating system and communication network behavior. Because this approach is computationally expensive, we are interested in its own parallelization specifically the parallelization of the discrete-event simulator. We describe methods suitable for parallelized direct execution simulation of message-passing parallel programs, and report on the performance of such a system, Large Application Parallel Simulation Environment (LAPSE), we have built on the Intel Paragon. On all codes measured to date, LAPSE predicts performance well typically within 10 percent relative error. Depending on the nature of the application code, we have observed low slowdowns (relative to natively executing code) and high relative speedups using up to 64 processors.

  12. Parallelizing quantum circuit synthesis

    NASA Astrophysics Data System (ADS)

    Di Matteo, Olivia; Mosca, Michele

    2016-03-01

    Quantum circuit synthesis is the process in which an arbitrary unitary operation is decomposed into a sequence of gates from a universal set, typically one which a quantum computer can implement both efficiently and fault-tolerantly. As physical implementations of quantum computers improve, the need is growing for tools that can effectively synthesize components of the circuits and algorithms they will run. Existing algorithms for exact, multi-qubit circuit synthesis scale exponentially in the number of qubits and circuit depth, leaving synthesis intractable for circuits on more than a handful of qubits. Even modest improvements in circuit synthesis procedures may lead to significant advances, pushing forward the boundaries of not only the size of solvable circuit synthesis problems, but also in what can be realized physically as a result of having more efficient circuits. We present a method for quantum circuit synthesis using deterministic walks. Also termed pseudorandom walks, these are walks in which once a starting point is chosen, its path is completely determined. We apply our method to construct a parallel framework for circuit synthesis, and implement one such version performing optimal T-count synthesis over the Clifford+T gate set. We use our software to present examples where parallelization offers a significant speedup on the runtime, as well as directly confirm that the 4-qubit 1-bit full adder has optimal T-count 7 and T-depth 3.

  13. Parallel ptychographic reconstruction

    PubMed Central

    Nashed, Youssef S. G.; Vine, David J.; Peterka, Tom; Deng, Junjing; Ross, Rob; Jacobsen, Chris

    2014-01-01

    Ptychography is an imaging method whereby a coherent beam is scanned across an object, and an image is obtained by iterative phasing of the set of diffraction patterns. It is able to be used to image extended objects at a resolution limited by scattering strength of the object and detector geometry, rather than at an optics-imposed limit. As technical advances allow larger fields to be imaged, computational challenges arise for reconstructing the correspondingly larger data volumes, yet at the same time there is also a need to deliver reconstructed images immediately so that one can evaluate the next steps to take in an experiment. Here we present a parallel method for real-time ptychographic phase retrieval. It uses a hybrid parallel strategy to divide the computation between multiple graphics processing units (GPUs) and then employs novel techniques to merge sub-datasets into a single complex phase and amplitude image. Results are shown on a simulated specimen and a real dataset from an X-ray experiment conducted at a synchrotron light source. PMID:25607174

  14. Tolerant (parallel) Programming

    NASA Technical Reports Server (NTRS)

    DiNucci, David C.; Bailey, David H. (Technical Monitor)

    1997-01-01

    In order to be truly portable, a program must be tolerant of a wide range of development and execution environments, and a parallel program is just one which must be tolerant of a very wide range. This paper first defines the term "tolerant programming", then describes many layers of tools to accomplish it. The primary focus is on F-Nets, a formal model for expressing computation as a folded partial-ordering of operations, thereby providing an architecture-independent expression of tolerant parallel algorithms. For implementing F-Nets, Cooperative Data Sharing (CDS) is a subroutine package for implementing communication efficiently in a large number of environments (e.g. shared memory and message passing). Software Cabling (SC), a very-high-level graphical programming language for building large F-Nets, possesses many of the features normally expected from today's computer languages (e.g. data abstraction, array operations). Finally, L2(sup 3) is a CASE tool which facilitates the construction, compilation, execution, and debugging of SC programs.

  15. Applied Parallel Metadata Indexing

    SciTech Connect

    Jacobi, Michael R

    2012-08-01

    The GPFS Archive is parallel archive is a parallel archive used by hundreds of users in the Turquoise collaboration network. It houses 4+ petabytes of data in more than 170 million files. Currently, users must navigate the file system to retrieve their data, requiring them to remember file paths and names. A better solution might allow users to tag data with meaningful labels and searach the archive using standard and user-defined metadata, while maintaining security. last summer, I developed the backend to a tool that adheres to these design goals. The backend works by importing GPFS metadata into a MongoDB cluster, which is then indexed on each attribute. This summer, the author implemented security and developed the user interfae for the search tool. To meet security requirements, each database table is associated with a single user, which only stores records that the user may read, and requires a set of credentials to access. The interface to the search tool is implemented using FUSE (Filesystem in USErspace). FUSE is an intermediate layer that intercepts file system calls and allows the developer to redefine how those calls behave. In the case of this tool, FUSE interfaces with MongoDB to issue queries and populate output. A FUSE implementation is desirable because it allows users to interact with the search tool using commands they are already familiar with. These security and interface additions are essential for a usable product.

  16. A systolic array parallelizing compiler

    SciTech Connect

    Tseng, P.S. )

    1990-01-01

    This book presents a completely new approach to the problem of systolic array parallelizing compiler. It describes the AL parallelizing compiler for the Warp systolic array, the first working systolic array parallelizing compiler which can generate efficient parallel code for complete LINPACK routines. This book begins by analyzing the architectural strength of the Warp systolic array. It proposes a model for mapping programs onto the machine and introduces the notion of data relations for optimizing the program mapping. Also presented are successful applications of the AL compiler in matrix computation and image processing. A complete listing of the source program and compiler-generated parallel code are given to clarify the overall picture of the compiler. The book concludes that systolic array parallelizing compiler can produce efficient parallel code, almost identical to what the user would have written by hand.

  17. Parallel Computing in SCALE

    SciTech Connect

    DeHart, Mark D; Williams, Mark L; Bowman, Stephen M

    2010-01-01

    The SCALE computational architecture has remained basically the same since its inception 30 years ago, although constituent modules and capabilities have changed significantly. This SCALE concept was intended to provide a framework whereby independent codes can be linked to provide a more comprehensive capability than possible with the individual programs - allowing flexibility to address a wide variety of applications. However, the current system was designed originally for mainframe computers with a single CPU and with significantly less memory than today's personal computers. It has been recognized that the present SCALE computation system could be restructured to take advantage of modern hardware and software capabilities, while retaining many of the modular features of the present system. Preliminary work is being done to define specifications and capabilities for a more advanced computational architecture. This paper describes the state of current SCALE development activities and plans for future development. With the release of SCALE 6.1 in 2010, a new phase of evolutionary development will be available to SCALE users within the TRITON and NEWT modules. The SCALE (Standardized Computer Analyses for Licensing Evaluation) code system developed by Oak Ridge National Laboratory (ORNL) provides a comprehensive and integrated package of codes and nuclear data for a wide range of applications in criticality safety, reactor physics, shielding, isotopic depletion and decay, and sensitivity/uncertainty (S/U) analysis. Over the last three years, since the release of version 5.1 in 2006, several important new codes have been introduced within SCALE, and significant advances applied to existing codes. Many of these new features became available with the release of SCALE 6.0 in early 2009. However, beginning with SCALE 6.1, a first generation of parallel computing is being introduced. In addition to near-term improvements, a plan for longer term SCALE enhancement

  18. Toward Parallel Document Clustering

    SciTech Connect

    Mogill, Jace A.; Haglin, David J.

    2011-09-01

    A key challenge to automated clustering of documents in large text corpora is the high cost of comparing documents in a multimillion dimensional document space. The Anchors Hierarchy is a fast data structure and algorithm for localizing data based on a triangle inequality obeying distance metric, the algorithm strives to minimize the number of distance calculations needed to cluster the documents into “anchors” around reference documents called “pivots”. We extend the original algorithm to increase the amount of available parallelism and consider two implementations: a complex data structure which affords efficient searching, and a simple data structure which requires repeated sorting. The sorting implementation is integrated with a text corpora “Bag of Words” program and initial performance results of end-to-end a document processing workflow are reported.

  19. Parallel Polarization State Generation

    NASA Astrophysics Data System (ADS)

    She, Alan; Capasso, Federico

    2016-05-01

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security.

  20. Parallel Polarization State Generation

    PubMed Central

    She, Alan; Capasso, Federico

    2016-01-01

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security. PMID:27184813

  1. A parallel programming environment supporting multiple data-parallel modules

    SciTech Connect

    Seevers, B.K.; Quinn, M.J. ); Hatcher, P.J. )

    1992-10-01

    We describe a system that allows programmers to take advantage of both control and data parallelism through multiple intercommunicating data-parallel modules. This programming environment extends C-type stream I/O to include intermodule communication channels. The progammer writes each module as a separate data-parallel program, then develops a channel linker specification describing how to connect the modules together. A channel linker we have developed loads the separate modules on the parallel machine and binds the communication channels together as specified. We present performance data that demonstrates a mixed control- and data-parallel solution can yield better performance than a strictly data-parallel solution. The system described currently runs on the Intel iWarp multicomputer.

  2. Parallel imaging microfluidic cytometer.

    PubMed

    Ehrlich, Daniel J; McKenna, Brian K; Evans, James G; Belkina, Anna C; Denis, Gerald V; Sherr, David H; Cheung, Man Ching

    2011-01-01

    By adding an additional degree of freedom from multichannel flow, the parallel microfluidic cytometer (PMC) combines some of the best features of fluorescence-activated flow cytometry (FCM) and microscope-based high-content screening (HCS). The PMC (i) lends itself to fast processing of large numbers of samples, (ii) adds a 1D imaging capability for intracellular localization assays (HCS), (iii) has a high rare-cell sensitivity, and (iv) has an unusual capability for time-synchronized sampling. An inability to practically handle large sample numbers has restricted applications of conventional flow cytometers and microscopes in combinatorial cell assays, network biology, and drug discovery. The PMC promises to relieve a bottleneck in these previously constrained applications. The PMC may also be a powerful tool for finding rare primary cells in the clinic. The multichannel architecture of current PMC prototypes allows 384 unique samples for a cell-based screen to be read out in ∼6-10 min, about 30 times the speed of most current FCM systems. In 1D intracellular imaging, the PMC can obtain protein localization using HCS marker strategies at many times for the sample throughput of charge-coupled device (CCD)-based microscopes or CCD-based single-channel flow cytometers. The PMC also permits the signal integration time to be varied over a larger range than is practical in conventional flow cytometers. The signal-to-noise advantages are useful, for example, in counting rare positive cells in the most difficult early stages of genome-wide screening. We review the status of parallel microfluidic cytometry and discuss some of the directions the new technology may take.

  3. Aerial photographic interpretation of lineaments and faults in late Cenozoic deposits in the eastern parts of the Saline Valley 1:100, 000 quadrangle, Nevada and California, and the Darwin Hills 1:100, 000 quadrangle, California

    SciTech Connect

    Reheis, M.C.

    1991-09-01

    Faults and fault-related lineaments in Quaternary and late Tertiary deposits in the southern part of the Walker Lane are potentially active and form patterns that are anomalous compared to those in most other areas of the Great Basin. Two maps at a scale of 1:100,000 summarize information about lineaments and faults in the area around and southwest of the Death Valley-Furnace Creek fault system based on extensive aerial-photo interpretation, limited field interpretation, limited field investigations, and published geologic maps. There are three major fault zones and two principal faults in the Saline Valley and Darwin Hills 1:100,000 quadrangles. (1) The Death Valley-Furnace Creek fault system and (2) the Hunter Mountain fault zone are northwest-trending right-lateral strike-slip fault zones. (3) The Panamint Valley fault zone and associated Towne Pass and Emigrant faults are north-trending normal faults. The intersection of the Hunter Mountain and Panamint Valley fault zones is marked by a large complex of faults and lineaments on the floor of Panamint Valley. Additional major faults include (4) the north-northwest-trending Ash Hill fault on the west side of Panamint Valley, and (5) the north-trending range-front Tin Mountain fault on the west side of the northern Cottonwood Mountains. The most active faults at present include those along the Death Valley-Furnace Creek fault system, the Tin Mountain fault, the northwest and southeast ends of the Hunter Mountain fault zone, the Ash Hill fault, and the fault bounding the west side of the Panamint Range south of Hall Canyon. Several large Quaternary landslides on the west sides of the Cottonwood Mountains and the Panamint Range apparently reflect slope instability due chiefly to rapid uplift of these ranges. 16 refs.

  4. Using high-precision 40Ar/39Ar geochronology to understand volcanic hazards within the Rio Grande rift and along the Jemez lineament, New Mexico

    NASA Astrophysics Data System (ADS)

    Zimmerer, M. J.; McIntosh, W. C.; Heizler, M. T.; Lafferty, J.

    2014-12-01

    High-precision Ar/Ar ages were generated for late Quaternary volcanic fields in the Rio Grande rift and along the Jemez Lineament, New Mexico, to assess the time-space patterns of volcanism and begin quantifying volcanic hazards for the region. The published chronology of most late Quaternary volcanic centers in the region is not sufficiently precise, accurate, or complete for a comprehensive volcanic hazard assessment. Ar/Ar ages generated as part of this study were determined using the high-sensitivity, multi-collector ARGUS VI mass spectrometer, which provides about an order of magnitude more precise isotopic measurements compared to older generation, single-detector mass spectrometers. Ar/Ar ages suggest an apparent increase in eruption frequency during the late Quaternary within the Raton-Clayton volcanic field, northeastern NM. Only four volcanoes erupted between 426±8 and 97±3 ka. Contrastingly, four volcanoes erupted between 55±2 and 32±5 ka. This last eruptive phase displays a west to east migration of volcanism, has repose periods of 0 to 17 ka, and an average recurrence rate of 1 eruption per 5750 ka. The Zuni-Bandera volcanic field, west-central NM, is composed of the ~100 late Quaternary basaltic vents. Preliminary results suggest that most of the Chain of Craters, the largest and oldest part of the Zuni-Bandera field, erupted between ~100 and 250 ka. Volcanism then migrated to the east, where published ages indicate at least seven eruptions between 50 and 3 ka. Both volcanic fields display a west to east migration of volcanism during the last ~500 ka, although the pattern is more pronounced in the Zuni-Bandera field. A reassessment of low-precision published ages for other late Quaternary volcanic fields in region indicates that most fields display a similar west to east migration of volcanism during the last ~500 ka. One possible mechanism to explain the observed patterns of volcanism is the westward migration of the North American plate relative

  5. Parallel processor engine model program

    NASA Technical Reports Server (NTRS)

    Mclaughlin, P.

    1984-01-01

    The Parallel Processor Engine Model Program is a generalized engineering tool intended to aid in the design of parallel processing real-time simulations of turbofan engines. It is written in the FORTRAN programming language and executes as a subset of the SOAPP simulation system. Input/output and execution control are provided by SOAPP; however, the analysis, emulation and simulation functions are completely self-contained. A framework in which a wide variety of parallel processing architectures could be evaluated and tools with which the parallel implementation of a real-time simulation technique could be assessed are provided.

  6. Parallel processing and expert systems

    NASA Technical Reports Server (NTRS)

    Yan, Jerry C.; Lau, Sonie

    1991-01-01

    Whether it be monitoring the thermal subsystem of Space Station Freedom, or controlling the navigation of the autonomous rover on Mars, NASA missions in the 90's cannot enjoy an increased level of autonomy without the efficient use of expert systems. Merely increasing the computational speed of uniprocessors may not be able to guarantee that real time demands are met for large expert systems. Speed-up via parallel processing must be pursued alongside the optimization of sequential implementations. Prototypes of parallel expert systems have been built at universities and industrial labs in the U.S. and Japan. The state-of-the-art research in progress related to parallel execution of expert systems was surveyed. The survey is divided into three major sections: (1) multiprocessors for parallel expert systems; (2) parallel languages for symbolic computations; and (3) measurements of parallelism of expert system. Results to date indicate that the parallelism achieved for these systems is small. In order to obtain greater speed-ups, data parallelism and application parallelism must be exploited.

  7. Parallel Programming in the Age of Ubiquitous Parallelism

    NASA Astrophysics Data System (ADS)

    Pingali, Keshav

    2014-04-01

    Multicore and manycore processors are now ubiquitous, but parallel programming remains as difficult as it was 30-40 years ago. During this time, our community has explored many promising approaches including functional and dataflow languages, logic programming, and automatic parallelization using program analysis and restructuring, but none of these approaches has succeeded except in a few niche application areas. In this talk, I will argue that these problems arise largely from the computation-centric foundations and abstractions that we currently use to think about parallelism. In their place, I will propose a novel data-centric foundation for parallel programming called the operator formulation in which algorithms are described in terms of actions on data. The operator formulation shows that a generalized form of data-parallelism called amorphous data-parallelism is ubiquitous even in complex, irregular graph applications such as mesh generation/refinement/partitioning and SAT solvers. Regular algorithms emerge as a special case of irregular ones, and many application-specific optimization techniques can be generalized to a broader context. The operator formulation also leads to a structural analysis of algorithms called TAO-analysis that provides implementation guidelines for exploiting parallelism efficiently. Finally, I will describe a system called Galois based on these ideas for exploiting amorphous data-parallelism on multicores and GPUs

  8. Trajectories in parallel optics.

    PubMed

    Klapp, Iftach; Sochen, Nir; Mendlovic, David

    2011-10-01

    In our previous work we showed the ability to improve the optical system's matrix condition by optical design, thereby improving its robustness to noise. It was shown that by using singular value decomposition, a target point-spread function (PSF) matrix can be defined for an auxiliary optical system, which works parallel to the original system to achieve such an improvement. In this paper, after briefly introducing the all optics implementation of the auxiliary system, we show a method to decompose the target PSF matrix. This is done through a series of shifted responses of auxiliary optics (named trajectories), where a complicated hardware filter is replaced by postprocessing. This process manipulates the pixel confined PSF response of simple auxiliary optics, which in turn creates an auxiliary system with the required PSF matrix. This method is simulated on two space variant systems and reduces their system condition number from 18,598 to 197 and from 87,640 to 5.75, respectively. We perform a study of the latter result and show significant improvement in image restoration performance, in comparison to a system without auxiliary optics and to other previously suggested hybrid solutions. Image restoration results show that in a range of low signal-to-noise ratio values, the trajectories method gives a significant advantage over alternative approaches. A third space invariant study case is explored only briefly, and we present a significant improvement in the matrix condition number from 1.9160e+013 to 34,526.

  9. High Performance Parallel Architectures

    NASA Technical Reports Server (NTRS)

    El-Ghazawi, Tarek; Kaewpijit, Sinthop

    1998-01-01

    Traditional remote sensing instruments are multispectral, where observations are collected at a few different spectral bands. Recently, many hyperspectral instruments, that can collect observations at hundreds of bands, have been operational. Furthermore, there have been ongoing research efforts on ultraspectral instruments that can produce observations at thousands of spectral bands. While these remote sensing technology developments hold great promise for new findings in the area of Earth and space science, they present many challenges. These include the need for faster processing of such increased data volumes, and methods for data reduction. Dimension Reduction is a spectral transformation, aimed at concentrating the vital information and discarding redundant data. One such transformation, which is widely used in remote sensing, is the Principal Components Analysis (PCA). This report summarizes our progress on the development of a parallel PCA and its implementation on two Beowulf cluster configuration; one with fast Ethernet switch and the other with a Myrinet interconnection. Details of the implementation and performance results, for typical sets of multispectral and hyperspectral NASA remote sensing data, are presented and analyzed based on the algorithm requirements and the underlying machine configuration. It will be shown that the PCA application is quite challenging and hard to scale on Ethernet-based clusters. However, the measurements also show that a high- performance interconnection network, such as Myrinet, better matches the high communication demand of PCA and can lead to a more efficient PCA execution.

  10. Geology, geochemistry, geochronology, and economic potential of Neogene volcanic rocks in the Laguna Pedernal and Salar de Aguas Calientes segments of the Archibarca lineament, northwest Argentina

    NASA Astrophysics Data System (ADS)

    Richards, J. P.; Jourdan, F.; Creaser, R. A.; Maldonado, G.; DuFrane, S. A.

    2013-05-01

    This study presents new geochemical, geochronological, isotopic, and mineralogical data, combined with new geological mapping for a 2400 km2 area of Neogene volcanic rocks in northwestern Argentina near the border with Chile, between 25°10‧S and 25°45‧S. The area covers the zone of intersection between the main axis of the Cordillera Occidental and a set of NW-SE-trending structures that form part of the transverse Archibarca lineament. This lineament has localized major ore deposits in Chile (e.g., the late Eocene La Escondida porphyry Cu deposit) and large volcanic centers such as the active Llullaillaco and Lastarría volcanoes on the border between Chile and Argentina, and the Neogene Archibarca, Antofalla, and Cerro Galán volcanoes in Argentina. Neogene volcanic rocks in the Laguna Pedernal and Salar de Aguas Calientes areas are mostly high-K calc-alkaline in composition, and range from basaltic andesites, through andesites and dacites, to rhyolites. Magmatic temperatures and oxidation states, estimated from mineral compositions, range from ~ 1000 °C and ∆FMQ ≈ 1.0-1.5 in andesites, to ~ 850 °C and ∆FMQ ≈ 1.5-2.0 in dacites and rhyolites. The oldest rocks consist of early-middle Miocene andesite-dacite plagioclase-pyroxene-phyric lava flows and ignimbrites, with 40Ar/39Ar ages ranging from 17.14 ± 0.10 Ma to 11.76 ± 0.27 Ma. Their major and trace element compositions are typical of the Andean Central Volcanic Zone, and show strong crustal contamination trends for highly incompatible elements such as Cs, Rb, Th, and U. These rocks are geochemically grouped as sub-suite 1. This widespread intermediate composition volcanism was followed in the middle-late Miocene by a period of more focused rhyodacitic flow-dome complex formation. These felsic rocks are characterized by less extreme enrichments in highly incompatible elements, and increasing depletion of heavy rare earth elements. These rocks are geochemically grouped as sub-suite 2. The

  11. Parallel Computational Protein Design

    PubMed Central

    Zhou, Yichao; Donald, Bruce R.; Zeng, Jianyang

    2016-01-01

    Computational structure-based protein design (CSPD) is an important problem in computational biology, which aims to design or improve a prescribed protein function based on a protein structure template. It provides a practical tool for real-world protein engineering applications. A popular CSPD method that guarantees to find the global minimum energy solution (GMEC) is to combine both dead-end elimination (DEE) and A* tree search algorithms. However, in this framework, the A* search algorithm can run in exponential time in the worst case, which may become the computation bottleneck of large-scale computational protein design process. To address this issue, we extend and add a new module to the OSPREY program that was previously developed in the Donald lab [1] to implement a GPU-based massively parallel A* algorithm for improving protein design pipeline. By exploiting the modern GPU computational framework and optimizing the computation of the heuristic function for A* search, our new program, called gOSPREY, can provide up to four orders of magnitude speedups in large protein design cases with a small memory overhead comparing to the traditional A* search algorithm implementation, while still guaranteeing the optimality. In addition, gOSPREY can be configured to run in a bounded-memory mode to tackle the problems in which the conformation space is too large and the global optimal solution cannot be computed previously. Furthermore, the GPU-based A* algorithm implemented in the gOSPREY program can be combined with the state-of-the-art rotamer pruning algorithms such as iMinDEE [2] and DEEPer [3] to also consider continuous backbone and side-chain flexibility. PMID:27914056

  12. Parallel Adaptive Mesh Refinement

    SciTech Connect

    Diachin, L; Hornung, R; Plassmann, P; WIssink, A

    2005-03-04

    As large-scale, parallel computers have become more widely available and numerical models and algorithms have advanced, the range of physical phenomena that can be simulated has expanded dramatically. Many important science and engineering problems exhibit solutions with localized behavior where highly-detailed salient features or large gradients appear in certain regions which are separated by much larger regions where the solution is smooth. Examples include chemically-reacting flows with radiative heat transfer, high Reynolds number flows interacting with solid objects, and combustion problems where the flame front is essentially a two-dimensional sheet occupying a small part of a three-dimensional domain. Modeling such problems numerically requires approximating the governing partial differential equations on a discrete domain, or grid. Grid spacing is an important factor in determining the accuracy and cost of a computation. A fine grid may be needed to resolve key local features while a much coarser grid may suffice elsewhere. Employing a fine grid everywhere may be inefficient at best and, at worst, may make an adequately resolved simulation impractical. Moreover, the location and resolution of fine grid required for an accurate solution is a dynamic property of a problem's transient features and may not be known a priori. Adaptive mesh refinement (AMR) is a technique that can be used with both structured and unstructured meshes to adjust local grid spacing dynamically to capture solution features with an appropriate degree of resolution. Thus, computational resources can be focused where and when they are needed most to efficiently achieve an accurate solution without incurring the cost of a globally-fine grid. Figure 1.1 shows two example computations using AMR; on the left is a structured mesh calculation of a impulsively-sheared contact surface and on the right is the fuselage and volume discretization of an RAH-66 Comanche helicopter [35]. Note the

  13. A Parallel Particle Swarm Optimizer

    DTIC Science & Technology

    2003-01-01

    by a computationally demanding biomechanical system identification problem, we introduce a parallel implementation of a stochastic population based...concurrent computation. The parallelization of the Particle Swarm Optimization (PSO) algorithm is detailed and its performance and characteristics demonstrated for the biomechanical system identification problem as example.

  14. The Galley Parallel File System

    NASA Technical Reports Server (NTRS)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  15. Parallel contingency statistics with Titan.

    SciTech Connect

    Thompson, David C.; Pebay, Philippe Pierre

    2009-09-01

    This report summarizes existing statistical engines in VTK/Titan and presents the recently parallelized contingency statistics engine. It is a sequel to [PT08] and [BPRT09] which studied the parallel descriptive, correlative, multi-correlative, and principal component analysis engines. The ease of use of this new parallel engines is illustrated by the means of C++ code snippets. Furthermore, this report justifies the design of these engines with parallel scalability in mind; however, the very nature of contingency tables prevent this new engine from exhibiting optimal parallel speed-up as the aforementioned engines do. This report therefore discusses the design trade-offs we made and study performance with up to 200 processors.

  16. Dextral Shear on the Olympic-Wallowa Lineament, Washington--Evidence from High-Resolution Aeromagnetic Anomalies and Implications for Cascadia Seismic Hazards

    NASA Astrophysics Data System (ADS)

    Blakely, R. J.; Sherrod, B. L.; Weaver, C. S.; Wells, R. E.

    2013-12-01

    The Olympic-Wallowa physiographic lineament (OWL) extends northwestward ~500 km from northeastern Oregon to Vancouver Island, British Columbia. The tectonic relevance of the OWL, particularly the degree to which horizontal shear has contributed to its evolution, is an important element in assessing kinematic connections between the Cascadia backarc and forearc, and the consequent seismic hazard it poses to the region. Past workers have come to rather different conclusions, some suggesting sinistral, others suggesting dextral, and still others suggesting no horizontal displacement on the OWL. North-northwest-striking dikes of the 8.5 Ma Ice Harbor Member of the Columbia River Basalt Group are offset and disrupted by the Wallula fault zone (WFZ), a 45-km-long segment of the OWL southeast of Kennewick. Although mostly concealed by young deposits, Ice Harbor dikes are clearly delineated in high-resolution aeromagnetic anomalies as near-vertical intrusions, affording an opportunity to estimate the sense and amount of offset along this part of the OWL. Aided by various derivative products calculated from magnetic anomalies, we interpret five piercing points from single-dike anomalies intersecting the Wallula fault (the northernmost strand of the WFZ), together indicating an average of 1.72 km of right-lateral offset on this single strand. Right-lateral offset across the entire WFZ is 6.9 km, an average rate of 0.8 mm/y since 8.5 Ma. We cannot rule out the possibility that tectonic offsets are only apparent because the injection process itself was offset during emplacement. However, consistent right-lateral offsets are observed in stream drainages along the OWL, supporting our magnetic anomaly interpretations. Aerial photography and airborne LiDAR surveys show seven offset streams along a fault scarp on the northeastern slope of Rattlesnake Mountain, located along the OWL about 70 km northwest of the WFZ. Incised in Pleistocene deposits, these stream offsets average about

  17. Parallel NPARC: Implementation and Performance

    NASA Technical Reports Server (NTRS)

    Townsend, S. E.

    1996-01-01

    Version 3 of the NPARC Navier-Stokes code includes support for large-grain (block level) parallelism using explicit message passing between a heterogeneous collection of computers. This capability has the potential for significant performance gains, depending upon the block data distribution. The parallel implementation uses a master/worker arrangement of processes. The master process assigns blocks to workers, controls worker actions, and provides remote file access for the workers. The processes communicate via explicit message passing using an interface library which provides portability to a number of message passing libraries, such as PVM (Parallel Virtual Machine). A Bourne shell script is used to simplify the task of selecting hosts, starting processes, retrieving remote files, and terminating a computation. This script also provides a simple form of fault tolerance. An analysis of the computational performance of NPARC is presented, using data sets from an F/A-18 inlet study and a Rocket Based Combined Cycle Engine analysis. Parallel speedup and overall computational efficiency were obtained for various NPARC run parameters on a cluster of IBM RS6000 workstations. The data show that although NPARC performance compares favorably with the estimated potential parallelism, typical data sets used with previous versions of NPARC will often need to be reblocked for optimum parallel performance. In one of the cases studied, reblocking increased peak parallel speedup from 3.2 to 11.8.

  18. Parallel processing and expert systems

    NASA Technical Reports Server (NTRS)

    Lau, Sonie; Yan, Jerry C.

    1991-01-01

    Whether it be monitoring the thermal subsystem of Space Station Freedom, or controlling the navigation of the autonomous rover on Mars, NASA missions in the 1990s cannot enjoy an increased level of autonomy without the efficient implementation of expert systems. Merely increasing the computational speed of uniprocessors may not be able to guarantee that real-time demands are met for larger systems. Speedup via parallel processing must be pursued alongside the optimization of sequential implementations. Prototypes of parallel expert systems have been built at universities and industrial laboratories in the U.S. and Japan. The state-of-the-art research in progress related to parallel execution of expert systems is surveyed. The survey discusses multiprocessors for expert systems, parallel languages for symbolic computations, and mapping expert systems to multiprocessors. Results to date indicate that the parallelism achieved for these systems is small. The main reasons are (1) the body of knowledge applicable in any given situation and the amount of computation executed by each rule firing are small, (2) dividing the problem solving process into relatively independent partitions is difficult, and (3) implementation decisions that enable expert systems to be incrementally refined hamper compile-time optimization. In order to obtain greater speedups, data parallelism and application parallelism must be exploited.

  19. Parallel integer sorting with medium and fine-scale parallelism

    NASA Technical Reports Server (NTRS)

    Dagum, Leonardo

    1993-01-01

    Two new parallel integer sorting algorithms, queue-sort and barrel-sort, are presented and analyzed in detail. These algorithms do not have optimal parallel complexity, yet they show very good performance in practice. Queue-sort designed for fine-scale parallel architectures which allow the queueing of multiple messages to the same destination. Barrel-sort is designed for medium-scale parallel architectures with a high message passing overhead. The performance results from the implementation of queue-sort on a Connection Machine CM-2 and barrel-sort on a 128 processor iPSC/860 are given. The two implementations are found to be comparable in performance but not as good as a fully vectorized bucket sort on the Cray YMP.

  20. EFFICIENT SCHEDULING OF PARALLEL JOBS ON MASSIVELY PARALLEL SYSTEMS

    SciTech Connect

    F. PETRINI; W. FENG

    1999-09-01

    We present buffered coscheduling, a new methodology to multitask parallel jobs in a message-passing environment and to develop parallel programs that can pave the way to the efficient implementation of a distributed operating system. Buffered coscheduling is based on three innovative techniques: communication buffering, strobing, and non-blocking communication. By leveraging these techniques, we can perform effective optimizations based on the global status of the parallel machine rather than on the limited knowledge available locally to each processor. The advantages of buffered coscheduling include higher resource utilization, reduced communication overhead, efficient implementation of low-control strategies and fault-tolerant protocols, accurate performance modeling, and a simplified yet still expressive parallel programming model. Preliminary experimental results show that buffered coscheduling is very effective in increasing the overall performance in the presence of load imbalance and communication-intensive workloads.

  1. Template based parallel checkpointing in a massively parallel computer system

    DOEpatents

    Archer, Charles Jens; Inglett, Todd Alan

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  2. Parallel Architecture For Robotics Computation

    NASA Technical Reports Server (NTRS)

    Fijany, Amir; Bejczy, Antal K.

    1990-01-01

    Universal Real-Time Robotic Controller and Simulator (URRCS) is highly parallel computing architecture for control and simulation of robot motion. Result of extensive algorithmic study of different kinematic and dynamic computational problems arising in control and simulation of robot motion. Study led to development of class of efficient parallel algorithms for these problems. Represents algorithmically specialized architecture, in sense capable of exploiting common properties of this class of parallel algorithms. System with both MIMD and SIMD capabilities. Regarded as processor attached to bus of external host processor, as part of bus memory.

  3. Multigrid on massively parallel architectures

    SciTech Connect

    Falgout, R D; Jones, J E

    1999-09-17

    The scalable implementation of multigrid methods for machines with several thousands of processors is investigated. Parallel performance models are presented for three different structured-grid multigrid algorithms, and a description is given of how these models can be used to guide implementation. Potential pitfalls are illustrated when moving from moderate-sized parallelism to large-scale parallelism, and results are given from existing multigrid codes to support the discussion. Finally, the use of mixed programming models is investigated for multigrid codes on clusters of SMPs.

  4. IOPA: I/O-aware parallelism adaption for parallel programs

    PubMed Central

    Liu, Tao; Liu, Yi; Qian, Chen; Qian, Depei

    2017-01-01

    With the development of multi-/many-core processors, applications need to be written as parallel programs to improve execution efficiency. For data-intensive applications that use multiple threads to read/write files simultaneously, an I/O sub-system can easily become a bottleneck when too many of these types of threads exist; on the contrary, too few threads will cause insufficient resource utilization and hurt performance. Therefore, programmers must pay much attention to parallelism control to find the appropriate number of I/O threads for an application. This paper proposes a parallelism control mechanism named IOPA that can adjust the parallelism of applications to adapt to the I/O capability of a system and balance computing resources and I/O bandwidth. The programming interface of IOPA is also provided to programmers to simplify parallel programming. IOPA is evaluated using multiple applications with both solid state and hard disk drives. The results show that the parallel applications using IOPA can achieve higher efficiency than those with a fixed number of threads. PMID:28278236

  5. Appendix E: Parallel Pascal development system

    NASA Technical Reports Server (NTRS)

    1985-01-01

    The Parallel Pascal Development System enables Parallel Pascal programs to be developed and tested on a conventional computer. It consists of several system programs, including a Parallel Pascal to standard Pascal translator, and a library of Parallel Pascal subprograms. The library includes subprograms for using Parallel Pascal on a parallel system with a fixed degree of parallelism, such as the Massively Parallel Processor, to conveniently manipulate arrays which have dimensions than the hardware. Programs can be conveninetly tested with small sized arrays on the conventional computer before attempting to run on a parallel system.

  6. Parallel hierarchical method in networks

    NASA Astrophysics Data System (ADS)

    Malinochka, Olha; Tymchenko, Leonid

    2007-09-01

    This method of parallel-hierarchical Q-transformation offers new approach to the creation of computing medium - of parallel -hierarchical (PH) networks, being investigated in the form of model of neurolike scheme of data processing [1-5]. The approach has a number of advantages as compared with other methods of formation of neurolike media (for example, already known methods of formation of artificial neural networks). The main advantage of the approach is the usage of multilevel parallel interaction dynamics of information signals at different hierarchy levels of computer networks, that enables to use such known natural features of computations organization as: topographic nature of mapping, simultaneity (parallelism) of signals operation, inlaid cortex, structure, rough hierarchy of the cortex, spatially correlated in time mechanism of perception and training [5].

  7. New NAS Parallel Benchmarks Results

    NASA Technical Reports Server (NTRS)

    Yarrow, Maurice; Saphir, William; VanderWijngaart, Rob; Woo, Alex; Kutler, Paul (Technical Monitor)

    1997-01-01

    NPB2 (NAS (NASA Advanced Supercomputing) Parallel Benchmarks 2) is an implementation, based on Fortran and the MPI (message passing interface) message passing standard, of the original NAS Parallel Benchmark specifications. NPB2 programs are run with little or no tuning, in contrast to NPB vendor implementations, which are highly optimized for specific architectures. NPB2 results complement, rather than replace, NPB results. Because they have not been optimized by vendors, NPB2 implementations approximate the performance a typical user can expect for a portable parallel program on distributed memory parallel computers. Together these results provide an insightful comparison of the real-world performance of high-performance computers. New NPB2 features: New implementation (CG), new workstation class problem sizes, new serial sample versions, more performance statistics.

  8. "Feeling" Series and Parallel Resistances.

    ERIC Educational Resources Information Center

    Morse, Robert A.

    1993-01-01

    Equipped with drinking straws and stirring straws, a teacher can help students understand how resistances in electric circuits combine in series and in parallel. Follow-up suggestions are provided. (ZWH)

  9. Demonstrating Forces between Parallel Wires.

    ERIC Educational Resources Information Center

    Baker, Blane

    2000-01-01

    Describes a physics demonstration that dramatically illustrates the mutual repulsion (attraction) between parallel conductors using insulated copper wire, wooden dowels, a high direct current power supply, electrical tape, and an overhead projector. (WRM)

  10. Parallel programming of industrial applications

    SciTech Connect

    Heroux, M; Koniges, A; Simon, H

    1998-07-21

    In the introductory material, we overview the typical MPP environment for real application computing and the special tools available such as parallel debuggers and performance analyzers. Next, we draw from a series of real applications codes and discuss the specific challenges and problems that are encountered in parallelizing these individual applications. The application areas drawn from include biomedical sciences, materials processing and design, plasma and fluid dynamics, and others. We show how it was possible to get a particular application to run efficiently and what steps were necessary. Finally we end with a summary of the lessons learned from these applications and predictions for the future of industrial parallel computing. This tutorial is based on material from a forthcoming book entitled: "Industrial Strength Parallel Computing" to be published by Morgan Kaufmann Publishers (ISBN l-55860-54).

  11. Distinguishing serial and parallel parsing.

    PubMed

    Gibson, E; Pearlmutter, N J

    2000-03-01

    This paper discusses ways of determining whether the human parser is serial maintaining at most, one structural interpretation at each parse state, or whether it is parallel, maintaining more than one structural interpretation in at least some circumstances. We make four points. The first two counterclaims made by Lewis (2000): (1) that the availability of alternative structures should not vary as a function of the disambiguating material in some ranked parallel models; and (2) that parallel models predict a slow down during the ambiguous region for more syntactically ambiguous structures. Our other points concern potential methods for seeking experimental evidence relevant to the serial/parallel question. We discuss effects of the plausibility of a secondary structure in the ambiguous region (Pearlmutter & Mendelsohn, 1999) and suggest examining the distribution of reaction times in the disambiguating region.

  12. Address tracing for parallel machines

    NASA Technical Reports Server (NTRS)

    Stunkel, Craig B.; Janssens, Bob; Fuchs, W. Kent

    1991-01-01

    Recently implemented parallel system address-tracing methods based on several metrics are surveyed. The issues specific to collection of traces for both shared and distributed memory parallel computers are highlighted. Five general categories of address-trace collection methods are examined: hardware-captured, interrupt-based, simulation-based, altered microcode-based, and instrumented program-based traces. The problems unique to shared memory and distributed memory multiprocessors are examined separately.

  13. Parallel Algorithms for Image Analysis.

    DTIC Science & Technology

    1982-06-01

    8217 _ _ _ _ _ _ _ 4. TITLE (aid Subtitle) S. TYPE OF REPORT & PERIOD COVERED PARALLEL ALGORITHMS FOR IMAGE ANALYSIS TECHNICAL 6. PERFORMING O4G. REPORT NUMBER TR-1180...Continue on reverse side it neceesary aid Identlfy by block number) Image processing; image analysis ; parallel processing; cellular computers. 20... IMAGE ANALYSIS TECHNICAL 6. PERFORMING ONG. REPORT NUMBER TR-1180 - 7. AUTHOR(&) S. CONTRACT OR GRANT NUMBER(s) Azriel Rosenfeld AFOSR-77-3271 9

  14. Debugging in a parallel environment

    SciTech Connect

    Wasserman, H.J.; Griffin, J.H.

    1985-01-01

    This paper describes the preliminary results of a project investigating approaches to dynamic debugging in parallel processing systems. Debugging programs in a multiprocessing environment is particularly difficult because of potential errors in synchronization of tasks, data dependencies, sharing of data among tasks, and irreproducibility of specific machine instruction sequences from one job to the next. The basic methodology involved in predicate-based debuggers is given as well as other desirable features of dynamic parallel debugging. 13 refs.

  15. Efficiency of parallel direct optimization

    NASA Technical Reports Server (NTRS)

    Janies, D. A.; Wheeler, W. C.

    2001-01-01

    Tremendous progress has been made at the level of sequential computation in phylogenetics. However, little attention has been paid to parallel computation. Parallel computing is particularly suited to phylogenetics because of the many ways large computational problems can be broken into parts that can be analyzed concurrently. In this paper, we investigate the scaling factors and efficiency of random addition and tree refinement strategies using the direct optimization software, POY, on a small (10 slave processors) and a large (256 slave processors) cluster of networked PCs running LINUX. These algorithms were tested on several data sets composed of DNA and morphology ranging from 40 to 500 taxa. Various algorithms in POY show fundamentally different properties within and between clusters. All algorithms are efficient on the small cluster for the 40-taxon data set. On the large cluster, multibuilding exhibits excellent parallel efficiency, whereas parallel building is inefficient. These results are independent of data set size. Branch swapping in parallel shows excellent speed-up for 16 slave processors on the large cluster. However, there is no appreciable speed-up for branch swapping with the further addition of slave processors (>16). This result is independent of data set size. Ratcheting in parallel is efficient with the addition of up to 32 processors in the large cluster. This result is independent of data set size. c2001 The Willi Hennig Society.

  16. Architectures for reasoning in parallel

    NASA Technical Reports Server (NTRS)

    Hall, Lawrence O.

    1989-01-01

    The research conducted has dealt with rule-based expert systems. The algorithms that may lead to effective parallelization of them were investigated. Both the forward and backward chained control paradigms were investigated in the course of this work. The best computer architecture for the developed and investigated algorithms has been researched. Two experimental vehicles were developed to facilitate this research. They are Backpac, a parallel backward chained rule-based reasoning system and Datapac, a parallel forward chained rule-based reasoning system. Both systems have been written in Multilisp, a version of Lisp which contains the parallel construct, future. Applying the future function to a function causes the function to become a task parallel to the spawning task. Additionally, Backpac and Datapac have been run on several disparate parallel processors. The machines are an Encore Multimax with 10 processors, the Concert Multiprocessor with 64 processors, and a 32 processor BBN GP1000. Both the Concert and the GP1000 are switch-based machines. The Multimax has all its processors hung off a common bus. All are shared memory machines, but have different schemes for sharing the memory and different locales for the shared memory. The main results of the investigations come from experiments on the 10 processor Encore and the Concert with partitions of 32 or less processors. Additionally, experiments have been run with a stripped down version of EMYCIN.

  17. Efficiency of parallel direct optimization.

    PubMed

    Janies, D A; Wheeler, W C

    2001-03-01

    Tremendous progress has been made at the level of sequential computation in phylogenetics. However, little attention has been paid to parallel computation. Parallel computing is particularly suited to phylogenetics because of the many ways large computational problems can be broken into parts that can be analyzed concurrently. In this paper, we investigate the scaling factors and efficiency of random addition and tree refinement strategies using the direct optimization software, POY, on a small (10 slave processors) and a large (256 slave processors) cluster of networked PCs running LINUX. These algorithms were tested on several data sets composed of DNA and morphology ranging from 40 to 500 taxa. Various algorithms in POY show fundamentally different properties within and between clusters. All algorithms are efficient on the small cluster for the 40-taxon data set. On the large cluster, multibuilding exhibits excellent parallel efficiency, whereas parallel building is inefficient. These results are independent of data set size. Branch swapping in parallel shows excellent speed-up for 16 slave processors on the large cluster. However, there is no appreciable speed-up for branch swapping with the further addition of slave processors (>16). This result is independent of data set size. Ratcheting in parallel is efficient with the addition of up to 32 processors in the large cluster. This result is independent of data set size.

  18. Role of local to regional-scale collisions in the closure history of the Southern Neotethys, exemplified by tectonic development of the Kyrenia Range active margin/collisional lineament, N Cyprus

    NASA Astrophysics Data System (ADS)

    Robertson, Alastair; Kinnaird, Tim; McCay, Gillian; Palamakumbura, Romesh; Chen, Guohui

    2016-04-01

    Active margin processes including subduction, accretion, arc magmatism and back-arc extension play a key role in the diachronous, and still incomplete closure of the S Neotethys. The S Neotethys rifted along the present-day Africa-Eurasia continental margin during the Late Triassic and, after sea-floor spreading, began to close related to northward subduction during the Late Cretaceous. The northern, active continental margin of the S Neotethys was bordered by several of the originally rifted continental fragments (e.g. Taurides). The present-day convergent lineament ranges from subaqueous (e.g. Mediterranean Ridge), to subaerial (e.g. SE Turkey). The active margin development is partially obscured by microcontinent-continent collision and post-collisional strike-slip deformation (e.g. Tauride-Arabian suture). However, the Kyrenia Range, N Cyprus provides an outstanding record of convergent margin to early stage collisional processes. It owes its existence to strong localised uplift during the Pleistocene, which probably resulted from the collision of a continental promontory of N Africa (Eratosthenes Seamount) with the long-lived S Neotethyan active margin to the north. A multi-stage convergence history is revealed, mainly from a combination of field structural, sedimentological and igneous geochemical studies. Initial Late Cretaceous convergence resulted in greenschist facies burial metamorphism that is likely to have been related to the collision, then rapid exhumation, of a continental fragment (stage 1). During the latest Cretaceous-Palaeogene, the Kyrenia lineament was characterised by subduction-influenced magmatism and syn-tectonic sediment deposition. Early to Mid-Eocene, S-directed thrusting and folding (stage 2) is likely to have been influenced by the suturing of the Izmir-Ankara-Erzincan ocean to the north ('N Neotethys'). Convergence continued during the Neogene, dominated by deep-water terrigenous gravity-flow accumulation in a foredeep setting

  19. Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer

    DOEpatents

    Archer, Charles J; Blocksome, Michael E; Ratterman, Joseph D; Smith, Brian E

    2014-02-11

    Endpoint-based parallel data processing in a parallel active messaging interface ('PAMI') of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective opeartion through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.

  20. Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer

    DOEpatents

    Archer, Charles J.; Blocksome, Michael A.; Ratterman, Joseph D.; Smith, Brian E.

    2014-08-12

    Endpoint-based parallel data processing in a parallel active messaging interface (`PAMI`) of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective operation through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.

  1. The economics of parallel trade.

    PubMed

    Danzon, P M

    1998-03-01

    The potential for parallel trade in the European Union (EU) has grown with the accession of low price countries and the harmonisation of registration requirements. Parallel trade implies a conflict between the principle of autonomy of member states to set their own pharmaceutical prices, the principle of free trade and the industrial policy goal of promoting innovative research and development (R&D). Parallel trade in pharmaceuticals does not yield the normal efficiency gains from trade because countries achieve low pharmaceutical prices by aggressive regulation, not through superior efficiency. In fact, parallel trade reduces economic welfare by undermining price differentials between markets. Pharmaceutical R&D is a global joint cost of serving all consumers worldwide; it accounts for roughly 30% of total costs. Optimal (welfare maximising) pricing to cover joint costs (Ramsey pricing) requires setting different prices in different markets, based on inverse demand elasticities. By contrast, parallel trade and regulation based on international price comparisons tend to force price convergence across markets. In response, manufacturers attempt to set a uniform 'euro' price. The primary losers from 'euro' pricing will be consumers in low income countries who will face higher prices or loss of access to new drugs. In the long run, even higher income countries are likely to be worse off with uniform prices, because fewer drugs will be developed. One policy option to preserve price differentials is to exempt on-patent products from parallel trade. An alternative is confidential contracting between individual manufacturers and governments to provide country-specific ex post discounts from the single 'euro' wholesale price, similar to rebates used by managed care in the US. This would preserve differentials in transactions prices even if parallel trade forces convergence of wholesale prices.

  2. Parallel Implicit Algorithms for CFD

    NASA Technical Reports Server (NTRS)

    Keyes, David E.

    1998-01-01

    The main goal of this project was efficient distributed parallel and workstation cluster implementations of Newton-Krylov-Schwarz (NKS) solvers for implicit Computational Fluid Dynamics (CFD.) "Newton" refers to a quadratically convergent nonlinear iteration using gradient information based on the true residual, "Krylov" to an inner linear iteration that accesses the Jacobian matrix only through highly parallelizable sparse matrix-vector products, and "Schwarz" to a domain decomposition form of preconditioning the inner Krylov iterations with primarily neighbor-only exchange of data between the processors. Prior experience has established that Newton-Krylov methods are competitive solvers in the CFD context and that Krylov-Schwarz methods port well to distributed memory computers. The combination of the techniques into Newton-Krylov-Schwarz was implemented on 2D and 3D unstructured Euler codes on the parallel testbeds that used to be at LaRC and on several other parallel computers operated by other agencies or made available by the vendors. Early implementations were made directly in Massively Parallel Integration (MPI) with parallel solvers we adapted from legacy NASA codes and enhanced for full NKS functionality. Later implementations were made in the framework of the PETSC library from Argonne National Laboratory, which now includes pseudo-transient continuation Newton-Krylov-Schwarz solver capability (as a result of demands we made upon PETSC during our early porting experiences). A secondary project pursued with funding from this contract was parallel implicit solvers in acoustics, specifically in the Helmholtz formulation. A 2D acoustic inverse problem has been solved in parallel within the PETSC framework.

  3. A parallel Jacobson-Oksman optimization algorithm. [parallel processing (computers)

    NASA Technical Reports Server (NTRS)

    Straeter, T. A.; Markos, A. T.

    1975-01-01

    A gradient-dependent optimization technique which exploits the vector-streaming or parallel-computing capabilities of some modern computers is presented. The algorithm, derived by assuming that the function to be minimized is homogeneous, is a modification of the Jacobson-Oksman serial minimization method. In addition to describing the algorithm, conditions insuring the convergence of the iterates of the algorithm and the results of numerical experiments on a group of sample test functions are presented. The results of these experiments indicate that this algorithm will solve optimization problems in less computing time than conventional serial methods on machines having vector-streaming or parallel-computing capabilities.

  4. Parallelizing Timed Petri Net simulations

    NASA Technical Reports Server (NTRS)

    Nicol, David M.

    1993-01-01

    The possibility of using parallel processing to accelerate the simulation of Timed Petri Nets (TPN's) was studied. It was recognized that complex system development tools often transform system descriptions into TPN's or TPN-like models, which are then simulated to obtain information about system behavior. Viewed this way, it was important that the parallelization of TPN's be as automatic as possible, to admit the possibility of the parallelization being embedded in the system design tool. Later years of the grant were devoted to examining the problem of joint performance and reliability analysis, to explore whether both types of analysis could be accomplished within a single framework. In this final report, the results of our studies are summarized. We believe that the problem of parallelizing TPN's automatically for MIMD architectures has been almost completely solved for a large and important class of problems. Our initial investigations into joint performance/reliability analysis are two-fold; it was shown that Monte Carlo simulation, with importance sampling, offers promise of joint analysis in the context of a single tool, and methods for the parallel simulation of general Continuous Time Markov Chains, a model framework within which joint performance/reliability models can be cast, were developed. However, very much more work is needed to determine the scope and generality of these approaches. The results obtained in our two studies, future directions for this type of work, and a list of publications are included.

  5. Visualizing Parallel Computer System Performance

    NASA Technical Reports Server (NTRS)

    Malony, Allen D.; Reed, Daniel A.

    1988-01-01

    Parallel computer systems are among the most complex of man's creations, making satisfactory performance characterization difficult. Despite this complexity, there are strong, indeed, almost irresistible, incentives to quantify parallel system performance using a single metric. The fallacy lies in succumbing to such temptations. A complete performance characterization requires not only an analysis of the system's constituent levels, it also requires both static and dynamic characterizations. Static or average behavior analysis may mask transients that dramatically alter system performance. Although the human visual system is remarkedly adept at interpreting and identifying anomalies in false color data, the importance of dynamic, visual scientific data presentation has only recently been recognized Large, complex parallel system pose equally vexing performance interpretation problems. Data from hardware and software performance monitors must be presented in ways that emphasize important events while eluding irrelevant details. Design approaches and tools for performance visualization are the subject of this paper.

  6. Features in Continuous Parallel Coordinates.

    PubMed

    Lehmann, Dirk J; Theisel, Holger

    2011-12-01

    Continuous Parallel Coordinates (CPC) are a contemporary visualization technique in order to combine several scalar fields, given over a common domain. They facilitate a continuous view for parallel coordinates by considering a smooth scalar field instead of a finite number of straight lines. We show that there are feature curves in CPC which appear to be the dominant structures of a CPC. We present methods to extract and classify them and demonstrate their usefulness to enhance the visualization of CPCs. In particular, we show that these feature curves are related to discontinuities in Continuous Scatterplots (CSP). We show this by exploiting a curve-curve duality between parallel and Cartesian coordinates, which is a generalization of the well-known point-line duality. Furthermore, we illustrate the theoretical considerations. Concluding, we discuss relations and aspects of the CPC's/CSP's features concerning the data analysis.

  7. PARAVT: Parallel Voronoi tessellation code

    NASA Astrophysics Data System (ADS)

    González, R. E.

    2016-10-01

    In this study, we present a new open source code for massive parallel computation of Voronoi tessellations (VT hereafter) in large data sets. The code is focused for astrophysical purposes where VT densities and neighbors are widely used. There are several serial Voronoi tessellation codes, however no open source and parallel implementations are available to handle the large number of particles/galaxies in current N-body simulations and sky surveys. Parallelization is implemented under MPI and VT using Qhull library. Domain decomposition takes into account consistent boundary computation between tasks, and includes periodic conditions. In addition, the code computes neighbors list, Voronoi density, Voronoi cell volume, density gradient for each particle, and densities on a regular grid. Code implementation and user guide are publicly available at https://github.com/regonzar/paravt.

  8. Parallel integrated frame synchronizer chip

    NASA Technical Reports Server (NTRS)

    Ghuman, Parminder Singh (Inventor); Solomon, Jeffrey Michael (Inventor); Bennett, Toby Dennis (Inventor)

    2000-01-01

    A parallel integrated frame synchronizer which implements a sequential pipeline process wherein serial data in the form of telemetry data or weather satellite data enters the synchronizer by means of a front-end subsystem and passes to a parallel correlator subsystem or a weather satellite data processing subsystem. When in a CCSDS mode, data from the parallel correlator subsystem passes through a window subsystem, then to a data alignment subsystem and then to a bit transition density (BTD)/cyclical redundancy check (CRC) decoding subsystem. Data from the BTD/CRC decoding subsystem or data from the weather satellite data processing subsystem is then fed to an output subsystem where it is output from a data output port.

  9. Fast data parallel polygon rendering

    SciTech Connect

    Ortega, F.A.; Hansen, C.D.

    1993-09-01

    This paper describes a parallel method for polygonal rendering on a massively parallel SIMD machine. This method, based on a simple shading model, is targeted for applications which require very fast polygon rendering for extremely large sets of polygons such as is found in many scientific visualization applications. The algorithms described in this paper are incorporated into a library of 3D graphics routines written for the Connection Machine. The routines are implemented on both the CM-200 and the CM-5. This library enables a scientists to display 3D shaded polygons directly from a parallel machine without the need to transmit huge amounts of data to a post-processing rendering system.

  10. Massively Parallel MRI Detector Arrays

    PubMed Central

    Keil, Boris; Wald, Lawrence L

    2013-01-01

    Originally proposed as a method to increase sensitivity by extending the locally high-sensitivity of small surface coil elements to larger areas, the term parallel imaging now includes the use of array coils to perform image encoding. This methodology has impacted clinical imaging to the point where many examinations are performed with an array comprising multiple smaller surface coil elements as the detector of the MR signal. This article reviews the theoretical and experimental basis for the trend towards higher channel counts relying on insights gained from modeling and experimental studies as well as the theoretical analysis of the so-called “ultimate” SNR and g-factor. We also review the methods for optimally combining array data and changes in RF methodology needed to construct massively parallel MRI detector arrays and show some examples of state-of-the-art for highly accelerated imaging with the resulting highly parallel arrays. PMID:23453758

  11. Parallel Adaptive Mesh Refinement Library

    NASA Technical Reports Server (NTRS)

    Mac-Neice, Peter; Olson, Kevin

    2005-01-01

    Parallel Adaptive Mesh Refinement Library (PARAMESH) is a package of Fortran 90 subroutines designed to provide a computer programmer with an easy route to extension of (1) a previously written serial code that uses a logically Cartesian structured mesh into (2) a parallel code with adaptive mesh refinement (AMR). Alternatively, in its simplest use, and with minimal effort, PARAMESH can operate as a domain-decomposition tool for users who want to parallelize their serial codes but who do not wish to utilize adaptivity. The package builds a hierarchy of sub-grids to cover the computational domain of a given application program, with spatial resolution varying to satisfy the demands of the application. The sub-grid blocks form the nodes of a tree data structure (a quad-tree in two or an oct-tree in three dimensions). Each grid block has a logically Cartesian mesh. The package supports one-, two- and three-dimensional models.

  12. Hybrid parallel programming with MPI and Unified Parallel C.

    SciTech Connect

    Dinan, J.; Balaji, P.; Lusk, E.; Sadayappan, P.; Thakur, R.; Mathematics and Computer Science; The Ohio State Univ.

    2010-01-01

    The Message Passing Interface (MPI) is one of the most widely used programming models for parallel computing. However, the amount of memory available to an MPI process is limited by the amount of local memory within a compute node. Partitioned Global Address Space (PGAS) models such as Unified Parallel C (UPC) are growing in popularity because of their ability to provide a shared global address space that spans the memories of multiple compute nodes. However, taking advantage of UPC can require a large recoding effort for existing parallel applications. In this paper, we explore a new hybrid parallel programming model that combines MPI and UPC. This model allows MPI programmers incremental access to a greater amount of memory, enabling memory-constrained MPI codes to process larger data sets. In addition, the hybrid model offers UPC programmers an opportunity to create static UPC groups that are connected over MPI. As we demonstrate, the use of such groups can significantly improve the scalability of locality-constrained UPC codes. This paper presents a detailed description of the hybrid model and demonstrates its effectiveness in two applications: a random access benchmark and the Barnes-Hut cosmological simulation. Experimental results indicate that the hybrid model can greatly enhance performance; using hybrid UPC groups that span two cluster nodes, RA performance increases by a factor of 1.33 and using groups that span four cluster nodes, Barnes-Hut experiences a twofold speedup at the expense of a 2% increase in code size.

  13. Medipix2 parallel readout system

    NASA Astrophysics Data System (ADS)

    Fanti, V.; Marzeddu, R.; Randaccio, P.

    2003-08-01

    A fast parallel readout system based on a PCI board has been developed in the framework of the Medipix collaboration. The readout electronics consists of two boards: the motherboard directly interfacing the Medipix2 chip, and the PCI board with digital I/O ports 32 bits wide. The device driver and readout software have been developed at low level in Assembler to allow fast data transfer and image reconstruction. The parallel readout permits a transfer rate up to 64 Mbytes/s. http://medipix.web.cern ch/MEDIPIX/

  14. Gang scheduling a parallel machine

    SciTech Connect

    Gorda, B.C.; Brooks, E.D. III.

    1991-03-01

    Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processors. User program and their gangs of processors are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantums are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory. 2 refs., 1 fig.

  15. Gang scheduling a parallel machine

    SciTech Connect

    Gorda, B.C.; Brooks, E.D. III.

    1991-12-01

    Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processes. User programs and their gangs of processes are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantum are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory.

  16. The Complexity of Parallel Algorithms,

    DTIC Science & Technology

    1985-11-01

    Much of this work was done in collaboration with my advisor, Ernst Mayr . He was also supported in part by ONR contract N00014-85-C-0731. F ’. Table...Helinbold and Mayr in their algorithn to compute an optimal two processor schedule [HM2]. One of the promising developments in parallel algorithms is that...lei can be solved by it fast parallel algorithmmmi if the nmlmmmibers are smiall. llehmibold and Mayr JIlM I] have slhowm that. if Ole job timies are

  17. File concepts for parallel I/O

    NASA Technical Reports Server (NTRS)

    Crockett, Thomas W.

    1989-01-01

    The subject of input/output (I/O) was often neglected in the design of parallel computer systems, although for many problems I/O rates will limit the speedup attainable. The I/O problem is addressed by considering the role of files in parallel systems. The notion of parallel files is introduced. Parallel files provide for concurrent access by multiple processes, and utilize parallelism in the I/O system to improve performance. Parallel files can also be used conventionally by sequential programs. A set of standard parallel file organizations is proposed, organizations are suggested, using multiple storage devices. Problem areas are also identified and discussed.

  18. Matpar: Parallel Extensions for MATLAB

    NASA Technical Reports Server (NTRS)

    Springer, P. L.

    1998-01-01

    Matpar is a set of client/server software that allows a MATLAB user to take advantage of a parallel computer for very large problems. The user can replace calls to certain built-in MATLAB functions with calls to Matpar functions.

  19. Parallel, Distributed Scripting with Python

    SciTech Connect

    Miller, P J

    2002-05-24

    Parallel computers used to be, for the most part, one-of-a-kind systems which were extremely difficult to program portably. With SMP architectures, the advent of the POSIX thread API and OpenMP gave developers ways to portably exploit on-the-box shared memory parallelism. Since these architectures didn't scale cost-effectively, distributed memory clusters were developed. The associated MPI message passing libraries gave these systems a portable paradigm too. Having programmers effectively use this paradigm is a somewhat different question. Distributed data has to be explicitly transported via the messaging system in order for it to be useful. In high level languages, the MPI library gives access to data distribution routines in C, C++, and FORTRAN. But we need more than that. Many reasonable and common tasks are best done in (or as extensions to) scripting languages. Consider sysadm tools such as password crackers, file purgers, etc ... These are simple to write in a scripting language such as Python (an open source, portable, and freely available interpreter). But these tasks beg to be done in parallel. Consider the a password checker that checks an encrypted password against a 25,000 word dictionary. This can take around 10 seconds in Python (6 seconds in C). It is trivial to parallelize if you can distribute the information and co-ordinate the work.

  20. Fast, Massively Parallel Data Processors

    NASA Technical Reports Server (NTRS)

    Heaton, Robert A.; Blevins, Donald W.; Davis, ED

    1994-01-01

    Proposed fast, massively parallel data processor contains 8x16 array of processing elements with efficient interconnection scheme and options for flexible local control. Processing elements communicate with each other on "X" interconnection grid with external memory via high-capacity input/output bus. This approach to conditional operation nearly doubles speed of various arithmetic operations.

  1. Optical Interferometric Parallel Data Processor

    NASA Technical Reports Server (NTRS)

    Breckinridge, J. B.

    1987-01-01

    Image data processed faster than in present electronic systems. Optical parallel-processing system effectively calculates two-dimensional Fourier transforms in time required by light to travel from plane 1 to plane 8. Coherence interferometer at plane 4 splits light into parts that form double image at plane 6 if projection screen placed there.

  2. Tutorial: Parallel Simulation on Supercomputers

    SciTech Connect

    Perumalla, Kalyan S

    2012-01-01

    This tutorial introduces typical hardware and software characteristics of extant and emerging supercomputing platforms, and presents issues and solutions in executing large-scale parallel discrete event simulation scenarios on such high performance computing systems. Covered topics include synchronization, model organization, example applications, and observed performance from illustrative large-scale runs.

  3. The physics of parallel machines

    NASA Technical Reports Server (NTRS)

    Chan, Tony F.

    1988-01-01

    The idea is considered that architectures for massively parallel computers must be designed to go beyond supporting a particular class of algorithms to supporting the underlying physical processes being modelled. Physical processes modelled by partial differential equations (PDEs) are discussed. Also discussed is the idea that an efficient architecture must go beyond nearest neighbor mesh interconnections and support global and hierarchical communications.

  4. Parallel distributed computing using Python

    NASA Astrophysics Data System (ADS)

    Dalcin, Lisandro D.; Paz, Rodrigo R.; Kler, Pablo A.; Cosimo, Alejandro

    2011-09-01

    This work presents two software components aimed to relieve the costs of accessing high-performance parallel computing resources within a Python programming environment: MPI for Python and PETSc for Python. MPI for Python is a general-purpose Python package that provides bindings for the Message Passing Interface (MPI) standard using any back-end MPI implementation. Its facilities allow parallel Python programs to easily exploit multiple processors using the message passing paradigm. PETSc for Python provides access to the Portable, Extensible Toolkit for Scientific Computation (PETSc) libraries. Its facilities allow sequential and parallel Python applications to exploit state of the art algorithms and data structures readily available in PETSc for the solution of large-scale problems in science and engineering. MPI for Python and PETSc for Python are fully integrated to PETSc-FEM, an MPI and PETSc based parallel, multiphysics, finite elements code developed at CIMEC laboratory. This software infrastructure supports research activities related to simulation of fluid flows with applications ranging from the design of microfluidic devices for biochemical analysis to modeling of large-scale stream/aquifer interactions.

  5. PALM: a Parallel Dynamic Coupler

    NASA Astrophysics Data System (ADS)

    Thevenin, A.; Morel, T.

    2008-12-01

    In order to efficiently represent complex systems, numerical modeling has to rely on many physical models at a time: an ocean model coupled with an atmospheric model is at the basis of climate modeling. The continuity of the solution is granted only if these models can constantly exchange information. PALM is a coupler allowing the concurrent execution and the intercommunication of programs not having been especially designed for that. With PALM, the dynamic coupling approach is introduced: a coupled component can be launched and can release computers' resources upon termination at any moment during the simulation. In order to exploit as much as possible computers' possibilities, the PALM coupler handles two levels of parallelism. The first level concerns the components themselves. While managing the resources, PALM allocates the number of processes which are necessary to any coupled component. These models can be parallel programs based on domain decomposition with MPI or applications multithreaded with OpenMP. The second level of parallelism is a task parallelism: one can define a coupling algorithm allowing two or more programs to be executed in parallel. PALM applications are implemented via a Graphical User Interface called PrePALM. In this GUI, the programmer initially defines the coupling algorithm then he describes the actual communications between the models. PALM offers a very high flexibility for testing different coupling techniques and for reaching the best load balance in a high performance computer. The transformation of computational independent code is almost straightforward. The other qualities of PALM are its easy set-up, its flexibility, its performances, the simple updates and evolutions of the coupled application and the many side services and functions that it offers.

  6. Task parallelism and high-performance languages

    SciTech Connect

    Foster, I.

    1996-03-01

    The definition of High Performance Fortran (HPF) is a significant event in the maturation of parallel computing: it represents the first parallel language that has gained widespread support from vendors and users. The subject of this paper is to incorporate support for task parallelism. The term task parallelism refers to the explicit creation of multiple threads of control, or tasks, which synchronize and communicate under programmer control. Task and data parallelism are complementary rather than competing programming models. While task parallelism is more general and can be used to implement algorithms that are not amenable to data-parallel solutions, many problems can benefit from a mixed approach, with for example a task-parallel coordination layer integrating multiple data-parallel computations. Other problems admit to both data- and task-parallel solutions, with the better solution depending on machine characteristics, compiler performance, or personal taste. For these reasons, we believe that a general-purpose high-performance language should integrate both task- and data-parallel constructs. The challenge is to do so in a way that provides the expressivity needed for applications, while preserving the flexibility and portability of a high-level language. In this paper, we examine and illustrate the considerations that motivate the use of task parallelism. We also describe one particular approach to task parallelism in Fortran, namely the Fortran M extensions. Finally, we contrast Fortran M with other proposed approaches and discuss the implications of this work for task parallelism and high-performance languages.

  7. A generalized parallel replica dynamics

    NASA Astrophysics Data System (ADS)

    Binder, Andrew; Lelièvre, Tony; Simpson, Gideon

    2015-03-01

    Metastability is a common obstacle to performing long molecular dynamics simulations. Many numerical methods have been proposed to overcome it. One method is parallel replica dynamics, which relies on the rapid convergence of the underlying stochastic process to a quasi-stationary distribution. Two requirements for applying parallel replica dynamics are knowledge of the time scale on which the process converges to the quasi-stationary distribution and a mechanism for generating samples from this distribution. By combining a Fleming-Viot particle system with convergence diagnostics to simultaneously identify when the process converges while also generating samples, we can address both points. This variation on the algorithm is illustrated with various numerical examples, including those with entropic barriers and the 2D Lennard-Jones cluster of seven atoms.

  8. Merlin - Massively parallel heterogeneous computing

    NASA Technical Reports Server (NTRS)

    Wittie, Larry; Maples, Creve

    1989-01-01

    Hardware and software for Merlin, a new kind of massively parallel computing system, are described. Eight computers are linked as a 300-MIPS prototype to develop system software for a larger Merlin network with 16 to 64 nodes, totaling 600 to 3000 MIPS. These working prototypes help refine a mapped reflective memory technique that offers a new, very general way of linking many types of computer to form supercomputers. Processors share data selectively and rapidly on a word-by-word basis. Fast firmware virtual circuits are reconfigured to match topological needs of individual application programs. Merlin's low-latency memory-sharing interfaces solve many problems in the design of high-performance computing systems. The Merlin prototypes are intended to run parallel programs for scientific applications and to determine hardware and software needs for a future Teraflops Merlin network.

  9. Parallel supercomputing with commodity components

    SciTech Connect

    Warren, M.S.; Goda, M.P.; Becker, D.J.

    1997-09-01

    We have implemented a parallel computer architecture based entirely upon commodity personal computer components. Using 16 Intel Pentium Pro microprocessors and switched fast ethernet as a communication fabric, we have obtained sustained performance on scientific applications in excess of one Gigaflop. During one production astrophysics treecode simulation, we performed 1.2 x 10{sup 15} floating point operations (1.2 Petaflops) over a three week period, with one phase of that simulation running continuously for two weeks without interruption. We report on a variety of disk, memory and network benchmarks. We also present results from the NAS parallel benchmark suite, which indicate that this architecture is competitive with current commercial architectures. In addition, we describe some software written to support efficient message passing, as well as a Linux device driver interface to the Pentium hardware performance monitoring registers.

  10. ASP: a parallel computing technology

    NASA Astrophysics Data System (ADS)

    Lea, R. M.

    1990-09-01

    ASP modules constitute the basis of a parallel computing technology platform for the rapid development of a broad range of numeric and symbolic information processing systems. Based on off-the-shelf general-purpose hardware and software modules ASP technology is intended to increase productivity in the development (and competitiveness in the marketing) of cost-effective low-MIMD/high-SIMD Massively Parallel Processor (MPPs). The paper discusses ASP module philosophy and demonstrates how ASP modules can satisfy the market algorithmic architectural and engineering requirements of such MPPs. In particular two specific ASP modules based on VLSI and WSI technologies are studied as case examples of ASP technology the latter reporting 1 TOPS/fl3 1 GOPS/W and 1 MOPS/$ as ball-park figures-of-merit of cost-effectiveness.

  11. Parallel processing spacecraft communication system

    NASA Technical Reports Server (NTRS)

    Bolotin, Gary S. (Inventor); Donaldson, James A. (Inventor); Luong, Huy H. (Inventor); Wood, Steven H. (Inventor)

    1998-01-01

    An uplink controlling assembly speeds data processing using a special parallel codeblock technique. A correct start sequence initiates processing of a frame. Two possible start sequences can be used; and the one which is used determines whether data polarity is inverted or non-inverted. Processing continues until uncorrectable errors are found. The frame ends by intentionally sending a block with an uncorrectable error. Each of the codeblocks in the frame has a channel ID. Each channel ID can be separately processed in parallel. This obviates the problem of waiting for error correction processing. If that channel number is zero, however, it indicates that the frame of data represents a critical command only. That data is handled in a special way, independent of the software. Otherwise, the processed data further handled using special double buffering techniques to avoid problems from overrun. When overrun does occur, the system takes action to lose only the oldest data.

  12. A generalized parallel replica dynamics

    SciTech Connect

    Binder, Andrew; Lelièvre, Tony; Simpson, Gideon

    2015-03-01

    Metastability is a common obstacle to performing long molecular dynamics simulations. Many numerical methods have been proposed to overcome it. One method is parallel replica dynamics, which relies on the rapid convergence of the underlying stochastic process to a quasi-stationary distribution. Two requirements for applying parallel replica dynamics are knowledge of the time scale on which the process converges to the quasi-stationary distribution and a mechanism for generating samples from this distribution. By combining a Fleming–Viot particle system with convergence diagnostics to simultaneously identify when the process converges while also generating samples, we can address both points. This variation on the algorithm is illustrated with various numerical examples, including those with entropic barriers and the 2D Lennard-Jones cluster of seven atoms.

  13. Parallel supercomputing with commodity components

    NASA Technical Reports Server (NTRS)

    Warren, M. S.; Goda, M. P.; Becker, D. J.

    1997-01-01

    We have implemented a parallel computer architecture based entirely upon commodity personal computer components. Using 16 Intel Pentium Pro microprocessors and switched fast ethernet as a communication fabric, we have obtained sustained performance on scientific applications in excess of one Gigaflop. During one production astrophysics treecode simulation, we performed 1.2 x 10(sup 15) floating point operations (1.2 Petaflops) over a three week period, with one phase of that simulation running continuously for two weeks without interruption. We report on a variety of disk, memory and network benchmarks. We also present results from the NAS parallel benchmark suite, which indicate that this architecture is competitive with current commercial architectures. In addition, we describe some software written to support efficient message passing, as well as a Linux device driver interface to the Pentium hardware performance monitoring registers.

  14. Parallel multiplex laser feedback interferometry

    SciTech Connect

    Zhang, Song; Tan, Yidong; Zhang, Shulian

    2013-12-15

    We present a parallel multiplex laser feedback interferometer based on spatial multiplexing which avoids the signal crosstalk in the former feedback interferometer. The interferometer outputs two close parallel laser beams, whose frequencies are shifted by two acousto-optic modulators by 2Ω simultaneously. A static reference mirror is inserted into one of the optical paths as the reference optical path. The other beam impinges on the target as the measurement optical path. Phase variations of the two feedback laser beams are simultaneously measured through heterodyne demodulation with two different detectors. Their subtraction accurately reflects the target displacement. Under typical room conditions, experimental results show a resolution of 1.6 nm and accuracy of 7.8 nm within the range of 100 μm.

  15. Parallelism in Manipulator Dynamics. Revision.

    DTIC Science & Technology

    1983-12-01

    excessive, and a VLSI implementation architecutre is suggested. We indicate possible appli- cations to incorporating dynamical considerations into...Inverse Dynamics problem. It investigates the high degree of parallelism inherent in the computations , and presents two "mathematically exact" formulations...and a 3 b Cases ............. ... 109 5 .9-- i 0. OVERVIEW The Inverse Dynamics problem consists (loosely) of computing the motor torques necessary to

  16. Parallel Symmetric Eigenvalue Problem Solvers

    DTIC Science & Technology

    2015-05-01

    graduate school. Debugging somebody else’s MPI code is an immensely frustrating experience, but he would regularly stay late at the oce to assist me...cessfully. In addition, I will describe the parallel kernels required by my code . 5 The next sections will describe my Fortran-based implementations of...Sandia’s publicly available Trace- Min code . Each of the methods has its own unique advantages and disadvantages, summarized in table 3.1. In short, I

  17. Parallel Algorithms for Computer Vision.

    DTIC Science & Technology

    1987-01-01

    73 755 P fiu.LEL ALORITHMS FOR CO PUTER VISIO (U) /MASSACHUSETTS INST OF TECH CRMORIDGE T P00010 ET AL.JAN 8? ETL-0456 DACA7-05-C-8IIO m 7E F/0 1...regularization principles, such as edge detection, stereo , motion, surface interpolation and shape from shading. The basic members of class I are convolution...them in collabo- ration with Thinking Machines Corporation): * Parallel convolution * Zero-crossing detection * Stereo -matching * Surface reconstruction

  18. Lightweight Specifications for Parallel Correctness

    DTIC Science & Technology

    2012-12-05

    this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 1215 Jefferson Davis Highway, Suite 1204... George Necula Professor David Wessel Fall 2012 1 Abstract Lightweight Specifications for Parallel Correctness by Jacob Samuels Burnim Doctor of Philosophy...enthusiasm and endless flow of ideas, and for his keen research sense. I would also like to thank George Necula for chairing my qualifying exam committee and

  19. National Combustion Code: Parallel Performance

    NASA Technical Reports Server (NTRS)

    Babrauckas, Theresa

    2001-01-01

    This report discusses the National Combustion Code (NCC). The NCC is an integrated system of codes for the design and analysis of combustion systems. The advanced features of the NCC meet designers' requirements for model accuracy and turn-around time. The fundamental features at the inception of the NCC were parallel processing and unstructured mesh. The design and performance of the NCC are discussed.

  20. Parallel Algorithms for Computer Vision.

    DTIC Science & Technology

    1989-01-01

    demonstrated the Vision Machine system processing images and recognizing objects through the inte- gration of several visual cues. The first version of the...achievements. n 2.1 The Vision Machine The overall organization of tie Vision Machine systeliis ased. o parallel processing of tie images by independent...smoothed and made dense by exploiting known constraints within each process (for example., that disparity is smooth). This is the stage of approximation

  1. Parallel strategies for SAR processing

    NASA Astrophysics Data System (ADS)

    Segoviano, Jesus A.

    2004-12-01

    This article proposes a series of strategies for improving the computer process of the Synthetic Aperture Radar (SAR) signal treatment, following the three usual lines of action to speed up the execution of any computer program. On the one hand, it is studied the optimization of both, the data structures and the application architecture used on it. On the other hand it is considered a hardware improvement. For the former, they are studied both, the usually employed SAR process data structures, proposing the use of parallel ones and the way the parallelization of the algorithms employed on the process is implemented. Besides, the parallel application architecture classifies processes between fine/coarse grain. These are assigned to individual processors or separated in a division among processors, all of them in their corresponding architectures. For the latter, it is studied the hardware employed on the computer parallel process used in the SAR handling. The improvement here refers to several kinds of platforms in which the SAR process is implemented, shared memory multicomputers, and distributed memory multiprocessors. A comparison between them gives us some guidelines to follow in order to get a maximum throughput with a minimum latency and a maximum effectiveness with a minimum cost, all together with a limited complexness. It is concluded and described, that the approach consisting of the processing of the algorithms in a GNU/Linux environment, together with a Beowulf cluster platform offers, under certain conditions, the best compromise between performance and cost, and promises the major development in the future for the Synthetic Aperture Radar computer power thirsty applications in the next years.

  2. Parallel Power Grid Simulation Toolkit

    SciTech Connect

    Smith, Steve; Kelley, Brian; Banks, Lawrence; Top, Philip; Woodward, Carol

    2015-09-14

    ParGrid is a 'wrapper' that integrates a coupled Power Grid Simulation toolkit consisting of a library to manage the synchronization and communication of independent simulations. The included library code in ParGid, named FSKIT, is intended to support the coupling multiple continuous and discrete even parallel simulations. The code is designed using modern object oriented C++ methods utilizing C++11 and current Boost libraries to ensure compatibility with multiple operating systems and environments.

  3. Parallel processing of genomics data

    NASA Astrophysics Data System (ADS)

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  4. Parallelism in integrated fluidic circuits

    NASA Astrophysics Data System (ADS)

    Bousse, Luc J.; Kopf-Sill, Anne R.; Parce, J. W.

    1998-04-01

    Many research groups around the world are working on integrated microfluidics. The goal of these projects is to automate and integrate the handling of liquid samples and reagents for measurement and assay procedures in chemistry and biology. Ultimately, it is hoped that this will lead to a revolution in chemical and biological procedures similar to that caused in electronics by the invention of the integrated circuit. The optimal size scale of channels for liquid flow is determined by basic constraints to be somewhere between 10 and 100 micrometers . In larger channels, mixing by diffusion takes too long; in smaller channels, the number of molecules present is so low it makes detection difficult. At Caliper, we are making fluidic systems in glass chips with channels in this size range, based on electroosmotic flow, and fluorescence detection. One application of this technology is rapid assays for drug screening, such as enzyme assays and binding assays. A further challenge in this area is to perform multiple functions on a chip in parallel, without a large increase in the number of inputs and outputs. A first step in this direction is a fluidic serial-to-parallel converter. Fluidic circuits will be shown with the ability to distribute an incoming serial sample stream to multiple parallel channels.

  5. Highly parallel sparse Cholesky factorization

    NASA Technical Reports Server (NTRS)

    Gilbert, John R.; Schreiber, Robert

    1990-01-01

    Several fine grained parallel algorithms were developed and compared to compute the Cholesky factorization of a sparse matrix. The experimental implementations are on the Connection Machine, a distributed memory SIMD machine whose programming model conceptually supplies one processor per data element. In contrast to special purpose algorithms in which the matrix structure conforms to the connection structure of the machine, the focus is on matrices with arbitrary sparsity structure. The most promising algorithm is one whose inner loop performs several dense factorizations simultaneously on a 2-D grid of processors. Virtually any massively parallel dense factorization algorithm can be used as the key subroutine. The sparse code attains execution rates comparable to those of the dense subroutine. Although at present architectural limitations prevent the dense factorization from realizing its potential efficiency, it is concluded that a regular data parallel architecture can be used efficiently to solve arbitrarily structured sparse problems. A performance model is also presented and it is used to analyze the algorithms.

  6. Parallel Environment for Quantum Computing

    NASA Astrophysics Data System (ADS)

    Tabakin, Frank; Diaz, Bruno Julia

    2009-03-01

    To facilitate numerical study of noise and decoherence in QC algorithms,and of the efficacy of error correction schemes, we have developed a Fortran 90 quantum computer simulator with parallel processing capabilities. It permits rapid evaluation of quantum algorithms for a large number of qubits and for various ``noise'' scenarios. State vectors are distributed over many processors, to employ a large number of qubits. Parallel processing is implemented by the Message-Passing Interface protocol. A description of how to spread the wave function components over many processors, along with how to efficiently describe the action of general one- and two-qubit operators on these state vectors will be delineated.Grover's search and Shor's factoring algorithms with noise will be discussed as examples. A major feature of this work is that concurrent versions of the algorithms can be evaluated with each version subject to diverse noise effects, corresponding to solving a stochastic Schrodinger equation. The density matrix for the ensemble of such noise cases is constructed using parallel distribution methods to evaluate its associated entropy. Applications of this powerful tool is made to delineate the stability and correction of QC processes using Hamiltonian based dynamics.

  7. Parallel Markov chain Monte Carlo simulations.

    PubMed

    Ren, Ruichao; Orkoulas, G

    2007-06-07

    With strict detailed balance, parallel Monte Carlo simulation through domain decomposition cannot be validated with conventional Markov chain theory, which describes an intrinsically serial stochastic process. In this work, the parallel version of Markov chain theory and its role in accelerating Monte Carlo simulations via cluster computing is explored. It is shown that sequential updating is the key to improving efficiency in parallel simulations through domain decomposition. A parallel scheme is proposed to reduce interprocessor communication or synchronization, which slows down parallel simulation with increasing number of processors. Parallel simulation results for the two-dimensional lattice gas model show substantial reduction of simulation time for systems of moderate and large size.

  8. Parallel Markov chain Monte Carlo simulations

    NASA Astrophysics Data System (ADS)

    Ren, Ruichao; Orkoulas, G.

    2007-06-01

    With strict detailed balance, parallel Monte Carlo simulation through domain decomposition cannot be validated with conventional Markov chain theory, which describes an intrinsically serial stochastic process. In this work, the parallel version of Markov chain theory and its role in accelerating Monte Carlo simulations via cluster computing is explored. It is shown that sequential updating is the key to improving efficiency in parallel simulations through domain decomposition. A parallel scheme is proposed to reduce interprocessor communication or synchronization, which slows down parallel simulation with increasing number of processors. Parallel simulation results for the two-dimensional lattice gas model show substantial reduction of simulation time for systems of moderate and large size.

  9. Parallelizing alternating direction implicit solver on GPUs

    Technology Transfer Automated Retrieval System (TEKTRAN)

    We present a parallel Alternating Direction Implicit (ADI) solver on GPUs. Our implementation significantly improves existing implementations in two aspects. First, we address the scalability issue of existing Parallel Cyclic Reduction (PCR) implementations by eliminating their hardware resource con...

  10. Implementing clips on a parallel computer

    NASA Technical Reports Server (NTRS)

    Riley, Gary

    1987-01-01

    The C language integrated production system (CLIPS) is a forward chaining rule based language to provide training and delivery for expert systems. Conceptually, rule based languages have great potential for benefiting from the inherent parallelism of the algorithms that they employ. During each cycle of execution, a knowledge base of information is compared against a set of rules to determine if any rules are applicable. Parallelism also can be employed for use with multiple cooperating expert systems. To investigate the potential benefits of using a parallel computer to speed up the comparison of facts to rules in expert systems, a parallel version of CLIPS was developed for the FLEX/32, a large grain parallel computer. The FLEX implementation takes a macroscopic approach in achieving parallelism by splitting whole sets of rules among several processors rather than by splitting the components of an individual rule among processors. The parallel CLIPS prototype demonstrates the potential advantages of integrating expert system tools with parallel computers.

  11. Parallel molecular dynamics: Communication requirements for massively parallel machines

    NASA Astrophysics Data System (ADS)

    Taylor, Valerie E.; Stevens, Rick L.; Arnold, Kathryn E.

    1995-05-01

    Molecular mechanics and dynamics are becoming widely used to perform simulations of molecular systems from large-scale computations of materials to the design and modeling of drug compounds. In this paper we address two major issues: a good decomposition method that can take advantage of future massively parallel processing systems for modest-sized problems in the range of 50,000 atoms and the communication requirements needed to achieve 30 to 40% efficiency on MPPs. We analyzed a scalable benchmark molecular dynamics program executing on the Intel Touchstone Deleta parallelized with an interaction decomposition method. Using a validated analytical performance model of the code, we determined that for an MPP with a four-dimensional mesh topology and 400 MHz processors the communication startup time must be at most 30 clock cycles and the network bandwidth must be at least 2.3 GB/s. This configuration results in 30 to 40% efficiency of the MPP for a problem with 50,000 atoms executing on 50,000 processors.

  12. Force user's manual: A portable, parallel FORTRAN

    NASA Technical Reports Server (NTRS)

    Jordan, Harry F.; Benten, Muhammad S.; Arenstorf, Norbert S.; Ramanan, Aruna V.

    1990-01-01

    The use of Force, a parallel, portable FORTRAN on shared memory parallel computers is described. Force simplifies writing code for parallel computers and, once the parallel code is written, it is easily ported to computers on which Force is installed. Although Force is nearly the same for all computers, specific details are included for the Cray-2, Cray-YMP, Convex 220, Flex/32, Encore, Sequent, Alliant computers on which it is installed.

  13. Automatic Multilevel Parallelization Using OpenMP

    NASA Technical Reports Server (NTRS)

    Jin, Hao-Qiang; Jost, Gabriele; Yan, Jerry; Ayguade, Eduard; Gonzalez, Marc; Martorell, Xavier; Biegel, Bryan (Technical Monitor)

    2002-01-01

    In this paper we describe the extension of the CAPO (CAPtools (Computer Aided Parallelization Toolkit) OpenMP) parallelization support tool to support multilevel parallelism based on OpenMP directives. CAPO generates OpenMP directives with extensions supported by the NanosCompiler to allow for directive nesting and definition of thread groups. We report some results for several benchmark codes and one full application that have been parallelized using our system.

  14. The electron signature of parallel electric fields

    NASA Astrophysics Data System (ADS)

    Burch, J. L.; Gurgiolo, C.; Menietti, J. D.

    1990-12-01

    Dynamics Explorer I High-Altitude Plasma Instrument electron data are presented. The electron distribution functions have characteristics expected of a region of parallel electric fields. The data are consistent with previous test-particle simulations for observations within parallel electric field regions which indicate that typical hole, bump, and loss-cone electron distributions, which contain evidence for parallel potential differences both above and below the point of observation, are not expected to occur in regions containing actual parallel electric fields.

  15. Debugging Parallel Programs with Instant Replay.

    DTIC Science & Technology

    1986-09-01

    produce the same results. In this paper we present a general solution for reproducing the execution behavior of parallel programs, termed Instant Replay...Instant Replay on the BBN Butterfly Parallel Processor, and discuss how it can be incorporated into the debugging cycle for parallel programs. This...program often do not produce the same results. In this paper we present a general solution for reproducing the execution behavior of parallel

  16. Parallel machine architecture and compiler design facilities

    NASA Technical Reports Server (NTRS)

    Kuck, David J.; Yew, Pen-Chung; Padua, David; Sameh, Ahmed; Veidenbaum, Alex

    1990-01-01

    The objective is to provide an integrated simulation environment for studying and evaluating various issues in designing parallel systems, including machine architectures, parallelizing compiler techniques, and parallel algorithms. The status of Delta project (which objective is to provide a facility to allow rapid prototyping of parallelized compilers that can target toward different machine architectures) is summarized. Included are the surveys of the program manipulation tools developed, the environmental software supporting Delta, and the compiler research projects in which Delta has played a role.

  17. Global Arrays Parallel Programming Toolkit

    SciTech Connect

    Nieplocha, Jaroslaw; Krishnan, Manoj Kumar; Palmer, Bruce J.; Tipparaju, Vinod; Harrison, Robert J.; Chavarría-Miranda, Daniel

    2011-01-01

    The two predominant classes of programming models for parallel computing are distributed memory and shared memory. Both shared memory and distributed memory models have advantages and shortcomings. Shared memory model is much easier to use but it ignores data locality/placement. Given the hierarchical nature of the memory subsystems in modern computers this characteristic can have a negative impact on performance and scalability. Careful code restructuring to increase data reuse and replacing fine grain load/stores with block access to shared data can address the problem and yield performance for shared memory that is competitive with message-passing. However, this performance comes at the cost of compromising the ease of use that the shared memory model advertises. Distributed memory models, such as message-passing or one-sided communication, offer performance and scalability but they are difficult to program. The Global Arrays toolkit attempts to offer the best features of both models. It implements a shared-memory programming model in which data locality is managed by the programmer. This management is achieved by calls to functions that transfer data between a global address space (a distributed array) and local storage. In this respect, the GA model has similarities to the distributed shared-memory models that provide an explicit acquire/release protocol. However, the GA model acknowledges that remote data is slower to access than local data and allows data locality to be specified by the programmer and hence managed. GA is related to the global address space languages such as UPC, Titanium, and, to a lesser extent, Co-Array Fortran. In addition, by providing a set of data-parallel operations, GA is also related to data-parallel languages such as HPF, ZPL, and Data Parallel C. However, the Global Array programming model is implemented as a library that works with most languages used for technical computing and does not rely on compiler technology for achieving

  18. High Performance Parallel Computational Nanotechnology

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Craw, James M. (Technical Monitor)

    1995-01-01

    At a recent press conference, NASA Administrator Dan Goldin encouraged NASA Ames Research Center to take a lead role in promoting research and development of advanced, high-performance computer technology, including nanotechnology. Manufacturers of leading-edge microprocessors currently perform large-scale simulations in the design and verification of semiconductor devices and microprocessors. Recently, the need for this intensive simulation and modeling analysis has greatly increased, due in part to the ever-increasing complexity of these devices, as well as the lessons of experiences such as the Pentium fiasco. Simulation, modeling, testing, and validation will be even more important for designing molecular computers because of the complex specification of millions of atoms, thousands of assembly steps, as well as the simulation and modeling needed to ensure reliable, robust and efficient fabrication of the molecular devices. The software for this capacity does not exist today, but it can be extrapolated from the software currently used in molecular modeling for other applications: semi-empirical methods, ab initio methods, self-consistent field methods, Hartree-Fock methods, molecular mechanics; and simulation methods for diamondoid structures. In as much as it seems clear that the application of such methods in nanotechnology will require powerful, highly powerful systems, this talk will discuss techniques and issues for performing these types of computations on parallel systems. We will describe system design issues (memory, I/O, mass storage, operating system requirements, special user interface issues, interconnects, bandwidths, and programming languages) involved in parallel methods for scalable classical, semiclassical, quantum, molecular mechanics, and continuum models; molecular nanotechnology computer-aided designs (NanoCAD) techniques; visualization using virtual reality techniques of structural models and assembly sequences; software required to

  19. Parallel Computing Using Web Servers and "Servlets".

    ERIC Educational Resources Information Center

    Lo, Alfred; Bloor, Chris; Choi, Y. K.

    2000-01-01

    Describes parallel computing and presents inexpensive ways to implement a virtual parallel computer with multiple Web servers. Highlights include performance measurement of parallel systems; models for using Java and intranet technology including single server, multiple clients and multiple servers, single client; and a comparison of CGI (common…

  20. Identifying, Quantifying, Extracting and Enhancing Implicit Parallelism

    ERIC Educational Resources Information Center

    Agarwal, Mayank

    2009-01-01

    The shift of the microprocessor industry towards multicore architectures has placed a huge burden on the programmers by requiring explicit parallelization for performance. Implicit Parallelization is an alternative that could ease the burden on programmers by parallelizing applications "under the covers" while maintaining sequential semantics…

  1. Parallel Processing at the High School Level.

    ERIC Educational Resources Information Center

    Sheary, Kathryn Anne

    This study investigated the ability of high school students to cognitively understand and implement parallel processing. Data indicates that most parallel processing is being taught at the university level. Instructional modules on C, Linux, and the parallel processing language, P4, were designed to show that high school students are highly…

  2. Coordination in serial-parallel image processing

    NASA Astrophysics Data System (ADS)

    Wójcik, Waldemar; Dubovoi, Vladymyr M.; Duda, Marina E.; Romaniuk, Ryszard S.; Yesmakhanova, Laura; Kozbakova, Ainur

    2015-12-01

    Serial-parallel systems used to convert the image. The control of their work results with the need to solve coordination problem. The paper summarizes the model of coordination of resource allocation in relation to the task of synchronizing parallel processes; the genetic algorithm of coordination developed, its adequacy verified in relation to the process of parallel image processing.

  3. Scalable Parallel Algebraic Multigrid Solvers

    SciTech Connect

    Bank, R; Lu, S; Tong, C; Vassilevski, P

    2005-03-23

    The authors propose a parallel algebraic multilevel algorithm (AMG), which has the novel feature that the subproblem residing in each processor is defined over the entire partition domain, although the vast majority of unknowns for each subproblem are associated with the partition owned by the corresponding processor. This feature ensures that a global coarse description of the problem is contained within each of the subproblems. The advantages of this approach are that interprocessor communication is minimized in the solution process while an optimal order of convergence rate is preserved; and the speed of local subproblem solvers can be maximized using the best existing sequential algebraic solvers.

  4. Parallel Assembly of LIGA Components

    SciTech Connect

    Christenson, T.R.; Feddema, J.T.

    1999-03-04

    In this paper, a prototype robotic workcell for the parallel assembly of LIGA components is described. A Cartesian robot is used to press 386 and 485 micron diameter pins into a LIGA substrate and then place a 3-inch diameter wafer with LIGA gears onto the pins. Upward and downward looking microscopes are used to locate holes in the LIGA substrate, pins to be pressed in the holes, and gears to be placed on the pins. This vision system can locate parts within 3 microns, while the Cartesian manipulator can place the parts within 0.4 microns.

  5. True Shear Parallel Plate Viscometer

    NASA Technical Reports Server (NTRS)

    Ethridge, Edwin; Kaukler, William

    2010-01-01

    This viscometer (which can also be used as a rheometer) is designed for use with liquids over a large temperature range. The device consists of horizontally disposed, similarly sized, parallel plates with a precisely known gap. The lower plate is driven laterally with a motor to apply shear to the liquid in the gap. The upper plate is freely suspended from a double-arm pendulum with a sufficiently long radius to reduce height variations during the swing to negligible levels. A sensitive load cell measures the shear force applied by the liquid to the upper plate. Viscosity is measured by taking the ratio of shear stress to shear rate.

  6. Scheduling Tasks In Parallel Processing

    NASA Technical Reports Server (NTRS)

    Price, Camille C.; Salama, Moktar A.

    1989-01-01

    Algorithms sought to minimize time and cost of computation. Report describes research on scheduling of computations tasks in system of multiple identical data processors operating in parallel. Computational intractability requires use of suboptimal heuristic algorithms. First algorithm called "list heuristic", variation of classical list scheduling. Second algorithm called "cluster heuristic" applied to tightly coupled tasks and consists of four phases. Third algorithm called "exchange heuristic", iterative-improvement algorithm beginning with initial feasible assignment of tasks to processors and periods of time. Fourth algorithm is iterative one for optimal assignment of tasks and based on concept called "simulated annealing" because of mathematical resemblance to aspects of physical annealing processes.

  7. Heart Fibrillation and Parallel Supercomputers

    NASA Technical Reports Server (NTRS)

    Kogan, B. Y.; Karplus, W. J.; Chudin, E. E.

    1997-01-01

    The Luo and Rudy 3 cardiac cell mathematical model is implemented on the parallel supercomputer CRAY - T3D. The splitting algorithm combined with variable time step and an explicit method of integration provide reasonable solution times and almost perfect scaling for rectilinear wave propagation. The computer simulation makes it possible to observe new phenomena: the break-up of spiral waves caused by intracellular calcium and dynamics and the non-uniformity of the calcium distribution in space during the onset of the spiral wave.

  8. Xyce parallel electronic simulator design.

    SciTech Connect

    Thornquist, Heidi K.; Rankin, Eric Lamont; Mei, Ting; Schiek, Richard Louis; Keiter, Eric Richard; Russo, Thomas V.

    2010-09-01

    This document is the Xyce Circuit Simulator developer guide. Xyce has been designed from the 'ground up' to be a SPICE-compatible, distributed memory parallel circuit simulator. While it is in many respects a research code, Xyce is intended to be a production simulator. As such, having software quality engineering (SQE) procedures in place to insure a high level of code quality and robustness are essential. Version control, issue tracking customer support, C++ style guildlines and the Xyce release process are all described. The Xyce Parallel Electronic Simulator has been under development at Sandia since 1999. Historically, Xyce has mostly been funded by ASC, the original focus of Xyce development has primarily been related to circuits for nuclear weapons. However, this has not been the only focus and it is expected that the project will diversify. Like many ASC projects, Xyce is a group development effort, which involves a number of researchers, engineers, scientists, mathmaticians and computer scientists. In addition to diversity of background, it is to be expected on long term projects for there to be a certain amount of staff turnover, as people move on to different projects. As a result, it is very important that the project maintain high software quality standards. The point of this document is to formally document a number of the software quality practices followed by the Xyce team in one place. Also, it is hoped that this document will be a good source of information for new developers.

  9. Parallel job-scheduling algorithms

    SciTech Connect

    Rodger, S.H.

    1989-01-01

    In this thesis, we consider solving job scheduling problems on the CREW PRAM model. We show how to adapt Cole's pipeline merge technique to yield several efficient parallel algorithms for a number of job scheduling problems and one optimal parallel algorithm for the following job scheduling problem: Given a set of n jobs defined by release times, deadlines and processing times, find a schedule that minimizes the maximum lateness of the jobs and allows preemption when the jobs are scheduled to run on one machine. In addition, we present the first NC algorithm for the following job scheduling problem: Given a set of n jobs defined by release times, deadlines and unit processing times, determine if there is a schedule of jobs on one machine, and calculate the schedule if it exists. We identify the notion of a canonical schedule, which is the type of schedule our algorithm computes if there is a schedule. Our algorithm runs in O((log n){sup 2}) time and uses O(n{sup 2}k{sup 2}) processors, where k is the minimum number of distinct offsets of release times or deadlines.

  10. Partitioning in parallel processing of production systems

    SciTech Connect

    Oflazer, K.

    1987-01-01

    This thesis presents research on certain issues related to parallel processing of production systems. It first presents a parallel production system interpreter that has been implemented on a four-processor multiprocessor. This parallel interpreter is based on Forgy's OPS5 interpreter and exploits production-level parallelism in production systems. Runs on the multiprocessor system indicate that it is possible to obtain speed-up of around 1.7 in the match computation for certain production systems when productions are split into three sets that are processed in parallel. The next issue addressed is that of partitioning a set of rules to processors in a parallel interpreter with production-level parallelism, and the extent of additional improvement in performance. The partitioning problem is formulated and an algorithm for approximate solutions is presented. The thesis next presents a parallel processing scheme for OPS5 production systems that allows some redundancy in the match computation. This redundancy enables the processing of a production to be divided into units of medium granularity each of which can be processed in parallel. Subsequently, a parallel processor architecture for implementing the parallel processing algorithm is presented.

  11. Satellite and surface geophysical expression of anomalous crustal structure in Kentucky and Tennessee

    NASA Technical Reports Server (NTRS)

    Mayhew, M. A.; Thomas, H. H.; Wasilewski, P. J.

    1981-01-01

    An equivalent layer magnetization model is discussed. Inversion of long wavelength satellite magnetic anomaly data indicates a very magnetic source region centered in south central Kentucky. Refraction profiles suggest that the source of the gravity anomaly is a large mass of rock occupying much of the crustal thickness. The outline of the source delineated by gravity contours is also discernible in aeromagnetic anomaly patterns. The mafic plutonic complex, and several lines of evidence are consistent with a rift association. The body is, however, clearly related to the inferred position of the Grenville Front. It is bounded on the north by the fault zones of the 38th Parallel Lineament. It is suggested that such magnetization levels are achieved with magnetic mineralogies produced by normal oxidation and metamorphic processes and enhanced by viscous build-up, especially in mafic rocks of alkaline character.

  12. Magnetic anomaly map of North America south of 50 degrees north from Pogo data

    NASA Technical Reports Server (NTRS)

    Mayhew, M. A.

    1976-01-01

    A magnetic anomaly map produced from Pogo data for North America and adjacent ocean areas is presented. At satellite elevations anomalies have wavelengths measured in hundreds of kilometers, and reflect regional structures on a large scale. Prominent features of the map are: (1) a large east-west high through the mid-continent, breached at the Mississippi Embayment; (2) a broad low over the Gulf of Mexico; (3) a strong gradient separating these features, which follows the Southern Appalachian-Ouachita curvature; and (4) a high over the Antilles-Bahamas Platform which extends to northern Florida. A possible relationship between the high of the mid-continent and the 38th parallel lineament is noted.

  13. International Instrumentation Symposium, 38th, Las Vegas, NV, Apr. 26-30, 1992, Proceedings

    NASA Astrophysics Data System (ADS)

    The present volume on aerospace instrumentation discusses computer applications, blast and shock, implementation of the Clean Air Act amendments, and thermal systems. Attention is given to measurement uncertainty/flow measurement, data acquisition and processing, force/acceleration/motion measurements, and hypersonics/reentry vehicle systems. Topics addressed include wind tunnels, real time systems, and pressure effects. Also discussed are a distributed data and control system for space simulation and thermal testing a stepwise shockwave velocity determinator, computer tracking and decision making, the use of silicon diodes for detecting the liquid-vapor interface in hydrogen, and practical methods for analysis of uncertainty propagation.

  14. The society for craniofacial genetics and developmental biology 38th annual meeting.

    PubMed

    Taneyhill, Lisa A; Hoover-Fong, Julie; Lozanoff, Scott; Marcucio, Ralph; Richtsmeier, Joan T; Trainor, Paul A

    2016-07-01

    The mission of the Society for Craniofacial Genetics and Developmental Biology (SCGDB) is to promote education, research, and communication about normal and abnormal development of the tissues and organs of the head. The SCGDB welcomes as members undergraduate students, graduate students, post doctoral researchers, clinicians, orthodontists, scientists, and academicians who share an interest in craniofacial biology. Each year our members come together to share their novel findings, build upon, and challenge current knowledge of craniofacial biology. © 2016 Wiley Periodicals, Inc.

  15. Annual Adult Education Research Conference Proceedings (38th, Stillwater, Oklahoma, May 16-18, 1997).

    ERIC Educational Resources Information Center

    Nolan, Robert E., Comp.; Chelesvig, Heath, Comp.

    The following are among 50 papers included: "The Politics of Planning Culturally Relevant AIDS Education for African-American Women" (Archie-Booker); "Developing White Consciousness through a Transformative Learning Process" (Barlas); "Executive Businesswomen's Learning in the Context of Organizational Culture"…

  16. Parallel processing for scientific computations

    NASA Technical Reports Server (NTRS)

    Alkhatib, Hasan S.

    1991-01-01

    The main contribution of the effort in the last two years is the introduction of the MOPPS system. After doing extensive literature search, we introduced the system which is described next. MOPPS employs a new solution to the problem of managing programs which solve scientific and engineering applications on a distributed processing environment. Autonomous computers cooperate efficiently in solving large scientific problems with this solution. MOPPS has the advantage of not assuming the presence of any particular network topology or configuration, computer architecture, or operating system. It imposes little overhead on network and processor resources while efficiently managing programs concurrently. The core of MOPPS is an intelligent program manager that builds a knowledge base of the execution performance of the parallel programs it is managing under various conditions. The manager applies this knowledge to improve the performance of future runs. The program manager learns from experience.

  17. Hybrid Optimization Parallel Search PACKage

    SciTech Connect

    2009-11-10

    HOPSPACK is open source software for solving optimization problems without derivatives. Application problems may have a fully nonlinear objective function, bound constraints, and linear and nonlinear constraints. Problem variables may be continuous, integer-valued, or a mixture of both. The software provides a framework that supports any derivative-free type of solver algorithm. Through the framework, solvers request parallel function evaluation, which may use MPI (multiple machines) or multithreading (multiple processors/cores on one machine). The framework provides a Cache and Pending Cache of saved evaluations that reduces execution time and facilitates restarts. Solvers can dynamically create other algorithms to solve subproblems, a useful technique for handling multiple start points and integer-valued variables. HOPSPACK ships with the Generating Set Search (GSS) algorithm, developed at Sandia as part of the APPSPACK open source software project.

  18. Parallel Performance Characterization of Columbia

    NASA Technical Reports Server (NTRS)

    Biswas, Rupak

    2004-01-01

    Using a collection of benchmark problems of increasing levels of realism and computational effort, we will characterize the strengths and limitations of the 10,240 processor Columbia system to deliver supercomputing value to application scientists. Scientists need to be able to determine if and how they can utilize Columbia to carry extreme workloads, either in terms of ultra-large applications that cannot be run otherwise (capability), or in terms of very large ensembles of medium-scale applications to populate response matrices (capacity). We select existing application benchmarks that scale from a small number of processors to the entire machine, and that highlight different issues in running supercomputing-calss applicaions, such as the various types of memory access, file I/O, inter- and intra-node communications and parallelization paradigms. http://www.nas.nasa.gov/Software/NPB/

  19. Information hiding in parallel programs

    SciTech Connect

    Foster, I.

    1992-01-30

    A fundamental principle in program design is to isolate difficult or changeable design decisions. Application of this principle to parallel programs requires identification of decisions that are difficult or subject to change, and the development of techniques for hiding these decisions. We experiment with three complex applications, and identify mapping, communication, and scheduling as areas in which decisions are particularly problematic. We develop computational abstractions that hide such decisions, and show that these abstractions can be used to develop elegant solutions to programming problems. In particular, they allow us to encode common structures, such as transforms, reductions, and meshes, as software cells and templates that can reused in different applications. An important characteristic of these structures is that they do not incorporate mapping, communication, or scheduling decisions: these aspects of the design are specified separately, when composing existing structures to form applications. This separation of concerns allows the same cells and templates to be reused in different contexts.

  20. Embodied and Distributed Parallel DJing.

    PubMed

    Cappelen, Birgitta; Andersson, Anders-Petter

    2016-01-01

    Everyone has a right to take part in cultural events and activities, such as music performances and music making. Enforcing that right, within Universal Design, is often limited to a focus on physical access to public areas, hearing aids etc., or groups of persons with special needs performing in traditional ways. The latter might be people with disabilities, being musicians playing traditional instruments, or actors playing theatre. In this paper we focus on the innovative potential of including people with special needs, when creating new cultural activities. In our project RHYME our goal was to create health promoting activities for children with severe disabilities, by developing new musical and multimedia technologies. Because of the users' extreme demands and rich contribution, we ended up creating both a new genre of musical instruments and a new art form. We call this new art form Embodied and Distributed Parallel DJing, and the new genre of instruments for Empowering Multi-Sensorial Things.

  1. Parallel spinors on flat manifolds

    NASA Astrophysics Data System (ADS)

    Sadowski, Michał

    2006-05-01

    Let p(M) be the dimension of the vector space of parallel spinors on a closed spin manifold M. We prove that every finite group G is the holonomy group of a closed flat spin manifold M(G) such that p(M(G))>0. If the holonomy group Hol(M) of M is cyclic, then we give an explicit formula for p(M) another than that given in [R.J. Miatello, R.A. Podesta, The spectrum of twisted Dirac operators on compact flat manifolds, Trans. Am. Math. Soc., in press]. We answer the question when p(M)>0 if Hol(M) is a cyclic group of prime order or dim⁡M≤4.

  2. Device for balancing parallel strings

    DOEpatents

    Mashikian, Matthew S.

    1985-01-01

    A battery plant is described which features magnetic circuit means in association with each of the battery strings in the battery plant for balancing the electrical current flow through the battery strings by equalizing the voltage across each of the battery strings. Each of the magnetic circuit means generally comprises means for sensing the electrical current flow through one of the battery strings, and a saturable reactor having a main winding connected electrically in series with the battery string, a bias winding connected to a source of alternating current and a control winding connected to a variable source of direct current controlled by the sensing means. Each of the battery strings is formed by a plurality of batteries connected electrically in series, and these battery strings are connected electrically in parallel across common bus conductors.

  3. Parallel network simulations with NEURON.

    PubMed

    Migliore, M; Cannia, C; Lytton, W W; Markram, Henry; Hines, M L

    2006-10-01

    The NEURON simulation environment has been extended to support parallel network simulations. Each processor integrates the equations for its subnet over an interval equal to the minimum (interprocessor) presynaptic spike generation to postsynaptic spike delivery connection delay. The performance of three published network models with very different spike patterns exhibits superlinear speedup on Beowulf clusters and demonstrates that spike communication overhead is often less than the benefit of an increased fraction of the entire problem fitting into high speed cache. On the EPFL IBM Blue Gene, almost linear speedup was obtained up to 100 processors. Increasing one model from 500 to 40,000 realistic cells exhibited almost linear speedup on 2,000 processors, with an integration time of 9.8 seconds and communication time of 1.3 seconds. The potential for speed-ups of several orders of magnitude makes practical the running of large network simulations that could otherwise not be explored.

  4. Parallel computing in enterprise modeling.

    SciTech Connect

    Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.

    2008-08-01

    This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.

  5. Integrated Task and Data Parallel Programming

    NASA Technical Reports Server (NTRS)

    Grimshaw, A. S.

    1998-01-01

    This research investigates the combination of task and data parallel language constructs within a single programming language. There are an number of applications that exhibit properties which would be well served by such an integrated language. Examples include global climate models, aircraft design problems, and multidisciplinary design optimization problems. Our approach incorporates data parallel language constructs into an existing, object oriented, task parallel language. The language will support creation and manipulation of parallel classes and objects of both types (task parallel and data parallel). Ultimately, the language will allow data parallel and task parallel classes to be used either as building blocks or managers of parallel objects of either type, thus allowing the development of single and multi-paradigm parallel applications. 1995 Research Accomplishments In February I presented a paper at Frontiers 1995 describing the design of the data parallel language subset. During the spring I wrote and defended my dissertation proposal. Since that time I have developed a runtime model for the language subset. I have begun implementing the model and hand-coding simple examples which demonstrate the language subset. I have identified an astrophysical fluid flow application which will validate the data parallel language subset. 1996 Research Agenda Milestones for the coming year include implementing a significant portion of the data parallel language subset over the Legion system. Using simple hand-coded methods, I plan to demonstrate (1) concurrent task and data parallel objects and (2) task parallel objects managing both task and data parallel objects. My next steps will focus on constructing a compiler and implementing the fluid flow application with the language. Concurrently, I will conduct a search for a real-world application exhibiting both task and data parallelism within the same program. Additional 1995 Activities During the fall I collaborated

  6. Computer-Aided Parallelizer and Optimizer

    NASA Technical Reports Server (NTRS)

    Jin, Haoqiang

    2011-01-01

    The Computer-Aided Parallelizer and Optimizer (CAPO) automates the insertion of compiler directives (see figure) to facilitate parallel processing on Shared Memory Parallel (SMP) machines. While CAPO currently is integrated seamlessly into CAPTools (developed at the University of Greenwich, now marketed as ParaWise), CAPO was independently developed at Ames Research Center as one of the components for the Legacy Code Modernization (LCM) project. The current version takes serial FORTRAN programs, performs interprocedural data dependence analysis, and generates OpenMP directives. Due to the widely supported OpenMP standard, the generated OpenMP codes have the potential to run on a wide range of SMP machines. CAPO relies on accurate interprocedural data dependence information currently provided by CAPTools. Compiler directives are generated through identification of parallel loops in the outermost level, construction of parallel regions around parallel loops and optimization of parallel regions, and insertion of directives with automatic identification of private, reduction, induction, and shared variables. Attempts also have been made to identify potential pipeline parallelism (implemented with point-to-point synchronization). Although directives are generated automatically, user interaction with the tool is still important for producing good parallel codes. A comprehensive graphical user interface is included for users to interact with the parallelization process.

  7. Parallel processing considerations for image recognition tasks

    NASA Astrophysics Data System (ADS)

    Simske, Steven J.

    2011-01-01

    Many image recognition tasks are well-suited to parallel processing. The most obvious example is that many imaging tasks require the analysis of multiple images. From this standpoint, then, parallel processing need be no more complicated than assigning individual images to individual processors. However, there are three less trivial categories of parallel processing that will be considered in this paper: parallel processing (1) by task; (2) by image region; and (3) by meta-algorithm. Parallel processing by task allows the assignment of multiple workflows-as diverse as optical character recognition [OCR], document classification and barcode reading-to parallel pipelines. This can substantially decrease time to completion for the document tasks. For this approach, each parallel pipeline is generally performing a different task. Parallel processing by image region allows a larger imaging task to be sub-divided into a set of parallel pipelines, each performing the same task but on a different data set. This type of image analysis is readily addressed by a map-reduce approach. Examples include document skew detection and multiple face detection and tracking. Finally, parallel processing by meta-algorithm allows different algorithms to be deployed on the same image simultaneously. This approach may result in improved accuracy.

  8. Towards Distributed Memory Parallel Program Analysis

    SciTech Connect

    Quinlan, D; Barany, G; Panas, T

    2008-06-17

    This paper presents a parallel attribute evaluation for distributed memory parallel computer architectures where previously only shared memory parallel support for this technique has been developed. Attribute evaluation is a part of how attribute grammars are used for program analysis within modern compilers. Within this work, we have extended ROSE, a open compiler infrastructure, with a distributed memory parallel attribute evaluation mechanism to support user defined global program analysis required for some forms of security analysis which can not be addressed by a file by file view of large scale applications. As a result, user defined security analyses may now run in parallel without the user having to specify the way data is communicated between processors. The automation of communication enables an extensible open-source parallel program analysis infrastructure.

  9. Runtime volume visualization for parallel CFD

    NASA Technical Reports Server (NTRS)

    Ma, Kwan-Liu

    1995-01-01

    This paper discusses some aspects of design of a data distributed, massively parallel volume rendering library for runtime visualization of parallel computational fluid dynamics simulations in a message-passing environment. Unlike the traditional scheme in which visualization is a postprocessing step, the rendering is done in place on each node processor. Computational scientists who run large-scale simulations on a massively parallel computer can thus perform interactive monitoring of their simulations. The current library provides an interface to handle volume data on rectilinear grids. The same design principles can be generalized to handle other types of grids. For demonstration, we run a parallel Navier-Stokes solver making use of this rendering library on the Intel Paragon XP/S. The interactive visual response achieved is found to be very useful. Performance studies show that the parallel rendering process is scalable with the size of the simulation as well as with the parallel computer.

  10. Parallel reactor systems for bioprocess development.

    PubMed

    Weuster-Botz, Dirk

    2005-01-01

    Controlled parallel bioreactor systems allow fed-batch operation at early stages of process development. The characteristics of shaken bioreactors operated in parallel (shake flask, microtiter plate), sparged bioreactors (small-scale bubble column) and stirred bioreactors (stirred-tank, stirred column) are briefly summarized. Parallel fed-batch operation is achieved with an intermittent feeding and pH-control system for up to 16 bioreactors operated in parallel on a scale of 100 ml. Examples of the scale-up and scale-down of pH-controlled microbial fed-batch processes demonstrate that controlled parallel reactor systems can result in more effective bioprocess development. Future developments are also outlined, including units of 48 parallel stirred-tank reactors with individual pH- and pO2-controls and automation as well as liquid handling system, operated on a scale of ml.

  11. Linearly exact parallel closures for slab geometry

    NASA Astrophysics Data System (ADS)

    Ji, Jeong-Young; Held, Eric D.; Jhang, Hogun

    2013-08-01

    Parallel closures are obtained by solving a linearized kinetic equation with a model collision operator using the Fourier transform method. The closures expressed in wave number space are exact for time-dependent linear problems to within the limits of the model collision operator. In the adiabatic, collisionless limit, an inverse Fourier transform is performed to obtain integral (nonlocal) parallel closures in real space; parallel heat flow and viscosity closures for density, temperature, and flow velocity equations replace Braginskii's parallel closure relations, and parallel flow velocity and heat flow closures for density and temperature equations replace Spitzer's parallel transport relations. It is verified that the closures reproduce the exact linear response function of Hammett and Perkins [Phys. Rev. Lett. 64, 3019 (1990)] for Landau damping given a temperature gradient. In contrast to their approximate closures where the vanishing viscosity coefficient numerically gives an exact response, our closures relate the heat flow and nonvanishing viscosity to temperature and flow velocity (gradients).

  12. Design considerations for parallel graphics libraries

    NASA Technical Reports Server (NTRS)

    Crockett, Thomas W.

    1994-01-01

    Applications which run on parallel supercomputers are often characterized by massive datasets. Converting these vast collections of numbers to visual form has proven to be a powerful aid to comprehension. For a variety of reasons, it may be desirable to provide this visual feedback at runtime. One way to accomplish this is to exploit the available parallelism to perform graphics operations in place. In order to do this, we need appropriate parallel rendering algorithms and library interfaces. This paper provides a tutorial introduction to some of the issues which arise in designing parallel graphics libraries and their underlying rendering algorithms. The focus is on polygon rendering for distributed memory message-passing systems. We illustrate our discussion with examples from PGL, a parallel graphics library which has been developed on the Intel family of parallel systems.

  13. Automatic Multilevel Parallelization Using OpenMP

    NASA Technical Reports Server (NTRS)

    Jin, Hao-Qiang; Jost, Gabriele; Yan, Jerry; Ayguade, Eduard; Gonzalez, Marc; Martorell, Xavier; Biegel, Bryan (Technical Monitor)

    2002-01-01

    In this paper we describe the extension of the CAPO parallelization support tool to support multilevel parallelism based on OpenMP directives. CAPO generates OpenMP directives with extensions supported by the NanosCompiler to allow for directive nesting and definition of thread groups. We report first results for several benchmark codes and one full application that have been parallelized using our system.

  14. Inverse Kinematics for a Parallel Myoelectric Elbow

    DTIC Science & Technology

    2001-10-25

    Inverse Kinematics for a Parallel Myoelectric Elbow A. Z. Escudero, Ja. Álvarez, L. Leija. Center of Research and Advanced Studies of the IPN...replacement above elbow are serial mechanisms driven by a DC motor and they include only one active articulation for the elbow [1]. Parallel mechanisms...are rather scarce [2]. The inverse kinematics model of a 3-degree of freedom parallel prosthetic elbow mechanism is reported. The mathematical

  15. Parallel computations and control of adaptive structures

    NASA Technical Reports Server (NTRS)

    Park, K. C.; Alvin, Kenneth F.; Belvin, W. Keith; Chong, K. P. (Editor); Liu, S. C. (Editor); Li, J. C. (Editor)

    1991-01-01

    The equations of motion for structures with adaptive elements for vibration control are presented for parallel computations to be used as a software package for real-time control of flexible space structures. A brief introduction of the state-of-the-art parallel computational capability is also presented. Time marching strategies are developed for an effective use of massive parallel mapping, partitioning, and the necessary arithmetic operations. An example is offered for the simulation of control-structure interaction on a parallel computer and the impact of the approach presented for applications in other disciplines than aerospace industry is assessed.

  16. Parallel auto-correlative statistics with VTK.

    SciTech Connect

    Pebay, Philippe Pierre; Bennett, Janine Camille

    2013-08-01

    This report summarizes existing statistical engines in VTK and presents both the serial and parallel auto-correlative statistics engines. It is a sequel to [PT08, BPRT09b, PT09, BPT09, PT10] which studied the parallel descriptive, correlative, multi-correlative, principal component analysis, contingency, k-means, and order statistics engines. The ease of use of the new parallel auto-correlative statistics engine is illustrated by the means of C++ code snippets and algorithm verification is provided. This report justifies the design of the statistics engines with parallel scalability in mind, and provides scalability and speed-up analysis results for the autocorrelative statistics engine.

  17. Parallel programming in Split-C

    SciTech Connect

    Culler, D.E.; Dusseau, A.; Goldstein, S.C.; Krishnamurthy, A.; Lumetta, S.; Eicken, T. von; Yelick, K.

    1993-12-31

    The authors introduce the Split-C language, a parallel extension of C intended for high performance programming on distributed memory multiprocessors, and demonstrate the use of the language in optimizing parallel programs. Split-C provides a global address space with a clear concept of locality and unusual assignment operators. These are used as tools to reduce the frequency and cost of remote access. The language allows a mixture of shared memory, message passing, and data parallel programming styles while providing efficient access to the underlying machine. They demonstrate the basic language concepts using regular and irregular parallel programs and give performance results for various stages of program optimization.

  18. Shared-memory parallel programming in C++

    SciTech Connect

    Beck, B. )

    1990-07-01

    This paper discusses how researchers have produced a set of portable parallel-programming constructs for C, implemented in M4 macros. These parallel-programming macros are available under the name Parmacs. The Parmacs macros let one write parallel C programs for shared-memory, distributed-memory, and mixed-memory (shared and distributed) systems. They have been implemented on several machines. Because Parmacs offers useful parallel-programming features, the author has considered how these problems might be overcome or avoided. The author thought that using C++, rather than C, would address these problems adequately, and describes the C++ features exploited. The work described addresses shared-memory constructs.

  19. Parallel Algorithms for the Exascale Era

    SciTech Connect

    Robey, Robert W.

    2016-10-19

    New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this work has been done by undergraduates and published in leading scientific journals.

  20. A parallel algorithm for global routing

    NASA Technical Reports Server (NTRS)

    Brouwer, Randall J.; Banerjee, Prithviraj

    1990-01-01

    A Parallel Hierarchical algorithm for Global Routing (PHIGURE) is presented. The router is based on the work of Burstein and Pelavin, but has many extensions for general global routing and parallel execution. Main features of the algorithm include structured hierarchical decomposition into separate independent tasks which are suitable for parallel execution and adaptive simplex solution for adding feedthroughs and adjusting channel heights for row-based layout. Alternative decomposition methods and the various levels of parallelism available in the algorithm are examined closely. The algorithm is described and results are presented for a shared-memory multiprocessor implementation.

  1. Conformal pure radiation with parallel rays

    NASA Astrophysics Data System (ADS)

    Leistner, Thomas; Nurowski, Paweł

    2012-03-01

    We define pure radiation metrics with parallel rays to be n-dimensional pseudo-Riemannian metrics that admit a parallel null line bundle K and whose Ricci tensor vanishes on vectors that are orthogonal to K. We give necessary conditions in terms of the Weyl, Cotton and Bach tensors for a pseudo-Riemannian metric to be conformal to a pure radiation metric with parallel rays. Then, we derive conditions in terms of the tractor calculus that are equivalent to the existence of a pure radiation metric with parallel rays in a conformal class. We also give analogous results for n-dimensional pseudo-Riemannian pp-waves.

  2. Parallel Genetic Algorithm for Alpha Spectra Fitting

    NASA Astrophysics Data System (ADS)

    García-Orellana, Carlos J.; Rubio-Montero, Pilar; González-Velasco, Horacio

    2005-01-01

    We present a performance study of alpha-particle spectra fitting using parallel Genetic Algorithm (GA). The method uses a two-step approach. In the first step we run parallel GA to find an initial solution for the second step, in which we use Levenberg-Marquardt (LM) method for a precise final fit. GA is a high resources-demanding method, so we use a Beowulf cluster for parallel simulation. The relationship between simulation time (and parallel efficiency) and processors number is studied using several alpha spectra, with the aim of obtaining a method to estimate the optimal processors number that must be used in a simulation.

  3. Parallel processing for scientific computations

    NASA Technical Reports Server (NTRS)

    Alkhatib, Hasan S.

    1995-01-01

    The scope of this project dealt with the investigation of the requirements to support distributed computing of scientific computations over a cluster of cooperative workstations. Various experiments on computations for the solution of simultaneous linear equations were performed in the early phase of the project to gain experience in the general nature and requirements of scientific applications. A specification of a distributed integrated computing environment, DICE, based on a distributed shared memory communication paradigm has been developed and evaluated. The distributed shared memory model facilitates porting existing parallel algorithms that have been designed for shared memory multiprocessor systems to the new environment. The potential of this new environment is to provide supercomputing capability through the utilization of the aggregate power of workstations cooperating in a cluster interconnected via a local area network. Workstations, generally, do not have the computing power to tackle complex scientific applications, making them primarily useful for visualization, data reduction, and filtering as far as complex scientific applications are concerned. There is a tremendous amount of computing power that is left unused in a network of workstations. Very often a workstation is simply sitting idle on a desk. A set of tools can be developed to take advantage of this potential computing power to create a platform suitable for large scientific computations. The integration of several workstations into a logical cluster of distributed, cooperative, computing stations presents an alternative to shared memory multiprocessor systems. In this project we designed and evaluated such a system.

  4. Parallel Computing for Brain Simulation.

    PubMed

    Pastur-Romay, L A; Porto-Pazos, A B; Cedrón, F; Pazos, A

    2016-11-04

    The human brain is the most complex system in the known universe, but it is the most unknown system. It allows the human beings to possess extraordinary capacities. However, we don´t understand yet how and why most of these capacities are produced. For decades, it have been tried that the computers reproduces these capacities. On one hand, to help understanding the nervous system. On the other hand, to process the data in a more efficient way than before. It is intended to make the computers process the information like the brain does it. The important technological developments and the big multidisciplinary projects have allowed create the first simulation with a number of neurons similar to the human brain neurons number. This paper presents an update review about the main research projects that are trying of simulate and/or emulate the human brain. They employ different types of computational models using parallel computing: digital models, analog models and hybrid models. This review includes the actual applications of these works and also the future trends. We have reviewed some works that look for a step forward in Neuroscience and other ones that look for a breakthrough in Computer Science (neuromorphic hardware, machine learning techniques). We summarize the most outstanding characteristics of them and present the latest advances and future plans. In addition, this review remarks the importance of considering not only neurons: the computational models of the brain should include glial cells, given the proven importance of the astrocytes in the information processing.

  5. Parallel methods for the flight simulation model

    SciTech Connect

    Xiong, Wei Zhong; Swietlik, C.

    1994-06-01

    The Advanced Computer Applications Center (ACAC) has been involved in evaluating advanced parallel architecture computers and the applicability of these machines to computer simulation models. The advanced systems investigated include parallel machines with shared. memory and distributed architectures consisting of an eight processor Alliant FX/8, a twenty four processor sor Sequent Symmetry, Cray XMP, IBM RISC 6000 model 550, and the Intel Touchstone eight processor Gamma and 512 processor Delta machines. Since parallelizing a truly efficient application program for the parallel machine is a difficult task, the implementation for these machines in a realistic setting has been largely overlooked. The ACAC has developed considerable expertise in optimizing and parallelizing application models on a collection of advanced multiprocessor systems. One of aspect of such an application model is the Flight Simulation Model, which used a set of differential equations to describe the flight characteristics of a launched missile by means of a trajectory. The Flight Simulation Model was written in the FORTRAN language with approximately 29,000 lines of source code. Depending on the number of trajectories, the computation can require several hours to full day of CPU time on DEC/VAX 8650 system. There is an impetus to reduce the execution time and utilize the advanced parallel architecture computing environment available. ACAC researchers developed a parallel method that allows the Flight Simulation Model to be able to run in parallel on the multiprocessor system. For the benchmark data tested, the parallel Flight Simulation Model implemented on the Alliant FX/8 has achieved nearly linear speedup. In this paper, we describe a parallel method for the Flight Simulation Model. We believe the method presented in this paper provides a general concept for the design of parallel applications. This concept, in most cases, can be adapted to many other sequential application programs.

  6. Recent Improvements to HST Parallel Scheduling

    NASA Astrophysics Data System (ADS)

    Henry, Ronald; Butschky, Mike

    The Hubble Space Telescope (HST) has several scientific instruments (SIs) that may be used at any given time. Most primary visits submitted by HST observers only use one SI, leaving the other SIs free to be requested by ``pure parallel'' observing programs. In order to accomplish this, separate scheduling units (SUs) for each parallel SI must be created and then scheduled by the Science Planning and Scheduling System (SPSS), taking into account numerous orbital and scientific constraints. The Parallel Observation Matching System (POMS) has the task of matching parallel visits to primary observations and ``crafting'' appropriate parallel SUs at each opportunity, taking scientific criteria and orbital constraints into account. The process for planning and scheduling parallel observations is thus quite different from the process for primary science. In the past, custom crafting rules for each parallel program were necessary, requiring full-time support from a software developer. In addition, because POMS ran as a standalone system, its ability to model how long parallel SUs would take was limited, especially with the flexible buffer-management schemes used for the second-generation SIs. A new version of POMS was developed in 1997. This version uses a formal proposal syntax (the same used for primary observations) for parallels, so that different proposals can be handled uniformly and without the need for customized ``crafting rules.'' In addition, POMS is integrated with the Transformation (TRANS) planning system in order to give it full knowledge of overheads within an SU, eliminating the need for ad hoc modeling. The power and versatility of this approach has paid off in improved utilization of parallel opportunities, greatly reduced maintenance costs, and an ability to gracefully handle new parallel proposals and new SIs with minimal software effort. This paper discusses the requirements, design, and operational results of the new POMS.

  7. High-Performance Psychometrics: The Parallel-E Parallel-M Algorithm for Generalized Latent Variable Models. Research Report. ETS RR-16-34

    ERIC Educational Resources Information Center

    von Davier, Matthias

    2016-01-01

    This report presents results on a parallel implementation of the expectation-maximization (EM) algorithm for multidimensional latent variable models. The developments presented here are based on code that parallelizes both the E step and the M step of the parallel-E parallel-M algorithm. Examples presented in this report include item response…

  8. The language parallel Pascal and other aspects of the massively parallel processor

    NASA Technical Reports Server (NTRS)

    Reeves, A. P.; Bruner, J. D.

    1982-01-01

    A high level language for the Massively Parallel Processor (MPP) was designed. This language, called Parallel Pascal, is described in detail. A description of the language design, a description of the intermediate language, Parallel P-Code, and details for the MPP implementation are included. Formal descriptions of Parallel Pascal and Parallel P-Code are given. A compiler was developed which converts programs in Parallel Pascal into the intermediate Parallel P-Code language. The code generator to complete the compiler for the MPP is being developed independently. A Parallel Pascal to Pascal translator was also developed. The architecture design for a VLSI version of the MPP was completed with a description of fault tolerant interconnection networks. The memory arrangement aspects of the MPP are discussed and a survey of other high level languages is given.

  9. Prototyping Parallel and Distributed Programs in Proteus

    DTIC Science & Technology

    1990-10-01

    Cole90, Gibb89]. " Highly-parallel processors - Applications for highly-parallel machines such as the CM- 2 or the iPSC are programmed using data...Programming, (Prentice-Hall, Englewood Cliffs, NJ) 1990. [Gibb89] Gibbons , P.B., "A more practical PRAM model", in: Proceedings of the First ACM

  10. Simulation Exploration through Immersive Parallel Planes: Preprint

    SciTech Connect

    Brunhart-Lupo, Nicholas; Bush, Brian W.; Gruchalla, Kenny; Smith, Steve

    2016-03-01

    We present a visualization-driven simulation system that tightly couples systems dynamics simulations with an immersive virtual environment to allow analysts to rapidly develop and test hypotheses in a high-dimensional parameter space. To accomplish this, we generalize the two-dimensional parallel-coordinates statistical graphic as an immersive 'parallel-planes' visualization for multivariate time series emitted by simulations running in parallel with the visualization. In contrast to traditional parallel coordinate's mapping the multivariate dimensions onto coordinate axes represented by a series of parallel lines, we map pairs of the multivariate dimensions onto a series of parallel rectangles. As in the case of parallel coordinates, each individual observation in the dataset is mapped to a polyline whose vertices coincide with its coordinate values. Regions of the rectangles can be 'brushed' to highlight and select observations of interest: a 'slider' control allows the user to filter the observations by their time coordinate. In an immersive virtual environment, users interact with the parallel planes using a joystick that can select regions on the planes, manipulate selection, and filter time. The brushing and selection actions are used to both explore existing data as well as to launch additional simulations corresponding to the visually selected portions of the input parameter space. As soon as the new simulations complete, their resulting observations are displayed in the virtual environment. This tight feedback loop between simulation and immersive analytics accelerates users' realization of insights about the simulation and its output.

  11. Parallel computation with the spectral element method

    SciTech Connect

    Ma, Hong

    1995-12-01

    Spectral element models for the shallow water equations and the Navier-Stokes equations have been successfully implemented on a data parallel supercomputer, the Connection Machine model CM-5. The nonstaggered grid formulations for both models are described, which are shown to be especially efficient in data parallel computing environment.

  12. Predicting Protein Structure Using Parallel Genetic Algorithms.

    DTIC Science & Technology

    1994-12-01

    By " Predicting rotein Structure D istribticfiar.. ................ Using Parallel Genetic Algorithms ,Avaiu " ’ •"... Dist THESIS I IGeorge H...iiLite-d Approved for public release; distribution unlimited AFIT/ GCS /ENG/94D-03 Predicting Protein Structure Using Parallel Genetic Algorithms ...1-1 1.2 Genetic Algorithms ......... ............................ 1-3 1.3 The Protein Folding Problem

  13. Parallel Activation in Bilingual Phonological Processing

    ERIC Educational Resources Information Center

    Lee, Su-Yeon

    2011-01-01

    In bilingual language processing, the parallel activation hypothesis suggests that bilinguals activate their two languages simultaneously during language processing. Support for the parallel activation mainly comes from studies of lexical (word-form) processing, with relatively less attention to phonological (sound) processing. According to…

  14. Serial Order: A Parallel Distributed Processing Approach.

    ERIC Educational Resources Information Center

    Jordan, Michael I.

    Human behavior shows a variety of serially ordered action sequences. This paper presents a theory of serial order which describes how sequences of actions might be learned and performed. In this theory, parallel interactions across time (coarticulation) and parallel interactions across space (dual-task interference) are viewed as two aspects of a…

  15. MULTIOBJECTIVE PARALLEL GENETIC ALGORITHM FOR WASTE MINIMIZATION

    EPA Science Inventory

    In this research we have developed an efficient multiobjective parallel genetic algorithm (MOPGA) for waste minimization problems. This MOPGA integrates PGAPack (Levine, 1996) and NSGA-II (Deb, 2000) with novel modifications. PGAPack is a master-slave parallel implementation of a...

  16. Parallel Narrative Structure in Paul Harding's "Tinkers"

    ERIC Educational Resources Information Center

    Çirakli, Mustafa Zeki

    2014-01-01

    The present paper explores the implications of parallel narrative structure in Paul Harding's "Tinkers" (2009). Besides primarily recounting the two sets of parallel narratives, "Tinkers" also comprises of seemingly unrelated fragments such as excerpts from clock repair manuals and diaries. The main stories, however, told…

  17. Parallel Computing Strategies for Irregular Algorithms

    NASA Technical Reports Server (NTRS)

    Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)

    2002-01-01

    Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.

  18. Parallel unstructured grid generation for computational aerosciences

    NASA Technical Reports Server (NTRS)

    Shephard, Mark S.

    1993-01-01

    The objective of this research project is to develop efficient parallel automatic grid generation procedures for use in computational aerosciences. This effort is focused on a parallel version of the Finite Octree grid generator. Progress made during the first six months is reported.

  19. Differences Between Distributed and Parallel Systems

    SciTech Connect

    Brightwell, R.; Maccabe, A.B.; Rissen, R.

    1998-10-01

    Distributed systems have been studied for twenty years and are now coming into wider use as fast networks and powerful workstations become more readily available. In many respects a massively parallel computer resembles a network of workstations and it is tempting to port a distributed operating system to such a machine. However, there are significant differences between these two environments and a parallel operating system is needed to get the best performance out of a massively parallel system. This report characterizes the differences between distributed systems, networks of workstations, and massively parallel systems and analyzes the impact of these differences on operating system design. In the second part of the report, we introduce Puma, an operating system specifically developed for massively parallel systems. We describe Puma portals, the basic building blocks for message passing paradigms implemented on top of Puma, and show how the differences observed in the first part of the report have influenced the design and implementation of Puma.

  20. Parallel-In-Time For Moving Meshes

    SciTech Connect

    Falgout, R. D.; Manteuffel, T. A.; Southworth, B.; Schroder, J. B.

    2016-02-04

    With steadily growing computational resources available, scientists must develop e ective ways to utilize the increased resources. High performance, highly parallel software has be- come a standard. However until recent years parallelism has focused primarily on the spatial domain. When solving a space-time partial di erential equation (PDE), this leads to a sequential bottleneck in the temporal dimension, particularly when taking a large number of time steps. The XBraid parallel-in-time library was developed as a practical way to add temporal parallelism to existing se- quential codes with only minor modi cations. In this work, a rezoning-type moving mesh is applied to a di usion problem and formulated in a parallel-in-time framework. Tests and scaling studies are run using XBraid and demonstrate excellent results for the simple model problem considered herein.

  1. Configuration space representation in parallel coordinates

    NASA Technical Reports Server (NTRS)

    Fiorini, Paolo; Inselberg, Alfred

    1989-01-01

    By means of a system of parallel coordinates, a nonprojective mapping from R exp N to R squared is obtained for any positive integer N. In this way multivariate data and relations can be represented in the Euclidean plane (embedded in the projective plane). Basically, R squared with Cartesian coordinates is augmented by N parallel axes, one for each variable. The N joint variables of a robotic device can be represented graphically by using parallel coordinates. It is pointed out that some properties of the relation are better perceived visually from the parallel coordinate representation, and that new algorithms and data structures can be obtained from this representation. The main features of parallel coordinates are described, and an example is presented of their use for configuration space representation of a mechanical arm (where Cartesian coordinates cannot be used).

  2. Implementation and performance of parallel Prolog interpreter

    SciTech Connect

    Wei, S.; Kale, L.V.; Balkrishna, R. . Dept. of Computer Science)

    1988-01-01

    In this paper, the authors discuss the implementation of a parallel Prolog interpreter on different parallel machines. The implementation is based on the REDUCE--OR process model which exploits both AND and OR parallelism in logic programs. It is machine independent as it runs on top of the chare-kernel--a machine-independent parallel programming system. The authors also give the performance of the interpreter running a diverse set of benchmark pargrams on parallel machines including shared memory systems: an Alliant FX/8, Sequent and a MultiMax, and a non-shared memory systems: Intel iPSC/32 hypercube, in addition to its performance on a multiprocessor simulation system.

  3. Parallel Algebraic Multigrid Methods - High Performance Preconditioners

    SciTech Connect

    Yang, U M

    2004-11-11

    The development of high performance, massively parallel computers and the increasing demands of computationally challenging applications have necessitated the development of scalable solvers and preconditioners. One of the most effective ways to achieve scalability is the use of multigrid or multilevel techniques. Algebraic multigrid (AMG) is a very efficient algorithm for solving large problems on unstructured grids. While much of it can be parallelized in a straightforward way, some components of the classical algorithm, particularly the coarsening process and some of the most efficient smoothers, are highly sequential, and require new parallel approaches. This chapter presents the basic principles of AMG and gives an overview of various parallel implementations of AMG, including descriptions of parallel coarsening schemes and smoothers, some numerical results as well as references to existing software packages.

  4. A parallel variable metric optimization algorithm

    NASA Technical Reports Server (NTRS)

    Straeter, T. A.

    1973-01-01

    An algorithm, designed to exploit the parallel computing or vector streaming (pipeline) capabilities of computers is presented. When p is the degree of parallelism, then one cycle of the parallel variable metric algorithm is defined as follows: first, the function and its gradient are computed in parallel at p different values of the independent variable; then the metric is modified by p rank-one corrections; and finally, a single univariant minimization is carried out in the Newton-like direction. Several properties of this algorithm are established. The convergence of the iterates to the solution is proved for a quadratic functional on a real separable Hilbert space. For a finite-dimensional space the convergence is in one cycle when p equals the dimension of the space. Results of numerical experiments indicate that the new algorithm will exploit parallel or pipeline computing capabilities to effect faster convergence than serial techniques.

  5. National Combustion Code: Parallel Implementation and Performance

    NASA Technical Reports Server (NTRS)

    Quealy, A.; Ryder, R.; Norris, A.; Liu, N.-S.

    2000-01-01

    The National Combustion Code (NCC) is being developed by an industry-government team for the design and analysis of combustion systems. CORSAIR-CCD is the current baseline reacting flow solver for NCC. This is a parallel, unstructured grid code which uses a distributed memory, message passing model for its parallel implementation. The focus of the present effort has been to improve the performance of the NCC flow solver to meet combustor designer requirements for model accuracy and analysis turnaround time. Improving the performance of this code contributes significantly to the overall reduction in time and cost of the combustor design cycle. This paper describes the parallel implementation of the NCC flow solver and summarizes its current parallel performance on an SGI Origin 2000. Earlier parallel performance results on an IBM SP-2 are also included. The performance improvements which have enabled a turnaround of less than 15 hours for a 1.3 million element fully reacting combustion simulation are described.

  6. Parallelization of a Compositional Reservoir Simulator

    NASA Astrophysics Data System (ADS)

    Reme, Hilde; Åge Øye, Geir; Espedal, Magne S.; Fladmark, Gunnar E.

    A finite volume dicretization has been used to solve compositional flow in porous media. Secondary migration in fractured rocks has been the main motivation for the work. Multipoint flux approximation has been implemented and adaptive local grid refinement, based on domain decomposition, is used at fractures and faults. The parallelization method, which is described in this paper, strongly promotes code reuse and gives a very high level of parallelization despite low implementation costs. The programming framework is also portable to other platforms or other applications. We have presented computer experiments to examine the parallel efficiency of the implemented parallel simulator with respect to scalability and speedup. Keywords: porous media, multipoint flux approximation, domain decomposition, parallelization

  7. Genetic Parallel Programming: design and implementation.

    PubMed

    Cheang, Sin Man; Leung, Kwong Sak; Lee, Kin Hong

    2006-01-01

    This paper presents a novel Genetic Parallel Programming (GPP) paradigm for evolving parallel programs running on a Multi-Arithmetic-Logic-Unit (Multi-ALU) Processor (MAP). The MAP is a Multiple Instruction-streams, Multiple Data-streams (MIMD), general-purpose register machine that can be implemented on modern Very Large-Scale Integrated Circuits (VLSIs) in order to evaluate genetic programs at high speed. For human programmers, writing parallel programs is more difficult than writing sequential programs. However, experimental results show that GPP evolves parallel programs with less computational effort than that of their sequential counterparts. It creates a new approach to evolving a feasible problem solution in parallel program form and then serializes it into a sequential program if required. The effectiveness and efficiency of GPP are investigated using a suite of 14 well-studied benchmark problems. Experimental results show that GPP speeds up evolution substantially.

  8. Parallel hypergraph partitioning for scientific computing.

    SciTech Connect

    Heaphy, Robert; Devine, Karen Dragon; Catalyurek, Umit; Bisseling, Robert; Hendrickson, Bruce Alan; Boman, Erik Gunnar

    2005-07-01

    Graph partitioning is often used for load balancing in parallel computing, but it is known that hypergraph partitioning has several advantages. First, hypergraphs more accurately model communication volume, and second, they are more expressive and can better represent nonsymmetric problems. Hypergraph partitioning is particularly suited to parallel sparse matrix-vector multiplication, a common kernel in scientific computing. We present a parallel software package for hypergraph (and sparse matrix) partitioning developed at Sandia National Labs. The algorithm is a variation on multilevel partitioning. Our parallel implementation is novel in that it uses a two-dimensional data distribution among processors. We present empirical results that show our parallel implementation achieves good speedup on several large problems (up to 33 million nonzeros) with up to 64 processors on a Linux cluster.

  9. Broadcasting a message in a parallel computer

    DOEpatents

    Berg, Jeremy E.; Faraj, Ahmad A.

    2011-08-02

    Methods, systems, and products are disclosed for broadcasting a message in a parallel computer. The parallel computer includes a plurality of compute nodes connected together using a data communications network. The data communications network optimized for point to point data communications and is characterized by at least two dimensions. The compute nodes are organized into at least one operational group of compute nodes for collective parallel operations of the parallel computer. One compute node of the operational group assigned to be a logical root. Broadcasting a message in a parallel computer includes: establishing a Hamiltonian path along all of the compute nodes in at least one plane of the data communications network and in the operational group; and broadcasting, by the logical root to the remaining compute nodes, the logical root's message along the established Hamiltonian path.

  10. Sequential bioequivalence approaches for parallel designs.

    PubMed

    Fuglsang, Anders

    2014-05-01

    Regulators in EU, USA and Canada allow the use of two-stage approaches for evaluation of bioequivalence. The purpose of this paper is to evaluate such designs for parallel groups using trial simulations. The methods developed by Diane Potvin and co-workers were adapted to parallel designs. Trials were simulated and evaluated on basis of either equal or unequal variances between treatment groups. Methods B and C of Potvin et al., when adapted for parallel designs, protected well against type I error rate inflation under all of the simulated scenarios. Performance characteristics of the new parallel design methods showed little dependence on the assumption of equality of the test and reference variances. This is the first paper to describe the performance of two-stage approaches for parallel designs used to evaluate bioequivalence. The results may prove useful to sponsors developing formulations where crossover designs for bioequivalence evaluation are undesirable.

  11. Code Parallelization with CAPO: A User Manual

    NASA Technical Reports Server (NTRS)

    Jin, Hao-Qiang; Frumkin, Michael; Yan, Jerry; Biegel, Bryan (Technical Monitor)

    2001-01-01

    A software tool has been developed to assist the parallelization of scientific codes. This tool, CAPO, extends an existing parallelization toolkit, CAPTools developed at the University of Greenwich, to generate OpenMP parallel codes for shared memory architectures. This is an interactive toolkit to transform a serial Fortran application code to an equivalent parallel version of the software - in a small fraction of the time normally required for a manual parallelization. We first discuss the way in which loop types are categorized and how efficient OpenMP directives can be defined and inserted into the existing code using the in-depth interprocedural analysis. The use of the toolkit on a number of application codes ranging from benchmark to real-world application codes is presented. This will demonstrate the great potential of using the toolkit to quickly parallelize serial programs as well as the good performance achievable on a large number of toolkit to quickly parallelize serial programs as well as the good performance achievable on a large number of processors. The second part of the document gives references to the parameters and the graphic user interface implemented in the toolkit. Finally a set of tutorials is included for hands-on experiences with this toolkit.

  12. Automatic Generation of Directive-Based Parallel Programs for Shared Memory Parallel Systems

    NASA Technical Reports Server (NTRS)

    Jin, Hao-Qiang; Yan, Jerry; Frumkin, Michael

    2000-01-01

    The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue of portability. Due to its ease of programming and its good performance, the technique has become very popular. In this study, we have extended CAPTools, a computer-aided parallelization toolkit, to automatically generate directive-based, OpenMP, parallel programs. We outline techniques used in the implementation of the tool and present test results on the NAS parallel benchmarks and ARC3D, a CFD application. This work demonstrates the great potential of using computer-aided tools to quickly port parallel programs and also achieve good performance.

  13. Xyce parallel electronic simulator : users' guide.

    SciTech Connect

    Mei, Ting; Rankin, Eric Lamont; Thornquist, Heidi K.; Santarelli, Keith R.; Fixel, Deborah A.; Coffey, Todd Stirling; Russo, Thomas V.; Schiek, Richard Louis; Warrender, Christina E.; Keiter, Eric Richard; Pawlowski, Roger Patrick

    2011-05-01

    This manual describes the use of the Xyce Parallel Electronic Simulator. Xyce has been designed as a SPICE-compatible, high-performance analog circuit simulator, and has been written to support the simulation needs of the Sandia National Laboratories electrical designers. This development has focused on improving capability over the current state-of-the-art in the following areas: (1) Capability to solve extremely large circuit problems by supporting large-scale parallel computing platforms (up to thousands of processors). Note that this includes support for most popular parallel and serial computers; (2) Improved performance for all numerical kernels (e.g., time integrator, nonlinear and linear solvers) through state-of-the-art algorithms and novel techniques. (3) Device models which are specifically tailored to meet Sandia's needs, including some radiation-aware devices (for Sandia users only); and (4) Object-oriented code design and implementation using modern coding practices that ensure that the Xyce Parallel Electronic Simulator will be maintainable and extensible far into the future. Xyce is a parallel code in the most general sense of the phrase - a message passing parallel implementation - which allows it to run efficiently on the widest possible number of computing platforms. These include serial, shared-memory and distributed-memory parallel as well as heterogeneous platforms. Careful attention has been paid to the specific nature of circuit-simulation problems to ensure that optimal parallel efficiency is achieved as the number of processors grows. The development of Xyce provides a platform for computational research and development aimed specifically at the needs of the Laboratory. With Xyce, Sandia has an 'in-house' capability with which both new electrical (e.g., device model development) and algorithmic (e.g., faster time-integration methods, parallel solver algorithms) research and development can be performed. As a result, Xyce is a unique

  14. Parallelization of the Implicit RPLUS Algorithm

    NASA Technical Reports Server (NTRS)

    Orkwis, Paul D.

    1997-01-01

    The multiblock reacting Navier-Stokes flow solver RPLUS2D was modified for parallel implementation. Results for non-reacting flow calculations of this code indicate parallelization efficiencies greater than 84% are possible for a typical test problem. Results tend to improve as the size of the problem increases. The convergence rate of the scheme is degraded slightly when additional artificial block boundaries are included for the purpose of parallelization. However, this degradation virtually disappears if the solution is converged near to machine zero. Recommendations are made for further code improvements to increase efficiency, correct bugs in the original version, and study decomposition effectiveness.

  15. Parallelization of the Implicit RPLUS Algorithm

    NASA Technical Reports Server (NTRS)

    Orkwis, Paul D.

    1994-01-01

    The multiblock reacting Navier-Stokes flow-solver RPLUS2D was modified for parallel implementation. Results for non-reacting flow calculations of this code indicate parallelization efficiencies greater than 84% are possible for a typical test problem. Results tend to improve as the size of the problem increases. The convergence rate of the scheme is degraded slightly when additional artificial block boundaries are included for the purpose of parallelization. However, this degradation virtually disappears if the solution is converged near to machine zero. Recommendations are made for further code improvements to increase efficiency, correct bugs in the original version, and study decomposition effectiveness.

  16. Knowledge representation into Ada parallel processing

    NASA Technical Reports Server (NTRS)

    Masotto, Tom; Babikyan, Carol; Harper, Richard

    1990-01-01

    The Knowledge Representation into Ada Parallel Processing project is a joint NASA and Air Force funded project to demonstrate the execution of intelligent systems in Ada on the Charles Stark Draper Laboratory fault-tolerant parallel processor (FTPP). Two applications were demonstrated - a portion of the adaptive tactical navigator and a real time controller. Both systems are implemented as Activation Framework Objects on the Activation Framework intelligent scheduling mechanism developed by Worcester Polytechnic Institute. The implementations, results of performance analyses showing speedup due to parallelism and initial efficiency improvements are detailed and further areas for performance improvements are suggested.

  17. Time-parallel multiscale/multiphysics framework

    SciTech Connect

    Frantziskonis, G.; Muralidharan, Krishna; Deymier, Pierre; Simunovic, Srdjan; Nukala, Phani K; Pannala, Sreekanth

    2009-01-01

    We introduce the time-parallel compound wavelet matrix method (tpCWM) for modeling the temporal evolution of multiscale and multiphysics systems. The method couples time parallel (TP) and CWM methods operating at different spatial and temporal scales. We demonstrate the efficiency of our approach on two examples: a chemical reaction kinetic system and a non-linear predator prey system. Our results indicate that the tpCWM technique is capable of accelerating time-to-solution by 2 3-orders of magnitude and is amenable to efficient parallel implementation.

  18. Language constructs for modular parallel programs

    SciTech Connect

    Foster, I.

    1996-03-01

    We describe programming language constructs that facilitate the application of modular design techniques in parallel programming. These constructs allow us to isolate resource management and processor scheduling decisions from the specification of individual modules, which can themselves encapsulate design decisions concerned with concurrence, communication, process mapping, and data distribution. This approach permits development of libraries of reusable parallel program components and the reuse of these components in different contexts. In particular, alternative mapping strategies can be explored without modifying other aspects of program logic. We describe how these constructs are incorporated in two practical parallel programming languages, PCN and Fortran M. Compilers have been developed for both languages, allowing experimentation in substantial applications.

  19. Distributed parallel messaging for multiprocessor systems

    SciTech Connect

    Chen, Dong; Heidelberger, Philip; Salapura, Valentina; Senger, Robert M; Steinmacher-Burrow, Burhard; Sugawara, Yutaka

    2013-06-04

    A method and apparatus for distributed parallel messaging in a parallel computing system. The apparatus includes, at each node of a multiprocessor network, multiple injection messaging engine units and reception messaging engine units, each implementing a DMA engine and each supporting both multiple packet injection into and multiple reception from a network, in parallel. The reception side of the messaging unit (MU) includes a switch interface enabling writing of data of a packet received from the network to the memory system. The transmission side of the messaging unit, includes switch interface for reading from the memory system when injecting packets into the network.

  20. Parallel path aspects of transmission modeling

    SciTech Connect

    Kavicky, J.A.; Shahidehpour, S.M.

    1996-11-01

    This paper examines the present methods and modeling techniques available to address the effects of parallel flows resulting from various firm and short-term energy transactions. A survey of significant methodologies is conducted to determine the present status of parallel flow transaction modeling. The strengths and weaknesses of these approaches are identified to suggest areas of further modeling improvements. The motivating force behind this research is to improve transfer capability assessment accuracy by suggesting a real-time modeling environment that adequately represents the influences of parallel flows while recognizing operational constraints and objectives.

  1. Fast combinatorial optimization with parallel digital computers.

    PubMed

    Kakeya, H; Okabe, Y

    2000-01-01

    This paper presents an algorithm which realizes fast search for the solutions of combinatorial optimization problems with parallel digital computers.With the standard weight matrices designed for combinatorial optimization, many iterations are required before convergence to a quasioptimal solution even when many digital processors can be used in parallel. By removing the components of the eingenvectors with eminent negative eigenvalues of the weight matrix, the proposed algorithm avoids oscillation and realizes energy reduction under synchronous discrete dynamics, which enables parallel digital computers to obtain quasi-optimal solutions with much less time than the conventional algorithm.

  2. Synchronization Of Parallel Discrete Event Simulations

    NASA Technical Reports Server (NTRS)

    Steinman, Jeffrey S.

    1992-01-01

    Adaptive, parallel, discrete-event-simulation-synchronization algorithm, Breathing Time Buckets, developed in Synchronous Parallel Environment for Emulation and Discrete Event Simulation (SPEEDES) operating system. Algorithm allows parallel simulations to process events optimistically in fluctuating time cycles that naturally adapt while simulation in progress. Combines best of optimistic and conservative synchronization strategies while avoiding major disadvantages. Algorithm processes events optimistically in time cycles adapting while simulation in progress. Well suited for modeling communication networks, for large-scale war games, for simulated flights of aircraft, for simulations of computer equipment, for mathematical modeling, for interactive engineering simulations, and for depictions of flows of information.

  3. Heterogeneous parallel programming capability. Final report

    SciTech Connect

    Flower, J.W.; Kolawa, A.

    1990-11-30

    In creating a heterogeneous parallel processing capability we are really trying to approach three basic problems with current systems: (1) Supercomputer and parallel computer hardware architectures vary widely but need to support one or two fairly standard programming languages and programming models. A particularly important issue concerns the short life cycle of individual hardware designs; (2) Many algorithms require capabilities beyond the reach of single superconducters but could be approached by several machines working together; and (3) Performing a given task requires integration of a system that may contain many components in addition to the super or parallel computer itself. Peripherals from many different manufacturers must be incorporated.

  4. Social Problems and Deviance: Some Parallel Issues

    ERIC Educational Resources Information Center

    Kitsuse, John I.; Spector, Malcolm

    1975-01-01

    Explores parallel developments in labeling theory and in the value conflict approach to social problems. Similarities in their critiques of functionalism and etiological theory as well as their emphasis on the definitional process are noted. (Author)

  5. Visualization and Tracking of Parallel CFD Simulations

    NASA Technical Reports Server (NTRS)

    Vaziri, Arsi; Kremenetsky, Mark

    1995-01-01

    We describe a system for interactive visualization and tracking of a 3-D unsteady computational fluid dynamics (CFD) simulation on a parallel computer. CM/AVS, a distributed, parallel implementation of a visualization environment (AVS) runs on the CM-5 parallel supercomputer. A CFD solver is run as a CM/AVS module on the CM-5. Data communication between the solver, other parallel visualization modules, and a graphics workstation, which is running AVS, are handled by CM/AVS. Partitioning of the visualization task, between CM-5 and the workstation, can be done interactively in the visual programming environment provided by AVS. Flow solver parameters can also be altered by programmable interactive widgets. This system partially removes the requirement of storing large solution files at frequent time steps, a characteristic of the traditional 'simulate (yields) store (yields) visualize' post-processing approach.

  6. Data parallel sorting for particle simulation

    NASA Technical Reports Server (NTRS)

    Dagum, Leonardo

    1992-01-01

    Sorting on a parallel architecture is a communications intensive event which can incur a high penalty in applications where it is required. In the case of particle simulation, only integer sorting is necessary, and sequential implementations easily attain the minimum performance bound of O (N) for N particles. Parallel implementations, however, have to cope with the parallel sorting problem which, in addition to incurring a heavy communications cost, can make the minimun performance bound difficult to attain. This paper demonstrates how the sorting problem in a particle simulation can be reduced to a merging problem, and describes an efficient data parallel algorithm to solve this merging problem in a particle simulation. The new algorithm is shown to be optimal under conditions usual for particle simulation, and its fieldwise implementation on the Connection Machine is analyzed in detail. The new algorithm is about four times faster than a fieldwise implementation of radix sort on the Connection Machine.

  7. Runtime support for parallelizing data mining algorithms

    NASA Astrophysics Data System (ADS)

    Jin, Ruoming; Agrawal, Gagan

    2002-03-01

    With recent technological advances, shared memory parallel machines have become more scalable, and offer large main memories and high bus bandwidths. They are emerging as good platforms for data warehousing and data mining. In this paper, we focus on shared memory parallelization of data mining algorithms. We have developed a series of techniques for parallelization of data mining algorithms, including full replication, full locking, fixed locking, optimized full locking, and cache-sensitive locking. Unlike previous work on shared memory parallelization of specific data mining algorithms, all of our techniques apply to a large number of common data mining algorithms. In addition, we propose a reduction-object based interface for specifying a data mining algorithm. We show how our runtime system can apply any of the technique we have developed starting from a common specification of the algorithm.

  8. Parallel processing of a rotating shaft simulation

    NASA Technical Reports Server (NTRS)

    Arpasi, Dale J.

    1989-01-01

    A FORTRAN program describing the vibration modes of a rotor-bearing system is analyzed for parellelism in this simulation using a Pascal-like structured language. Potential vector operations are also identified. A critical path through the simulation is identified and used in conjunction with somewhat fictitious processor characteristics to determine the time to calculate the problem on a parallel processing system having those characteristics. A parallel processing overhead time is included as a parameter for proper evaluation of the gain over serial calculation. The serial calculation time is determined for the same fictitious system. An improvement of up to 640 percent is possible depending on the value of the overhead time. Based on the analysis, certain conclusions are drawn pertaining to the development needs of parallel processing technology, and to the specification of parallel processing systems to meet computational needs.

  9. NAS Parallel Benchmarks, Multi-Zone Versions

    NASA Technical Reports Server (NTRS)

    vanderWijngaart, Rob F.; Haopiang, Jin

    2003-01-01

    We describe an extension of the NAS Parallel Benchmarks (NPB) suite that involves solving the application benchmarks LU, BT and SP on collections of loosely coupled discretization meshes. The solutions on the meshes are updated independently, but after each time step they exchange boundary value information. This strategy, which is common among structured-mesh production flow solver codes in use at NASA Ames and elsewhere, provides relatively easily exploitable coarse-grain parallelism between meshes. Since the individual application benchmarks also allow fine-grain parallelism themselves, this NPB extension, named NPB Multi-Zone (NPB-MZ), is a good candidate for testing hybrid and multi-level parallelization tools and strategies.

  10. Asynchronous parallel pattern search for nonlinear optimization

    SciTech Connect

    P. D. Hough; T. G. Kolda; V. J. Torczon

    2000-01-01

    Parallel pattern search (PPS) can be quite useful for engineering optimization problems characterized by a small number of variables (say 10--50) and by expensive objective function evaluations such as complex simulations that take from minutes to hours to run. However, PPS, which was originally designed for execution on homogeneous and tightly-coupled parallel machine, is not well suited to the more heterogeneous, loosely-coupled, and even fault-prone parallel systems available today. Specifically, PPS is hindered by synchronization penalties and cannot recover in the event of a failure. The authors introduce a new asynchronous and fault tolerant parallel pattern search (AAPS) method and demonstrate its effectiveness on both simple test problems as well as some engineering optimization problems

  11. Parallel programming with PCN. Revision 1

    SciTech Connect

    Foster, I.; Tuecke, S.

    1991-12-01

    PCN is a system for developing and executing parallel programs. It comprises a high-level programming language, tools for developing and debugging programs in this language, and interfaces to Fortran and C that allow the reuse of existing code in multilingual parallel programs. Programs developed using PCN are portable across many different workstations, networks, and parallel computers. This document provides all the information required to develop parallel programs with the PCN programming system. In includes both tutorial and reference material. It also presents the basic concepts that underly PCN, particularly where these are likely to be unfamiliar to the reader, and provides pointers to other documentation on the PCN language, programming techniques, and tools. PCN is in the public domain. The latest version of both the software and this manual can be obtained by anonymous FTP from Argonne National Laboratory in the directory pub/pcn at info.mcs.anl.gov (c.f. Appendix A).

  12. A Nomograph for Resistors in Parallel

    NASA Astrophysics Data System (ADS)

    Greenslade, Thomas B.

    2002-11-01

    The author has a large collection of 19th century textbooks, and found a quotation in one of them concerning a construction for finding the equivalent resistance of two resistors in parallel. This note discusses the equations.

  13. Parallel Implementation of the Discontinuous Galerkin Method

    NASA Technical Reports Server (NTRS)

    Baggag, Abdalkader; Atkins, Harold; Keyes, David

    1999-01-01

    This paper describes a parallel implementation of the discontinuous Galerkin method. Discontinuous Galerkin is a spatially compact method that retains its accuracy and robustness on non-smooth unstructured grids and is well suited for time dependent simulations. Several parallelization approaches are studied and evaluated. The most natural and symmetric of the approaches has been implemented in all object-oriented code used to simulate aeroacoustic scattering. The parallel implementation is MPI-based and has been tested on various parallel platforms such as the SGI Origin, IBM SP2, and clusters of SGI and Sun workstations. The scalability results presented for the SGI Origin show slightly superlinear speedup on a fixed-size problem due to cache effects.

  14. Feature Clustering for Accelerating Parallel Coordinate Descent

    SciTech Connect

    Scherrer, Chad; Tewari, Ambuj; Halappanavar, Mahantesh; Haglin, David J.

    2012-12-06

    We demonstrate an approach for accelerating calculation of the regularization path for L1 sparse logistic regression problems. We show the benefit of feature clustering as a preconditioning step for parallel block-greedy coordinate descent algorithms.

  15. Parallel algorithms for dynamically partitioning unstructured grids

    SciTech Connect

    Diniz, P.; Plimpton, S.; Hendrickson, B.; Leland, R.

    1994-10-01

    Grid partitioning is the method of choice for decomposing a wide variety of computational problems into naturally parallel pieces. In problems where computational load on the grid or the grid itself changes as the simulation progresses, the ability to repartition dynamically and in parallel is attractive for achieving higher performance. We describe three algorithms suitable for parallel dynamic load-balancing which attempt to partition unstructured grids so that computational load is balanced and communication is minimized. The execution time of algorithms and the quality of the partitions they generate are compared to results from serial partitioners for two large grids. The integration of the algorithms into a parallel particle simulation is also briefly discussed.

  16. The PISCES 2 parallel programming environment

    NASA Technical Reports Server (NTRS)

    Pratt, Terrence W.

    1987-01-01

    PISCES 2 is a programming environment for scientific and engineering computations on MIMD parallel computers. It is currently implemented on a flexible FLEX/32 at NASA Langley, a 20 processor machine with both shared and local memories. The environment provides an extended Fortran for applications programming, a configuration environment for setting up a run on the parallel machine, and a run-time environment for monitoring and controlling program execution. This paper describes the overall design of the system and its implementation on the FLEX/32. Emphasis is placed on several novel aspects of the design: the use of a carefully defined virtual machine, programmer control of the mapping of virtual machine to actual hardware, forces for medium-granularity parallelism, and windows for parallel distribution of data. Some preliminary measurements of storage use are included.

  17. Massively Parallel Computing: A Sandia Perspective

    SciTech Connect

    Dosanjh, Sudip S.; Greenberg, David S.; Hendrickson, Bruce; Heroux, Michael A.; Plimpton, Steve J.; Tomkins, James L.; Womble, David E.

    1999-05-06

    The computing power available to scientists and engineers has increased dramatically in the past decade, due in part to progress in making massively parallel computing practical and available. The expectation for these machines has been great. The reality is that progress has been slower than expected. Nevertheless, massively parallel computing is beginning to realize its potential for enabling significant break-throughs in science and engineering. This paper provides a perspective on the state of the field, colored by the authors' experiences using large scale parallel machines at Sandia National Laboratories. We address trends in hardware, system software and algorithms, and we also offer our view of the forces shaping the parallel computing industry.

  18. Improved chopper circuit uses parallel transistors

    NASA Technical Reports Server (NTRS)

    1966-01-01

    Parallel transistor chopper circuit operates with one transistor in the forward mode and the other in the inverse mode. By using this method, it acts as a single, symmetrical, bidirectional transistor, and reduces and stabilizes the offset voltage.

  19. Molecular dynamics on hypercube parallel computers

    NASA Astrophysics Data System (ADS)

    Smith, W.

    1991-03-01

    The implementation of molecular dynamics on parallel computers is described, with particular reference to hypercube computers. Three particular algorithms are described: replicated data (RD); systolic loop (SLS-G), and parallelised link-cells (PLC), all of which have good load balancing. The performance characteristics of each algorithm and the factors affecting their scaling properties are discussed. The article is pedagogic in intent, to introduce a novice to the main aspects of parallel computing in molecular dynamics.

  20. Graphics-Based Parallel Programming Tools

    DTIC Science & Technology

    1991-09-01

    more general context by implementing perspective views within the Voyeur system [121. Voyeur is a more conventional tool for displaying application...Varadaraju. Interfacing Belvedere with Voyeur . Master’s Thesis, COINS Department, University of Massachusetts (June 1991). 13 David Socha and Mary L...Bailey and David Notkin, " Voyeur : Graphi- cal Views of Parallel Programs", SIGPLAN Workshop on Parallel and Distributed Debugging, pp. 206-215 (1988). 14

  1. Graphics-Based Parallel Programming Tools

    DTIC Science & Technology

    1992-01-01

    the Voyeur system [121. Voyeur is a more conventional tool for displaying application-specific visualizations of parallel programs [131 and it provides...Department. University of Massachusetts (June 1991). 13 David Socha and Mary L. Bailey and David Notkin. "’ Voyeur : Graphi- cal Views of Parallel Programs...Massachusetts (September 1991). Nandakumar Varadaraju. Interfacing Belvedere with Voyeur . Master’s The- sis. COINS Department. University of Massachusetts

  2. Voyeur: Graphical Views of Parallel Programs

    DTIC Science & Technology

    1988-04-01

    visualization, parallel debugging, moni torino 20. ABSTRACT (Continue ci reveree, eide if necessary and Identify by block number) ~9 Voyeur is a prototype...PAGE Dh- eaa,:~ Voyeur : Graphical Views of Parallel Programs David Socha, Mary Bailey and David Notkin Department of Computer Science, FR-35 University...of Washington Seattle, Washington 98195 TR 88-04-03 April 1988 Voyeur is a prototype system that facilitates the construction of application-specific

  3. Partitioning And Packing Equations For Parallel Processing

    NASA Technical Reports Server (NTRS)

    Arpasi, Dale J.; Milner, Edward J.

    1989-01-01

    Algorithm developed to identify parallelism in set of coupled ordinary differential equations that describe physical system and to divide set into parallel computational paths, along with parts of solution proceeds independently of others during at least part of time. Path-identifying algorithm creates number of paths consisting of equations that must be computed serially and table that gives dependent and independent arguments and "can start," "can end," and "must end" times of each equation. "Must end" time used subsequently by packing algorithm.

  4. Enhancing Scalability of Parallel Structured AMR Calculations

    SciTech Connect

    Wissink, A M; Hysom, D; Hornung, R D

    2003-02-10

    This paper discusses parallel scaling performance of large scale parallel structured adaptive mesh refinement (SAMR) calculations in SAMRAI. Previous work revealed that poor scaling qualities in the adaptive gridding operations in SAMR calculations cause them to become dominant for cases run on up to 512 processors. This work describes algorithms we have developed to enhance the efficiency of the adaptive gridding operations. Performance of the algorithms is evaluated for two adaptive benchmarks run on up 512 processors of an IBM SP system.

  5. Joint Experimentation on Scalable Parallel Processors (JESPP)

    DTIC Science & Technology

    2006-04-01

    SCALABLE PARALLEL PROCESSORS (JESPP) 6. AUTHOR(S) Dan M. Davis, Robert F. Lucas, Ke-Thia Yao, Gene Wagenbreth 5. FUNDING NUMBERS C...List of Papers • Robert J. Graebener, Gregory Rafuse, Robert Miller & Ke-Thia Yao, “The Road to Successful Joint Experimentation Starts at the...2003. • Robert F. Lucas & Dan M. Davis, “Joint Experimentation on Scalable Parallel Processors“, Interservice/Industry Training, Simulation, and

  6. LDV Measurement of Confined Parallel Jet Mixing

    SciTech Connect

    R.F. Kunz; S.W. D'Amico; P.F. Vassallo; M.A. Zaccaria

    2001-01-31

    Laser Doppler Velocimetry (LDV) measurements were taken in a confinement, bounded by two parallel walls, into which issues a row of parallel jets. Two-component measurements were taken of two mean velocity components and three Reynolds stress components. As observed in isolated three dimensional wall bounded jets, the transverse diffusion of the jets is quite large. The data indicate that this rapid mixing process is due to strong secondary flows, transport of large inlet intensities and Reynolds stress anisotropy effects.

  7. A survey of parallel programming tools

    NASA Technical Reports Server (NTRS)

    Cheng, Doreen Y.

    1991-01-01

    This survey examines 39 parallel programming tools. Focus is placed on those tool capabilites needed for parallel scientific programming rather than for general computer science. The tools are classified with current and future needs of Numerical Aerodynamic Simulator (NAS) in mind: existing and anticipated NAS supercomputers and workstations; operating systems; programming languages; and applications. They are divided into four categories: suggested acquisitions, tools already brought in; tools worth tracking; and tools eliminated from further consideration at this time.

  8. HOPSPACK: Hybrid Optimization Parallel Search Package.

    SciTech Connect

    Gray, Genetha Anne.; Kolda, Tamara G.; Griffin, Joshua; Taddy, Matt; Martinez-Canales, Monica L.

    2008-12-01

    In this paper, we describe the technical details of HOPSPACK (Hybrid Optimization Parallel SearchPackage), a new software platform which facilitates combining multiple optimization routines into asingle, tightly-coupled, hybrid algorithm that supports parallel function evaluations. The frameworkis designed such that existing optimization source code can be easily incorporated with minimalcode modification. By maintaining the integrity of each individual solver, the strengths and codesophistication of the original optimization package are retained and exploited.4

  9. Parallel algorithms for contour extraction and coding

    NASA Astrophysics Data System (ADS)

    Dinstein, Its'hak; Landau, Gad M.

    1990-07-01

    A parallel approach to contour extraction and coding on an Exclusive Read Exclusive Write (EREW) Parallel Random Access Machine (PRAM) is presented and analyzed. The algorithm is intended for binary images. The labeled contours can be represented by lists of coordinates, and/or chain codes, and/or any other user designed codes. Using O(n2/log n) processors, the algorithm runs in O(logn) time, where n by n is the size of the processed binary image.

  10. Parallel and Distributed Computing Combinatorial Algorithms

    DTIC Science & Technology

    1993-10-01

    FUPNDKC %2,•, PARALLEL AND DISTRIBUTED COMPUTING COMBINATORIAL ALGORITHMS 6. AUTHOR(S) 2304/DS F49620-92-J-0125 DR. LEIGHTON 7 PERFORMING ORGANIZATION NAME...on several problems involving parallel and distributed computing and combinatorial optimization. This research is reported in the numerous papers that...network decom- position. In Proceedings of the Eleventh Annual ACM Symposium on Principles of Distributed Computing , August 1992. [15] B. Awerbuch, B

  11. Computational electromagnetics and parallel dense matrix computations

    SciTech Connect

    Forsman, K.; Kettunen, L.; Gropp, W.; Levine, D.

    1995-06-01

    We present computational results using CORAL, a parallel, three-dimensional, nonlinear magnetostatic code based on a volume integral equation formulation. A key feature of CORAL is the ability to solve, in parallel, the large, dense systems of linear equations that are inherent in the use of integral equation methods. Using the Chameleon and PSLES libraries ensures portability and access to the latest linear algebra solution technology.

  12. Fast Parallel Computation Of Multibody Dynamics

    NASA Technical Reports Server (NTRS)

    Fijany, Amir; Kwan, Gregory L.; Bagherzadeh, Nader

    1996-01-01

    Constraint-force algorithm fast, efficient, parallel-computation algorithm for solving forward dynamics problem of multibody system like robot arm or vehicle. Solves problem in minimum time proportional to log(N) by use of optimal number of processors proportional to N, where N is number of dynamical degrees of freedom: in this sense, constraint-force algorithm both time-optimal and processor-optimal parallel-processing algorithm.

  13. Parallel computer methods for eigenvalue extraction

    NASA Technical Reports Server (NTRS)

    Akl, Fred

    1988-01-01

    A new numerical algorithm for the solution of large-order eigenproblems typically encountered in linear elastic finite element systems is presented. The architecture of parallel processing is used in the algorithm to achieve increased speed and efficiency of calculations. The algorithm is based on the frontal technique for the solution of linear simultaneous equations and the modified subspace eigenanalysis method for the solution of the eigenproblem. The advantages of this new algorithm in parallel computer architecture are discussed.

  14. Parallel algorithms for unconstrained optimizations by multisplitting

    SciTech Connect

    He, Qing

    1994-12-31

    In this paper a new parallel iterative algorithm for unconstrained optimization using the idea of multisplitting is proposed. This algorithm uses the existing sequential algorithms without any parallelization. Some convergence and numerical results for this algorithm are presented. The experiments are performed on an Intel iPSC/860 Hyper Cube with 64 nodes. It is interesting that the sequential implementation on one node shows that if the problem is split properly, the algorithm converges much faster than one without splitting.

  15. Computational electromagnetics and parallel dense matrix computations

    SciTech Connect

    Forsman, K.; Kettunen, L.; Gropp, W.

    1995-12-01

    We present computational results using CORAL, a parallel, three-dimensional, nonlinear magnetostatic code based on a volume integral equation formulation. A key feature of CORAL is the ability to solve, in parallel, the large, dense systems of linear equations that are inherent in the use of integral equation methods. Using the Chameleon and PSLES libraries ensures portability and access to the latest linear algebra solution technology.

  16. A parallel PCG solver for MODFLOW.

    PubMed

    Dong, Yanhui; Li, Guomin

    2009-01-01

    In order to simulate large-scale ground water flow problems more efficiently with MODFLOW, the OpenMP programming paradigm was used to parallelize the preconditioned conjugate-gradient (PCG) solver with in this study. Incremental parallelization, the significant advantage supported by OpenMP on a shared-memory computer, made the solver transit to a parallel program smoothly one block of code at a time. The parallel PCG solver, suitable for both MODFLOW-2000 and MODFLOW-2005, is verified using an 8-processor computer. Both the impact of compilers and different model domain sizes were considered in the numerical experiments. Based on the timing results, execution times using the parallel PCG solver are typically about 1.40 to 5.31 times faster than those using the serial one. In addition, the simulation results are the exact same as the original PCG solver, because the majority of serial codes were not changed. It is worth noting that this parallelizing approach reduces cost in terms of software maintenance because only a single source PCG solver code needs to be maintained in the MODFLOW source tree.

  17. The Xyce Parallel Electronic Simulator - An Overview

    SciTech Connect

    HUTCHINSON,SCOTT A.; KEITER,ERIC R.; HOEKSTRA,ROBERT J.; WATTS,HERMAN A.; WATERS,ARLON J.; SCHELLS,REGINA L.; WIX,STEVEN D.

    2000-12-08

    The Xyce{trademark} Parallel Electronic Simulator has been written to support the simulation needs of the Sandia National Laboratories electrical designers. As such, the development has focused on providing the capability to solve extremely large circuit problems by supporting large-scale parallel computing platforms (up to thousands of processors). In addition, they are providing improved performance for numerical kernels using state-of-the-art algorithms, support for modeling circuit phenomena at a variety of abstraction levels and using object-oriented and modern coding-practices that ensure the code will be maintainable and extensible far into the future. The code is a parallel code in the most general sense of the phrase--a message passing parallel implementation--which allows it to run efficiently on the widest possible number of computing platforms. These include serial, shared-memory and distributed-memory parallel as well as heterogeneous platforms. Furthermore, careful attention has been paid to the specific nature of circuit-simulation problems to ensure that optimal parallel efficiency is achieved even as the number of processors grows.

  18. Parallel adaptive wavelet collocation method for PDEs

    SciTech Connect

    Nejadmalayeri, Alireza; Vezolainen, Alexei; Brown-Dymkoski, Eric; Vasilyev, Oleg V.

    2015-10-01

    A parallel adaptive wavelet collocation method for solving a large class of Partial Differential Equations is presented. The parallelization is achieved by developing an asynchronous parallel wavelet transform, which allows one to perform parallel wavelet transform and derivative calculations with only one data synchronization at the highest level of resolution. The data are stored using tree-like structure with tree roots starting at a priori defined level of resolution. Both static and dynamic domain partitioning approaches are developed. For the dynamic domain partitioning, trees are considered to be the minimum quanta of data to be migrated between the processes. This allows fully automated and efficient handling of non-simply connected partitioning of a computational domain. Dynamic load balancing is achieved via domain repartitioning during the grid adaptation step and reassigning trees to the appropriate processes to ensure approximately the same number of grid points on each process. The parallel efficiency of the approach is discussed based on parallel adaptive wavelet-based Coherent Vortex Simulations of homogeneous turbulence with linear forcing at effective non-adaptive resolutions up to 2048{sup 3} using as many as 2048 CPU cores.

  19. Parallel adaptive wavelet collocation method for PDEs

    NASA Astrophysics Data System (ADS)

    Nejadmalayeri, Alireza; Vezolainen, Alexei; Brown-Dymkoski, Eric; Vasilyev, Oleg V.

    2015-10-01

    A parallel adaptive wavelet collocation method for solving a large class of Partial Differential Equations is presented. The parallelization is achieved by developing an asynchronous parallel wavelet transform, which allows one to perform parallel wavelet transform and derivative calculations with only one data synchronization at the highest level of resolution. The data are stored using tree-like structure with tree roots starting at a priori defined level of resolution. Both static and dynamic domain partitioning approaches are developed. For the dynamic domain partitioning, trees are considered to be the minimum quanta of data to be migrated between the processes. This allows fully automated and efficient handling of non-simply connected partitioning of a computational domain. Dynamic load balancing is achieved via domain repartitioning during the grid adaptation step and reassigning trees to the appropriate processes to ensure approximately the same number of grid points on each process. The parallel efficiency of the approach is discussed based on parallel adaptive wavelet-based Coherent Vortex Simulations of homogeneous turbulence with linear forcing at effective non-adaptive resolutions up to 20483 using as many as 2048 CPU cores.

  20. Integrating GIS-based geologic mapping, LiDAR-based lineament analysis and site specific rock slope data to delineate a zone of existing and potential rock slope instability located along the grandfather mountain window-Linville Falls shear zone contact, Southern Appalachian Mountains, Watauga County, North Carolina

    USGS Publications Warehouse

    Gillon, K.A.; Wooten, R.M.; Latham, R.L.; Witt, A.W.; Douglas, T.J.; Bauer, J.B.; Fuemmeler, S.J.

    2009-01-01

    Landslide hazard maps of Watauga County identify >2200 landslides, model debris flow susceptibility, and evaluate a 14km x 0.5km zone of existing and potential rock slope instability (ZEPRSI) near the Town of Boone. The ZEPRSI encompasses west-northwest trending (WNWT) topographic ridges where 14 active/past-active rock/weathered rock slides occur mainly in rocks of the Grandfather Mountain Window (GMW). The north side of this ridgeline is the GMW / Linville Falls Fault (LFF) contact. Sheared rocks of the Linville Falls Shear Zone (LFSZ) occur along the ridge and locally in the valley north of the contact. The valley is underlain principally by layered granitic gneiss comprising the Linville Falls/Beech Mountain/Stone Mountain Thrust Sheet. The integration of ArcGIS??? - format digital geologic and lineament mapping on a 6m LiDAR (Light Detecting and Ranging) digital elevation model (DEM) base, and kinematic analyses of site specific rock slope data (e.g., presence and degree of ductile and brittle deformation fabrics, rock type, rock weathering state) indicate: WNWT lineaments are expressions of a regionally extensive zone of fractures and faults; and ZEPRSI rock slope failures concentrate along excavated, north-facing LFF/LFSZ slopes where brittle fabrics overprint older metamorphic foliations, and other fractures create side and back release surfaces. Copyright 2009 ARMA, American Rock Mechanics Association.

  1. Start/Pat; A parallel-programming toolkit

    SciTech Connect

    Appelbe, B.; Smith, K. ); McDowell, C. )

    1989-07-01

    How can you make Fortran code parallel without isolating the programmer from learning to understand and exploit parallelism effectively. With an interactive toolkit that automates parallelization as it educates. This paper discusses the Start/Pat toolkit.

  2. Parallel Architectures and Parallel Algorithms for Integrated Vision Systems. Ph.D. Thesis

    NASA Technical Reports Server (NTRS)

    Choudhary, Alok Nidhi

    1989-01-01

    Computer vision is regarded as one of the most complex and computationally intensive problems. An integrated vision system (IVS) is a system that uses vision algorithms from all levels of processing to perform for a high level application (e.g., object recognition). An IVS normally involves algorithms from low level, intermediate level, and high level vision. Designing parallel architectures for vision systems is of tremendous interest to researchers. Several issues are addressed in parallel architectures and parallel algorithms for integrated vision systems.

  3. Parallel phase model : a programming model for high-end parallel machines with manycores.

    SciTech Connect

    Wu, Junfeng; Wen, Zhaofang; Heroux, Michael Allen; Brightwell, Ronald Brian

    2009-04-01

    This paper presents a parallel programming model, Parallel Phase Model (PPM), for next-generation high-end parallel machines based on a distributed memory architecture consisting of a networked cluster of nodes with a large number of cores on each node. PPM has a unified high-level programming abstraction that facilitates the design and implementation of parallel algorithms to exploit both the parallelism of the many cores and the parallelism at the cluster level. The programming abstraction will be suitable for expressing both fine-grained and coarse-grained parallelism. It includes a few high-level parallel programming language constructs that can be added as an extension to an existing (sequential or parallel) programming language such as C; and the implementation of PPM also includes a light-weight runtime library that runs on top of an existing network communication software layer (e.g. MPI). Design philosophy of PPM and details of the programming abstraction are also presented. Several unstructured applications that inherently require high-volume random fine-grained data accesses have been implemented in PPM with very promising results.

  4. PISCES: An environment for parallel scientific computation

    NASA Technical Reports Server (NTRS)

    Pratt, T. W.

    1985-01-01

    The parallel implementation of scientific computing environment (PISCES) is a project to provide high-level programming environments for parallel MIMD computers. Pisces 1, the first of these environments, is a FORTRAN 77 based environment which runs under the UNIX operating system. The Pisces 1 user programs in Pisces FORTRAN, an extension of FORTRAN 77 for parallel processing. The major emphasis in the Pisces 1 design is in providing a carefully specified virtual machine that defines the run-time environment within which Pisces FORTRAN programs are executed. Each implementation then provides the same virtual machine, regardless of differences in the underlying architecture. The design is intended to be portable to a variety of architectures. Currently Pisces 1 is implemented on a network of Apollo workstations and on a DEC VAX uniprocessor via simulation of the task level parallelism. An implementation for the Flexible Computing Corp. FLEX/32 is under construction. An introduction to the Pisces 1 virtual computer and the FORTRAN 77 extensions is presented. An example of an algorithm for the iterative solution of a system of equations is given. The most notable features of the design are the provision for several granularities of parallelism in programs and the provision of a window mechanism for distributed access to large arrays of data.

  5. Equalizer: a scalable parallel rendering framework.

    PubMed

    Eilemann, Stefan; Makhinya, Maxim; Pajarola, Renato

    2009-01-01

    Continuing improvements in CPU and GPU performances as well as increasing multi-core processor and cluster-based parallelism demand for flexible and scalable parallel rendering solutions that can exploit multipipe hardware accelerated graphics. In fact, to achieve interactive visualization, scalable rendering systems are essential to cope with the rapid growth of data sets. However, parallel rendering systems are non-trivial to develop and often only application specific implementations have been proposed. The task of developing a scalable parallel rendering framework is even more difficult if it should be generic to support various types of data and visualization applications, and at the same time work efficiently on a cluster with distributed graphics cards. In this paper we introduce a novel system called Equalizer, a toolkit for scalable parallel rendering based on OpenGL which provides an application programming interface (API) to develop scalable graphics applications for a wide range of systems ranging from large distributed visualization clusters and multi-processor multipipe graphics systems to single-processor single-pipe desktop machines. We describe the system architecture, the basic API, discuss its advantages over previous approaches, present example configurations and usage scenarios as well as scalability results.

  6. A parallel algorithm for random searches

    NASA Astrophysics Data System (ADS)

    Wosniack, M. E.; Raposo, E. P.; Viswanathan, G. M.; da Luz, M. G. E.

    2015-11-01

    We discuss a parallelization procedure for a two-dimensional random search of a single individual, a typical sequential process. To assure the same features of the sequential random search in the parallel version, we analyze the former spatial patterns of the encountered targets for different search strategies and densities of homogeneously distributed targets. We identify a lognormal tendency for the distribution of distances between consecutively detected targets. Then, by assigning the distinct mean and standard deviation of this distribution for each corresponding configuration in the parallel simulations (constituted by parallel random walkers), we are able to recover important statistical properties, e.g., the target detection efficiency, of the original problem. The proposed parallel approach presents a speedup of nearly one order of magnitude compared with the sequential implementation. This algorithm can be easily adapted to different instances, as searches in three dimensions. Its possible range of applicability covers problems in areas as diverse as automated computer searchers in high-capacity databases and animal foraging.

  7. Parallel algorithms for the spectral transform method

    SciTech Connect

    Foster, I.T.; Worley, P.H.

    1997-05-01

    The spectral transform method is a standard numerical technique for solving partial differential equations on a sphere and is widely used in atmospheric circulation models. Recent research has identified several promising algorithms for implementing this method on massively parallel computers; however, no detailed comparison of the different algorithms has previously been attempted. In this paper, the authors describe these different parallel algorithms and report on computational experiments that they have conducted to evaluate their efficiency on parallel computers. The experiments used a testbed code that solves the nonlinear shallow water equations on a sphere; considerable care was taken to ensure that the experiments provide a fair comparison of the different algorithms and that the results are relevant to global models. The authors focus on hypercube- and mesh-connected multicomputers with cut-through routing, such as the Intel iPSC/860, DELTA, and Paragon, and the nCUBE/2, but they also indicate how the results extend to other parallel computer architectures. The results of this study are relevant not only to the spectral transform method but also to multidimensional fast Fourier transforms (FFTs) and other parallel transforms.

  8. Iteration schemes for parallelizing models of superconductivity

    SciTech Connect

    Gray, P.A.

    1996-12-31

    The time dependent Lawrence-Doniach model, valid for high fields and high values of the Ginzburg-Landau parameter, is often used for studying vortex dynamics in layered high-T{sub c} superconductors. When solving these equations numerically, the added degrees of complexity due to the coupling and nonlinearity of the model often warrant the use of high-performance computers for their solution. However, the interdependence between the layers can be manipulated so as to allow parallelization of the computations at an individual layer level. The reduced parallel tasks may then be solved independently using a heterogeneous cluster of networked workstations connected together with Parallel Virtual Machine (PVM) software. Here, this parallelization of the model is discussed and several computational implementations of varying degrees of parallelism are presented. Computational results are also given which contrast properties of convergence speed, stability, and consistency of these implementations. Included in these results are models involving the motion of vortices due to an applied current and pinning effects due to various material properties.

  9. IMPAIR: massively parallel deconvolution on the GPU

    NASA Astrophysics Data System (ADS)

    Sherry, Michael; Shearer, Andy

    2013-02-01

    The IMPAIR software is a high throughput image deconvolution tool for processing large out-of-core datasets of images, varying from large images with spatially varying PSFs to large numbers of images with spatially invariant PSFs. IMPAIR implements a parallel version of the tried and tested Richardson-Lucy deconvolution algorithm regularised via a custom wavelet thresholding library. It exploits the inherently parallel nature of the convolution operation to achieve quality results on consumer grade hardware: through the NVIDIA Tesla GPU implementation, the multi-core OpenMP implementation, and the cluster computing MPI implementation of the software. IMPAIR aims to address the problem of parallel processing in both top-down and bottom-up approaches: by managing the input data at the image level, and by managing the execution at the instruction level. These combined techniques will lead to a scalable solution with minimal resource consumption and maximal load balancing. IMPAIR is being developed as both a stand-alone tool for image processing, and as a library which can be embedded into non-parallel code to transparently provide parallel high throughput deconvolution.

  10. Toward a science of parallel computation

    SciTech Connect

    Worlton, W.J.

    1986-01-01

    The evolution of parallel processing over the past several decades can be viewed as the development of a new scientific discipline. Parallel processing has been, and is, undergoing the same evolutionary stages that are common to the development of scientific disciplines in general: exploration, focusing, and maturity. That parallel processing is not yet a science can readily be appreciated by its lack of some of the characteristics typical of mature sciences, such as prescriptive terminology, comprehensive taxonomies, and authoritative fundamental principles. A great deal of outstanding work has been done and the field is experiencing the beginnings of its ''focusing'' phase, i.e., support is being concentrated in a set of the more promising approaches selected from among the larger set of exploratory projects. However, the possible set of parallel-processing concepts is so extensive that exploratory work will probably continue for one or two more decades. In the meantime, the growing maturity of the field will be reflected in the increasing clarity and precision of the terminology, the development of systematic classification of the domain of discourse, the development of basic principles, and the growing number of commercial products that are the outcome of the research and development projects on which support is being focused. In this paper we develop some generalizations of taxonomies and use basic principles to draw conclusions about the extensibility of parallel processor architectures. 7 refs., 5 figs., 2 tabs.

  11. Linear Bregman algorithm implemented in parallel GPU

    NASA Astrophysics Data System (ADS)

    Li, Pengyan; Ke, Jue; Sui, Dong; Wei, Ping

    2015-08-01

    At present, most compressed sensing (CS) algorithms have poor converging speed, thus are difficult to run on PC. To deal with this issue, we use a parallel GPU, to implement a broadly used compressed sensing algorithm, the Linear Bregman algorithm. Linear iterative Bregman algorithm is a reconstruction algorithm proposed by Osher and Cai. Compared with other CS reconstruction algorithms, the linear Bregman algorithm only involves the vector and matrix multiplication and thresholding operation, and is simpler and more efficient for programming. We use C as a development language and adopt CUDA (Compute Unified Device Architecture) as parallel computing architectures. In this paper, we compared the parallel Bregman algorithm with traditional CPU realized Bregaman algorithm. In addition, we also compared the parallel Bregman algorithm with other CS reconstruction algorithms, such as OMP and TwIST algorithms. Compared with these two algorithms, the result of this paper shows that, the parallel Bregman algorithm needs shorter time, and thus is more convenient for real-time object reconstruction, which is important to people's fast growing demand to information technology.

  12. A feasibility study of multiplexing parallel beam.

    PubMed

    Ma, Jiayi; Zhao, Jingwu; Shi, Xiaodong; Huang, Runshen

    2013-05-01

    Single-photon emission computed tomography (SPECT) is a suitable tool for clinically localizing deep-sited tumors; SPECT with high spatial resolution has the ability to localize deep-sited tumors precisely. However, because of its poor sensitivity, in China SPECT now only plays a complementary role. To improve the sensitivity of the parallel beam collimator mainly used in China, a multiplexing parallel beam collimator is proposed, which can improve sensitivity while maintaining higher spatial resolution by using theoretical prediction and Monte Carlo simulation. The improved sensitivity-to-spatial resolution ratio has an optimal value. In addition, a set of gamma ray channels, introduced only in the transverse direction, did not have any effect in the axial direction. In the transverse direction, the projection data are the sum of the parallel beam and two oblique parallel beams. From visual assessment obtained using computer simulations with equal sensitivity, the reconstructed image at deep-sited was noticeably better than that with the high sensitivity parallel beam.

  13. Parallel algorithms for the spectral transform method

    SciTech Connect

    Foster, I.T.; Worley, P.H.

    1994-04-01

    The spectral transform method is a standard numerical technique for solving partial differential equations on a sphere and is widely used in atmospheric circulation models. Recent research has identified several promising algorithms for implementing this method on massively parallel computers; however, no detailed comparison of the different algorithms has previously been attempted. In this paper, we describe these different parallel algorithms and report on computational experiments that we have conducted to evaluate their efficiency on parallel computers. The experiments used a testbed code that solves the nonlinear shallow water equations or a sphere; considerable care was taken to ensure that the experiments provide a fair comparison of the different algorithms and that the results are relevant to global models. We focus on hypercube- and mesh-connected multicomputers with cut-through routing, such as the Intel iPSC/860, DELTA, and Paragon, and the nCUBE/2, but also indicate how the results extend to other parallel computer architectures. The results of this study are relevant not only to the spectral transform method but also to multidimensional FFTs and other parallel transforms.

  14. Relative Debugging of Automatically Parallelized Programs

    NASA Technical Reports Server (NTRS)

    Jost, Gabriele; Hood, Robert; Biegel, Bryan (Technical Monitor)

    2002-01-01

    We describe a system that simplifies the process of debugging programs produced by computer-aided parallelization tools. The system uses relative debugging techniques to compare serial and parallel executions in order to show where the computations begin to differ. If the original serial code is correct, errors due to parallelization will be isolated by the comparison. One of the primary goals of the system is to minimize the effort required of the user. To that end, the debugging system uses information produced by the parallelization tool to drive the comparison process. In particular, the debugging system relies on the parallelization tool to provide information about where variables may have been modified and how arrays are distributed across multiple processes. User effort is also reduced through the use of dynamic instrumentation. This allows us to modify, the program execution with out changing the way the user builds the executable. The use of dynamic instrumentation also permits us to compare the executions in a fine-grained fashion and only involve the debugger when a difference has been detected. This reduces the overhead of executing instrumentation.

  15. Support for Debugging Automatically Parallelized Programs

    NASA Technical Reports Server (NTRS)

    Jost, Gabriele; Hood, Robert; Biegel, Bryan (Technical Monitor)

    2001-01-01

    We describe a system that simplifies the process of debugging programs produced by computer-aided parallelization tools. The system uses relative debugging techniques to compare serial and parallel executions in order to show where the computations begin to differ. If the original serial code is correct, errors due to parallelization will be isolated by the comparison. One of the primary goals of the system is to minimize the effort required of the user. To that end, the debugging system uses information produced by the parallelization tool to drive the comparison process. In particular the debugging system relies on the parallelization tool to provide information about where variables may have been modified and how arrays are distributed across multiple processes. User effort is also reduced through the use of dynamic instrumentation. This allows us to modify the program execution without changing the way the user builds the executable. The use of dynamic instrumentation also permits us to compare the executions in a fine-grained fashion and only involve the debugger when a difference has been detected. This reduces the overhead of executing instrumentation.

  16. A parallel adaptive mesh refinement algorithm

    NASA Technical Reports Server (NTRS)

    Quirk, James J.; Hanebutte, Ulf R.

    1993-01-01

    Over recent years, Adaptive Mesh Refinement (AMR) algorithms which dynamically match the local resolution of the computational grid to the numerical solution being sought have emerged as powerful tools for solving problems that contain disparate length and time scales. In particular, several workers have demonstrated the effectiveness of employing an adaptive, block-structured hierarchical grid system for simulations of complex shock wave phenomena. Unfortunately, from the parallel algorithm developer's viewpoint, this class of scheme is quite involved; these schemes cannot be distilled down to a small kernel upon which various parallelizing strategies may be tested. However, because of their block-structured nature such schemes are inherently parallel, so all is not lost. In this paper we describe the method by which Quirk's AMR algorithm has been parallelized. This method is built upon just a few simple message passing routines and so it may be implemented across a broad class of MIMD machines. Moreover, the method of parallelization is such that the original serial code is left virtually intact, and so we are left with just a single product to support. The importance of this fact should not be underestimated given the size and complexity of the original algorithm.

  17. Simulating Billion-Task Parallel Programs

    SciTech Connect

    Perumalla, Kalyan S; Park, Alfred J

    2014-01-01

    In simulating large parallel systems, bottom-up approaches exercise detailed hardware models with effects from simplified software models or traces, whereas top-down approaches evaluate the timing and functionality of detailed software models over coarse hardware models. Here, we focus on the top-down approach and significantly advance the scale of the simulated parallel programs. Via the direct execution technique combined with parallel discrete event simulation, we stretch the limits of the top-down approach by simulating message passing interface (MPI) programs with millions of tasks. Using a timing-validated benchmark application, a proof-of-concept scaling level is achieved to over 0.22 billion virtual MPI processes on 216,000 cores of a Cray XT5 supercomputer, representing one of the largest direct execution simulations to date, combined with a multiplexing ratio of 1024 simulated tasks per real task.

  18. A Parallel Rendering Algorithm for MIMD Architectures

    NASA Technical Reports Server (NTRS)

    Crockett, Thomas W.; Orloff, Tobias

    1991-01-01

    Applications such as animation and scientific visualization demand high performance rendering of complex three dimensional scenes. To deliver the necessary rendering rates, highly parallel hardware architectures are required. The challenge is then to design algorithms and software which effectively use the hardware parallelism. A rendering algorithm targeted to distributed memory MIMD architectures is described. For maximum performance, the algorithm exploits both object-level and pixel-level parallelism. The behavior of the algorithm is examined both analytically and experimentally. Its performance for large numbers of processors is found to be limited primarily by communication overheads. An experimental implementation for the Intel iPSC/860 shows increasing performance from 1 to 128 processors across a wide range of scene complexities. It is shown that minimal modifications to the algorithm will adapt it for use on shared memory architectures as well.

  19. Improved CDMA Performance Using Parallel Interference Cancellation

    NASA Technical Reports Server (NTRS)

    Simon, Marvin; Divsalar, Dariush

    1995-01-01

    This report considers a general parallel interference cancellation scheme that significantly reduces the degradation effect of user interference but with a lesser implementation complexity than the maximum-likelihood technique. The scheme operates on the fact that parallel processing simultaneously removes from each user the interference produced by the remaining users accessing the channel in an amount proportional to their reliability. The parallel processing can be done in multiple stages. The proposed scheme uses tentative decision devices with different optimum thresholds at the multiple stages to produce the most reliably received data for generation and cancellation of user interference. The 1-stage interference cancellation is analyzed for three types of tentative decision devices, namely, hard, null zone, and soft decision, and two types of user power distribution, namely, equal and unequal powers. Simulation results are given for a multitude of different situations, in particular, those cases for which the analysis is too complex.

  20. Single-agent parallel window search

    NASA Technical Reports Server (NTRS)

    Powley, Curt; Korf, Richard E.

    1991-01-01

    Parallel window search is applied to single-agent problems by having different processes simultaneously perform iterations of Iterative-Deepening-A(asterisk) (IDA-asterisk) on the same problem but with different cost thresholds. This approach is limited by the time to perform the goal iteration. To overcome this disadvantage, the authors consider node ordering. They discuss how global node ordering by minimum h among nodes with equal f = g + h values can reduce the time complexity of serial IDA-asterisk by reducing the time to perform the iterations prior to the goal iteration. Finally, the two ideas of parallel window search and node ordering are combined to eliminate the weaknesses of each approach while retaining the strengths. The resulting approach, called simply parallel window search, can be used to find a near-optimal solution quickly, improve the solution until it is optimal, and then finally guarantee optimality, depending on the amount of time available.

  1. On the parallelization of molecular dynamics codes

    NASA Astrophysics Data System (ADS)

    Trabado, G. P.; Plata, O.; Zapata, E. L.

    2002-08-01

    Molecular dynamics (MD) codes present a high degree of spatial data locality and a significant amount of independent computations. However, most of the parallelization strategies are usually based on the manual transformation of sequential programs either by completely rewriting the code with message passing routines or using specific libraries intended for writing new MD programs. In this paper we propose a new library-based approach (DDLY) which supports parallelization of existing short-range MD sequential codes. The novelty of this approach is that it can directly handle the distribution of common data structures used in MD codes to represent data (arrays, Verlet lists, link cells), using domain decomposition. Thus, the insertion of run-time support for distribution and communication in a MD program does not imply significant changes to its structure. The method is simple, efficient and portable. It may be also used to extend existing parallel programming languages, such as HPF.

  2. New parallel SOR method by domain partitioning

    SciTech Connect

    Xie, D.; Adams, L.

    1999-07-01

    In this paper the authors propose and analyze a new parallel SOR method, the PSOR method, formulated by using domain partitioning and interprocessor data communication techniques. They prove that the PSOR method has the same asymptotic rate of convergence as the Red/Black (R/B) SOR method for the five-point stencil on both strip and block partitions, and as the four-color (R/B/G/O) SOR method for the nine-point stencil on strip partitions. They also demonstrate the parallel performance of the PSOR method on four different MIMD multiprocessors (a KSR1, an Intel Delta, a Paragon, and an IBM SP2). Finally, they compare the parallel performance of PSOR, R/B SOR, and R/B/G/O SOR. Numerical results on the Paragon indicate that PSOR is more efficient than R/B SOR and R/B/G/O SOR in both computation and interprocessor data communication.

  3. Parallel Harness for Informatic Stream Hashing

    SciTech Connect

    Steve Plimpton, Tim Shead

    2012-09-11

    PHISH is a lightweight framework which a set of independent processes can use to exchange data as they run on the same desktop machine, on processors of a parallel machine, or on different machines across a network. This enables them to work in a coordinated parallel fashion to perform computations on either streaming, archived, or self-generated data. The PHISH distribution includes a simple, portable library for performing data exchanges in useful patterns either via MPI message-passing or ZMQ sockets. PHISH input scripts are used to describe a data-processing algorithm, and additional tools provided in the PHISH distribution convert the script into a form that can be launched as a parallel job.

  4. Optimal expression evaluation for data parallel architectures

    NASA Technical Reports Server (NTRS)

    Gilbert, John R.; Schreiber, Robert

    1991-01-01

    A data parallel machine represents an array or other composits data structure by allocating one processor per data item. A pointwise operation can be performed between two such arrays in unit time, provided their corresponding elements are allocated in the same processors. If the arrays are not aligned in this fashion, the cost of moving one or both of them is part of the cost of operation. The choice of where to perform the operation then affects this cost. If an expression with several operands is to be evaluated, there may be many choices of where to perform the intermediate operations. An efficient algorithm is given to find the minimum cost way to evaluate an expression, for several different data parallel architectures. The algorithm applies to any architecture in which the metric describing the cost of moving an array has a property called robustness. This encompasses most of the common data parallel communication architectures, including meshes of arbitrary dimension and hypercubes.

  5. Optimal expression evaluation for data parallel architectures

    NASA Technical Reports Server (NTRS)

    Gilbert, J. R.; Schreiber, R.

    1990-01-01

    A data parallel machine represents an array or other composite data structure by allocating one processor per data item. A pointwise operation can be performed between two such arrays in unit time, provided their corresponding elements are allocated in the same processors. If the arrays are not aligned in this fashion, the cost of moving one or both of them is part of the cost of operation. The choice of where to perform the operation then affects this cost. If an expression with several operands is to be evaluated, there may be many choices of where to perform the intermediate operations. An efficient algorithm is given to find the minimum cost way to evaluate an expression, for several different data parallel architectures. The algorithm applies to any architecture in which the metric describing the cost of moving an array has a property called robustness. This encompasses most of the common data parallel communication architectures, including meshes of arbitrary dimension and hypercubes.

  6. Globality and speed of optical parallel processors.

    PubMed

    Lohmann, A W; Marathay, A S

    1989-09-15

    The chances of optical computing are probably best if a large number of processing elements act in parallel. The efficiency of parallel processors depends, among other things, on the time it takes to communicate signals from one processor to any other processor. In an optical parallel processor one hopes to be able to transmit a signal from one processor to any other processor within only one cycle period, no matter how far apart the processors are. Such a global communications network is desirable especially for algorithms with global interactions. The fast Fourier algorithm is an example. We define a degree of globality and we show how speed and globality are related. Our result applies to a specific architecture based on spatial filtering.

  7. PADRE: a parallel asynchronous data routing environment

    SciTech Connect

    Gunney, B; Quinlan, D

    2001-01-08

    Increasingly in industry, software design and implementation is object-oriented, developed in C++ or Java, and relies heavily on pre-existing software libraries (e.g. the Microsoft Foundation Classes for C++, the Java API for Java). A similar but more tentative trend is developing in high-performance parallel scientific computing. The transition from serial to parallel application development considerably increases the need for library support: task creation and management, data distribution and dynamic redistribution, and inter-process and inter-processor communication and synchronization must be supported. PADRE is a library to support the interoperability of parallel applications. We feel there is significant need for just such a tool to compliment the many domain-specific application frameworks presently available today, but which are generally not interoperable.

  8. Parallelization of the Lagrangian Particle Dispersion Model

    SciTech Connect

    Buckley, R.L.; O`Steen, B.L.

    1997-08-01

    An advanced stochastic Lagrangian Particle Dispersion Model (LPDM) is used by the Atmospheric Technologies Group (ATG) to simulate contaminant transport. The model uses time-dependent three-dimensional fields of wind and turbulence to determine the location of individual particles released into the atmosphere. This report describes modifications to LPDM using the Message Passing Interface (MPI) which allows for execution in a parallel configuration on the Cray Supercomputer facility at the SRS. Use of a parallel version allows for many more particles to be released in a given simulation, with little or no increase in computational time. This significantly lowers (greater than an order of magnitude) the minimum resolvable concentration levels without ad hoc averaging schemes and/or without reducing spatial resolution. The general changes made to LPDM are discussed and a series of tests are performed comparing the serial (single processor) and parallel versions of the code.

  9. Extending HPF for advanced data parallel applications

    NASA Technical Reports Server (NTRS)

    Chapman, Barbara; Mehrotra, Piyush; Zima, Hans

    1994-01-01

    The stated goal of High Performance Fortran (HPF) was to 'address the problems of writing data parallel programs where the distribution of data affects performance'. After examining the current version of the language we are led to the conclusion that HPF has not fully achieved this goal. While the basic distribution functions offered by the language - regular block, cyclic, and block cyclic distributions - can support regular numerical algorithms, advanced applications such as particle-in-cell codes or unstructured mesh solvers cannot be expressed adequately. We believe that this is a major weakness of HPF, significantly reducing its chances of becoming accepted in the numeric community. The paper discusses the data distribution and alignment issues in detail, points out some flaws in the basic language, and outlines possible future paths of development. Furthermore, we briefly deal with the issue of task parallelism and its integration with the data parallel paradigm of HPF.

  10. National Combustion Code Parallel Performance Enhancements

    NASA Technical Reports Server (NTRS)

    Quealy, Angela; Benyo, Theresa (Technical Monitor)

    2002-01-01

    The National Combustion Code (NCC) is being developed by an industry-government team for the design and analysis of combustion systems. The unstructured grid, reacting flow code uses a distributed memory, message passing model for its parallel implementation. The focus of the present effort has been to improve the performance of the NCC code to meet combustor designer requirements for model accuracy and analysis turnaround time. Improving the performance of this code contributes significantly to the overall reduction in time and cost of the combustor design cycle. This report describes recent parallel processing modifications to NCC that have improved the parallel scalability of the code, enabling a two hour turnaround for a 1.3 million element fully reacting combustion simulation on an SGI Origin 2000.

  11. Computing association probabilities using parallel Boltzmann machines.

    PubMed

    Iltis, R A; Ting, P Y

    1993-01-01

    A new computational method is presented for solving the data association problem using parallel Boltzmann machines. It is shown that the association probabilities can be computed with arbitrarily small errors if a sufficient number of parallel Boltzmann machines are available. The probability beta(i)(j) that the i th measurement emanated from the jth target can be obtained simply by observing the relative frequency with which neuron v(i,j) in a two-dimensional network is on throughout the layers. Some simple tracking examples comparing the performance of the Boltzmann algorithm to the exact data association solution and with the performance of an alternative parallel method using the Hopfield neural network are also presented.

  12. Learning in Parallel: Using Parallel Corpora to Enhance Written Language Acquisition at the Beginning Level

    ERIC Educational Resources Information Center

    Bluemel, Brody

    2014-01-01

    This article illustrates the pedagogical value of incorporating parallel corpora in foreign language education. It explores the development of a Chinese/English parallel corpus designed specifically for pedagogical application. The corpus tool was created to aid language learners in reading comprehension and writing development by making foreign…

  13. Flow of a Rarefied Gas between Parallel and Almost Parallel Plates

    DTIC Science & Technology

    2005-07-13

    Flow of a Rarefied Gas between Parallel and Almost Parallel Plates Carlo Cercignani, Maria Lampis and Silvia Lorenzani Dipartimento di Matematica ...UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Dipartimento di Matematica , Politecnico di Milano, Milano, Italy 20133 8. PERFORMING

  14. Aerodynamic simulation on massively parallel systems

    NASA Technical Reports Server (NTRS)

    Haeuser, Jochem; Simon, Horst D.

    1992-01-01

    This paper briefly addresses the computational requirements for the analysis of complete configurations of aircraft and spacecraft currently under design to be used for advanced transportation in commercial applications as well as in space flight. The discussion clearly shows that massively parallel systems are the only alternative which is both cost effective and on the other hand can provide the necessary TeraFlops, needed to satisfy the narrow design margins of modern vehicles. It is assumed that the solution of the governing physical equations, i.e., the Navier-Stokes equations which may be complemented by chemistry and turbulence models, is done on multiblock grids. This technique is situated between the fully structured approach of classical boundary fitted grids and the fully unstructured tetrahedra grids. A fully structured grid best represents the flow physics, while the unstructured grid gives best geometrical flexibility. The multiblock grid employed is structured within a block, but completely unstructured on the block level. While a completely unstructured grid is not straightforward to parallelize, the above mentioned multiblock grid is inherently parallel, in particular for multiple instruction multiple datastream (MIMD) machines. In this paper guidelines are provided for setting up or modifying an existing sequential code so that a direct parallelization on a massively parallel system is possible. Results are presented for three parallel systems, namely the Intel hypercube, the Ncube hypercube, and the FPS 500 system. Some preliminary results for an 8K CM2 machine will also be mentioned. The code run is the two dimensional grid generation module of Grid, which is a general two dimensional and three dimensional grid generation code for complex geometries. A system of nonlinear Poisson equations is solved. This code is also a good testcase for complex fluid dynamics codes, since the same datastructures are used. All systems provided good speedups, but

  15. Data communications in a parallel active messaging interface of a parallel computer

    DOEpatents

    Archer, Charles J; Blocksome, Michael A; Ratterman, Joseph D; Smith, Brian E

    2013-11-12

    Data communications in a parallel active messaging interface (`PAMI`) of a parallel computer composed of compute nodes that execute a parallel application, each compute node including application processors that execute the parallel application and at least one management processor dedicated to gathering information regarding data communications. The PAMI is composed of data communications endpoints, each endpoint composed of a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes and the endpoints coupled for data communications through the PAMI and through data communications resources. Embodiments function by gathering call site statistics describing data communications resulting from execution of data communications instructions and identifying in dependence upon the call cite statistics a data communications algorithm for use in executing a data communications instruction at a call site in the parallel application.

  16. Directions in parallel programming: HPF, shared virtual memory and object parallelism in pC++

    NASA Technical Reports Server (NTRS)

    Bodin, Francois; Priol, Thierry; Mehrotra, Piyush; Gannon, Dennis

    1994-01-01

    Fortran and C++ are the dominant programming languages used in scientific computation. Consequently, extensions to these languages are the most popular for programming massively parallel computers. We discuss two such approaches to parallel Fortran and one approach to C++. The High Performance Fortran Forum has designed HPF with the intent of supporting data parallelism on Fortran 90 applications. HPF works by asking the user to help the compiler distribute and align the data structures with the distributed memory modules in the system. Fortran-S takes a different approach in which the data distribution is managed by the operating system and the user provides annotations to indicate parallel control regions. In the case of C++, we look at pC++ which is based on a concurrent aggregate parallel model.

  17. A scalable parallel black oil simulator on distributed memory parallel computers

    NASA Astrophysics Data System (ADS)

    Wang, Kun; Liu, Hui; Chen, Zhangxin

    2015-11-01

    This paper presents our work on developing a parallel black oil simulator for distributed memory computers based on our in-house parallel platform. The parallel simulator is designed to overcome the performance issues of common simulators that are implemented for personal computers and workstations. The finite difference method is applied to discretize the black oil model. In addition, some advanced techniques are employed to strengthen the robustness and parallel scalability of the simulator, including an inexact Newton method, matrix decoupling methods, and algebraic multigrid methods. A new multi-stage preconditioner is proposed to accelerate the solution of linear systems from the Newton methods. Numerical experiments show that our simulator is scalable and efficient, and is capable of simulating extremely large-scale black oil problems with tens of millions of grid blocks using thousands of MPI processes on parallel computers.

  18. Fast Parallel Computation Of Manipulator Inverse Dynamics

    NASA Technical Reports Server (NTRS)

    Fijany, Amir; Bejczy, Antal K.

    1991-01-01

    Method for fast parallel computation of inverse dynamics problem, essential for real-time dynamic control and simulation of robot manipulators, undergoing development. Enables exploitation of high degree of parallelism and, achievement of significant computational efficiency, while minimizing various communication and synchronization overheads as well as complexity of required computer architecture. Universal real-time robotic controller and simulator (URRCS) consists of internal host processor and several SIMD processors with ring topology. Architecture modular and expandable: more SIMD processors added to match size of problem. Operate asynchronously and in MIMD fashion.

  19. Concurrency-based approaches to parallel programming

    NASA Technical Reports Server (NTRS)

    Kale, L.V.; Chrisochoides, N.; Kohl, J.; Yelick, K.

    1995-01-01

    The inevitable transition to parallel programming can be facilitated by appropriate tools, including languages and libraries. After describing the needs of applications developers, this paper presents three specific approaches aimed at development of efficient and reusable parallel software for irregular and dynamic-structured problems. A salient feature of all three approaches in their exploitation of concurrency within a processor. Benefits of individual approaches such as these can be leveraged by an interoperability environment which permits modules written using different approaches to co-exist in single applications.

  20. Parallel preconditioning techniques for sparse CG solvers

    SciTech Connect

    Basermann, A.; Reichel, B.; Schelthoff, C.

    1996-12-31

    Conjugate gradient (CG) methods to solve sparse systems of linear equations play an important role in numerical methods for solving discretized partial differential equations. The large size and the condition of many technical or physical applications in this area result in the need for efficient parallelization and preconditioning techniques of the CG method. In particular for very ill-conditioned matrices, sophisticated preconditioner are necessary to obtain both acceptable convergence and accuracy of CG. Here, we investigate variants of polynomial and incomplete Cholesky preconditioners that markedly reduce the iterations of the simply diagonally scaled CG and are shown to be well suited for massively parallel machines.

  1. Unsteady flow simulation on a parallel computer

    NASA Astrophysics Data System (ADS)

    Faden, M.; Pokorny, S.; Engel, K.

    For the simulation of the flow through compressor stages, an interactive flow simulation system is set up on an MIMD-type parallel computer. An explicit scheme is used in order to resolve the time-dependent interaction between the blades. The 2D Navier-Stokes equations are transformed into their general moving coordinates. The parallelization of the solver is based on the idea of domain decomposition. Results are presented for a problem of fixed size (4096 grid nodes for the Hakkinen case).

  2. Optimal parallel evaluation of AND trees

    NASA Technical Reports Server (NTRS)

    Wah, Benjamin W.; Li, Guo-Jie

    1990-01-01

    A quantitative analysis based on both preemptive and nonpreemptive critical-path scheduling algorithms is presently conducted for the optimal degree of parallelism required in evaluating a given AND tree. The optimal degree of parallelism is found to depend on problem complexity, precedence-graph shape, and task-time distribution along each path. In addition to demonstrating the optimality of the preemptive critical-path scheduling algorithm for evaluating an arbitrary AND tree on a fixed number of processors, the possibility of efficiently ascertaining tight bounds on the number of processors for optimal processor-time efficiency is illustrated.

  3. Parallel Digital Phase-Locked Loops

    NASA Technical Reports Server (NTRS)

    Sadr, Ramin; Shah, Biren N.; Hinedi, Sami M.

    1995-01-01

    Wide-band microwave receivers of proposed type include digital phase-locked loops in which band-pass filtering and down-conversion of input signals implemented by banks of multirate digital filters operating in parallel. Called "parallel digital phase-locked loops" to distinguish them from other digital phase-locked loops. Systems conceived as cost-effective solution to problem of filtering signals at high sampling rates needed to accommodate wide input frequency bands. Each of M filters process 1/M of spectrum of signal.

  4. Compositional reservoir simulation in parallel supercomputing environments

    SciTech Connect

    Briens, F.J.L. ); Wu, C.H. ); Gazdag, J.; Wang, H.H. )

    1991-09-01

    A large-scale compositional reservoir simulation ({gt}1,000 cells) is not often run on a conventional mainframe computer owing to excessive turnaround times. This paper presents programming and computational techniques that fully exploit the capabilities of parallel supercomputers for a large-scale compositional simulation. A novel algorithm called sequential staging of tasks (SST) that can take full advantage of parallel-vector processing to speed up the solution of a large linear system is introduced. The effectiveness of SST is illustrated with results from computer experiments conducted on an IBM 3090-600E.

  5. Operation of high power converters in parallel

    NASA Technical Reports Server (NTRS)

    Decker, D. K.; Inouye, L. Y.

    1993-01-01

    High power converters that are used in space power subsystems are limited in power handling capability due to component and thermal limitations. For applications, such as Space Station Freedom, where multi-kilowatts of power must be delivered to user loads, parallel operation of converters becomes an attractive option when considering overall power subsystem topologies. TRW developed three different unequal power sharing approaches for parallel operation of converters. These approaches, known as droop, master-slave, and proportional adjustment, are discussed and test results are presented.

  6. Concurrency-based approaches to parallel programming

    SciTech Connect

    Kale, L.V.; Chrisochoides, N.; Kohl, J.

    1995-07-17

    The inevitable transition to parallel programming can be facilitated by appropriate tools, including languages and libraries. After describing the needs of applications developers, this paper presents three specific approaches aimed at development of efficient and reusable parallel software for irregular and dynamic-structured problems. A salient feature of all three approaches in their exploitation of concurrency within a processor. Benefits of individual approaches such as these can be leveraged by an interoperability environment which permits modules written using different approaches to co-exist in single applications.

  7. Xyce parallel electronic simulator release notes.

    SciTech Connect

    Keiter, Eric R; Hoekstra, Robert John; Mei, Ting; Russo, Thomas V.; Schiek, Richard Louis; Thornquist, Heidi K.; Rankin, Eric Lamont; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    The Xyce Parallel Electronic Simulator has been written to support, in a rigorous manner, the simulation needs of the Sandia National Laboratories electrical designers. Specific requirements include, among others, the ability to solve extremely large circuit problems by supporting large-scale parallel computing platforms, improved numerical performance and object-oriented code design and implementation. The Xyce release notes describe: Hardware and software requirements New features and enhancements Any defects fixed since the last release Current known defects and defect workarounds For up-to-date information not available at the time these notes were produced, please visit the Xyce web page at http://www.cs.sandia.gov/xyce.

  8. Gang scheduling a parallel machine. Revision 1

    SciTech Connect

    Gorda, B.C.; Brooks, E.D. III

    1991-12-01

    Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processes. User programs and their gangs of processes are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantum are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory.

  9. Structured building model reduction toward parallel simulation

    SciTech Connect

    Dobbs, Justin R.; Hencey, Brondon M.

    2013-08-26

    Building energy model reduction exchanges accuracy for improved simulation speed by reducing the number of dynamical equations. Parallel computing aims to improve simulation times without loss of accuracy but is poorly utilized by contemporary simulators and is inherently limited by inter-processor communication. This paper bridges these disparate techniques to implement efficient parallel building thermal simulation. We begin with a survey of three structured reduction approaches that compares their performance to a leading unstructured method. We then use structured model reduction to find thermal clusters in the building energy model and allocate processing resources. Experimental results demonstrate faster simulation and low error without any interprocessor communication.

  10. Object-oriented parallel polygon rendering

    SciTech Connect

    Heiland, R.W.

    1994-09-01

    Since many scientific datasets can be visualized using some polygonal representation, a polygon renderer has broad use for scientific visualization. With today`s high performance computing applications producing very large datasets, a parallel polygon renderer is a necessary tool for keeping the compute-visualize cycle at a minimum. This paper presents a DOIV on renderer that combines the shared-memory and message-passing models of parallel programming. It uses the Global Arrays library, a shared-memory programming toolkit for distributed memory machines. The experience of using an object oriented approach for software design and development is also discussed.

  11. Interfacing Computer Aided Parallelization and Performance Analysis

    NASA Technical Reports Server (NTRS)

    Jost, Gabriele; Jin, Haoqiang; Labarta, Jesus; Gimenez, Judit; Biegel, Bryan A. (Technical Monitor)

    2003-01-01

    When porting sequential applications to parallel computer architectures, the program developer will typically go through several cycles of source code optimization and performance analysis. We have started a project to develop an environment where the user can jointly navigate through program structure and performance data information in order to make efficient optimization decisions. In a prototype implementation we have interfaced the CAPO computer aided parallelization tool with the Paraver performance analysis tool. We describe both tools and their interface and give an example for how the interface helps within the program development cycle of a benchmark code.

  12. PADMA: PArallel Data Mining Agents for scalable text classification

    SciTech Connect

    Kargupta, H.; Hamzaoglu, I.; Stafford, B.

    1997-03-01

    This paper introduces PADMA (PArallel Data Mining Agents), a parallel agent based system for scalable text classification. PADMA contains modules for (1) parallel data accessing operations, (2) parallel hierarchical clustering, and (3) web-based data visualization. This paper introduces the general architecture of PADMA and presents a detailed description of its different modules.

  13. Parallel solution of sparse one-dimensional dynamic programming problems

    NASA Technical Reports Server (NTRS)

    Nicol, David M.

    1989-01-01

    Parallel computation offers the potential for quickly solving large computational problems. However, it is often a non-trivial task to effectively use parallel computers. Solution methods must sometimes be reformulated to exploit parallelism; the reformulations are often more complex than their slower serial counterparts. We illustrate these points by studying the parallelization of sparse one-dimensional dynamic programming problems, those which do not obviously admit substantial parallelization. We propose a new method for parallelizing such problems, develop analytic models which help us to identify problems which parallelize well, and compare the performance of our algorithm with existing algorithms on a multiprocessor.

  14. Collectively loading an application in a parallel computer

    SciTech Connect

    Aho, Michael E.; Attinella, John E.; Gooding, Thomas M.; Miller, Samuel J.; Mundy, Michael B.

    2016-01-05

    Collectively loading an application in a parallel computer, the parallel computer comprising a plurality of compute nodes, including: identifying, by a parallel computer control system, a subset of compute nodes in the parallel computer to execute a job; selecting, by the parallel computer control system, one of the subset of compute nodes in the parallel computer as a job leader compute node; retrieving, by the job leader compute node from computer memory, an application for executing the job; and broadcasting, by the job leader to the subset of compute nodes in the parallel computer, the application for executing the job.

  15. Parallelized reliability estimation of reconfigurable computer networks

    NASA Technical Reports Server (NTRS)

    Nicol, David M.; Das, Subhendu; Palumbo, Dan

    1990-01-01

    A parallelized system, ASSURE, for computing the reliability of embedded avionics flight control systems which are able to reconfigure themselves in the event of failure is described. ASSURE accepts a grammar that describes a reliability semi-Markov state-space. From this it creates a parallel program that simultaneously generates and analyzes the state-space, placing upper and lower bounds on the probability of system failure. ASSURE is implemented on a 32-node Intel iPSC/860, and has achieved high processor efficiencies on real problems. Through a combination of improved algorithms, exploitation of parallelism, and use of an advanced microprocessor architecture, ASSURE has reduced the execution time on substantial problems by a factor of one thousand over previous workstation implementations. Furthermore, ASSURE's parallel execution rate on the iPSC/860 is an order of magnitude faster than its serial execution rate on a Cray-2 supercomputer. While dynamic load balancing is necessary for ASSURE's good performance, it is needed only infrequently; the particular method of load balancing used does not substantially affect performance.

  16. Functional MRI using regularized parallel imaging acquisition.

    PubMed

    Lin, Fa-Hsuan; Huang, Teng-Yi; Chen, Nan-Kuei; Wang, Fu-Nien; Stufflebeam, Steven M; Belliveau, John W; Wald, Lawrence L; Kwong, Kenneth K

    2005-08-01

    Parallel MRI techniques reconstruct full-FOV images from undersampled k-space data by using the uncorrelated information from RF array coil elements. One disadvantage of parallel MRI is that the image signal-to-noise ratio (SNR) is degraded because of the reduced data samples and the spatially correlated nature of multiple RF receivers. Regularization has been proposed to mitigate the SNR loss originating due to the latter reason. Since it is necessary to utilize static prior to regularization, the dynamic contrast-to-noise ratio (CNR) in parallel MRI will be affected. In this paper we investigate the CNR of regularized sensitivity encoding (SENSE) acquisitions. We propose to implement regularized parallel MRI acquisitions in functional MRI (fMRI) experiments by incorporating the prior from combined segmented echo-planar imaging (EPI) acquisition into SENSE reconstructions. We investigated the impact of regularization on the CNR by performing parametric simulations at various BOLD contrasts, acceleration rates, and sizes of the active brain areas. As quantified by receiver operating characteristic (ROC) analysis, the simulations suggest that the detection power of SENSE fMRI can be improved by regularized reconstructions, compared to unregularized reconstructions. Human motor and visual fMRI data acquired at different field strengths and array coils also demonstrate that regularized SENSE improves the detection of functionally active brain regions.

  17. Functional MRI Using Regularized Parallel Imaging Acquisition

    PubMed Central

    Lin, Fa-Hsuan; Huang, Teng-Yi; Chen, Nan-Kuei; Wang, Fu-Nien; Stufflebeam, Steven M.; Belliveau, John W.; Wald, Lawrence L.; Kwong, Kenneth K.

    2013-01-01

    Parallel MRI techniques reconstruct full-FOV images from undersampled k-space data by using the uncorrelated information from RF array coil elements. One disadvantage of parallel MRI is that the image signal-to-noise ratio (SNR) is degraded because of the reduced data samples and the spatially correlated nature of multiple RF receivers. Regularization has been proposed to mitigate the SNR loss originating due to the latter reason. Since it is necessary to utilize static prior to regularization, the dynamic contrast-to-noise ratio (CNR) in parallel MRI will be affected. In this paper we investigate the CNR of regularized sensitivity encoding (SENSE) acquisitions. We propose to implement regularized parallel MRI acquisitions in functional MRI (fMRI) experiments by incorporating the prior from combined segmented echo-planar imaging (EPI) acquisition into SENSE reconstructions. We investigated the impact of regularization on the CNR by performing parametric simulations at various BOLD contrasts, acceleration rates, and sizes of the active brain areas. As quantified by receiver operating characteristic (ROC) analysis, the simulations suggest that the detection power of SENSE fMRI can be improved by regularized reconstructions, compared to unregularized reconstructions. Human motor and visual fMRI data acquired at different field strengths and array coils also demonstrate that regularized SENSE improves the detection of functionally active brain regions. PMID:16032694

  18. Adapting implicit methods to parallel processors

    SciTech Connect

    Reeves, L.; McMillin, B.; Okunbor, D.; Riggins, D.

    1994-12-31

    When numerically solving many types of partial differential equations, it is advantageous to use implicit methods because of their better stability and more flexible parameter choice, (e.g. larger time steps). However, since implicit methods usually require simultaneous knowledge of the entire computational domain, these methods axe difficult to implement directly on distributed memory parallel processors. This leads to infrequent use of implicit methods on parallel/distributed systems. The usual implementation of implicit methods is inefficient due to the nature of parallel systems where it is common to take the computational domain and distribute the grid points over the processors so as to maintain a relatively even workload per processor. This creates a problem at the locations in the domain where adjacent points are not on the same processor. In order for the values at these points to be calculated, messages have to be exchanged between the corresponding processors. Without special adaptation, this will result in idle processors during part of the computation, and as the number of idle processors increases, the lower the effective speed improvement by using a parallel processor.

  19. Parallel processing: The Cm/sup */ experience

    SciTech Connect

    Siewiorek, D.; Gehringer, E.; Segall, Z.

    1986-01-01

    This book describes the parallel-processing research with CM/sup */ at Carnegie-Mellon University. Cm/sup */ is a tightly coupled 50-processor multiprocessing system that has been in operation since 1977. Two complete operating systems-StarOS and Medusa-are part of its development along with a number of applications.

  20. Ray tracing on distributed memory parallel systems

    NASA Technical Reports Server (NTRS)

    Jensen, David W.; Reed, Daniel A.

    1990-01-01

    Among the many techniques in computer graphics, ray tracing is prized because it can render realistic images, albeit at great computational expense. In this note, the performance of several approaches to ray tracing on a distributed memory parallel system is evaluated. A set of performance instrumentation tools and their associated visualization software are used to identify the underlying causes of performance differences.