STOMP Sparse Vegetation Evapotranspiration Model for the Water-Air-Energy Operational Mode
Ward, Anderson L.; White, Mark D.; Freeman, Eugene J.; Zhang, Z. F.
2005-09-15
The Water-Air-Energy (WAE) Operational Mode of the Subsurface Transport Over Multiple Phases (STOMP) numerical simulator solves the coupled conservation equations for water mass, air mass, and thermal energy in multiple dimensions. This addendum describes the theory, input file formatting, and application of a soil-vegetation-atmosphere transfer (SVAT) scheme for STOMP that is based on a sparse vegetation evapotranspiration model. The SVAT scheme is implemented as a boundary condition on the upper surface of the computational domain and has capabilities for simulating evaporation from bare surfaces as well as evapotranspiration from sparsely vegetated surfaces populated with single or multiple plant species in response to meteorological forcings. With this extension, the model calculates water mass, air mass and thermal energy across a boundary surface in addition to root-water transport between the subsurface and atmosphere. This mode represents the barrier extension of the WAE mode and is designated as STOMP-WAE-B. Input for STOMP-WAE-B is specified via three input cards and include: atmospheric conditions through the Atmospheric Conditions Card; time-invariant plant species data through the Plant Properties Card; and time varying plant species data through the Boundary Conditions Card. Two optional cards, the Observed Data and UCODE Control Cards allow use of STOMP-WAE with UCODE in an inverse mode to estimate model parameters. STOMP-WAE was validated by solving a number of test problems from the literature that included experimental observations as well as analytical or numerical solutions. Several of the UNSAT-H verification problems are included along with a benchmark simulation derived from a recently published intercode comparison for barrier design tools. Results show that STOMP is able to meet, and in most cases, exceed performance of other commonly used simulation codes without having to resort to may of their simplifying assumptions. Use of the fully
ERIC Educational Resources Information Center
Martinez, Luz Adriana
2007-01-01
The Specialized Training of Military Parents, better known by many as STOMP, is a federally funded Parent Training and Information (PTI) center established to assist military families who have children with specialized education or health needs. STOMP exists to empower military parents, individuals with disabilities, and service providers with…
Energy budgets and resistances to energy transport in sparsely vegetated rangeland
Nichols, W.D.
1992-01-01
Partitioning available energy between plants and bare soil in sparsely vegetated rangelands will allow hydrologists and others to gain a greater understanding of water use by native vegetation, especially phreatophytes. Standard methods of conducting energy budget studies result in measurements of latent and sensible heat fluxes above the plant canopy which therefore include the energy fluxes from both the canopy and the soil. One-dimensional theoretical numerical models have been proposed recently for the partitioning of energy in sparse crops. Bowen ratio and other micrometeorological data collected over phreatophytes growing in areas of shallow ground water in central Nevada were used to evaluate the feasibility of using these models, which are based on surface and within-canopy aerodynamic resistances, to determine heat and water vapor transport in sparsely vegetated rangelands. The models appear to provide reasonably good estimates of sensible heat flux from the soil and latent heat flux from the canopy. Estimates of latent heat flux from the soil were less satisfactory. Sensible heat flux from the canopy was not well predicted by the present resistance formulations. Also, estimates of total above-canopy fluxes were not satisfactory when using a single value for above-canopy bulk aerodynamic resistance. ?? 1992.
STOMP Subsurface Transport Over Multiple Phases: STOMP-CO2 and STOMP-CO2e Guide: Version 1.0
White, Mark D.; Bacon, Diana H.; McGrail, B. Peter; Watson, David J.; White, Signe K.; Zhang, Z. F.
2012-04-03
This STOMP (Subsurface Transport Over Multiple Phases) guide document describes the theory, use, and application of the STOMP-CO2 and STOMP-CO2e operational modes. These operational modes of the STOMP simulator are configured to solve problems involving the sequestration of CO2 in geologic saline reservoirs. STOMP-CO2 is the isothermal version and STOMP-CO2e is the nonisothermal version. These core operational modes solve the governing conservation equations for component flow and transport through geologic media; where, the STOMP-CO2 components are water, CO2 and salt and the STOMP-CO2e operational mode also includes an energy conservation equation. Geochemistry can be included in the problem solution via the ECKEChem (Equilibrium-Conservation-Kinetic-Equation Chemistry) module, and geomechanics via the EPRMech (Elastic-Plastic-Rock Mechanics) module. This addendum is designed to provide the new user with a full guide for the core capabilities of the STOMP-CO2 and -CO2e simulators, and to provide the experienced user with a quick reference on implementing features. Several benchmark problems are provided in this addendum, which serve as starting points for developing inputs for more complex problems and as demonstrations of the simulator’s capabilities.
Biomechanical assessment of the injury risk of stomping.
Schirmer, Frederike; Muggenthaler, Holger; Hubig, Michael; Schenkl, Sebastian; Koch, Markus; Blickhan, Reinhard; Mall, Gita
2016-05-01
Forensic case work as well as literature shows that severe head injuries, e.g., with basilar fractures and cerebral hemorrhages due to stomps can be seen; however, there is no data basis concerning contact forces and potential influencing factors. The objective of this work was to generate a data basis of contact forces in stomping by performing experimental measurements and subsequent statistical analyses. Fifty-five volunteers participated in the present study. Each participant performed several stomps onto force plates with sturdy/soft footwear as well as with/without an elastic layer imitating the scalp. Ground reaction forces induced by jumps were also measured for sturdy and soft footwear. The results show statistically significant dependencies between maximum ground reaction forces and body weight and body height. A statistically significant influence of footwear on stomping force could only be found in tests with an elastic layer and in the jumping setup. Mean maximum stomping forces for the female volunteers were between 4694 and 5970 N; male volunteers were able to produce mean peak stomping forces between 8494 and 9016 N. Jumping forces were approximately twice the stomping forces for both male and female test persons. Regardless of footwear and gender, it can be claimed that a forceful stomp or jump to someone's head supported on the ground can cause facial and skull fractures. Thus, forceful stomps or jumps to someone's head can cause potential fatal injuries independent of footwear, gender, or fitness level. PMID:27075915
NASA Technical Reports Server (NTRS)
Otterman, J.; Susskind, J.; Brakke, T.; Kimes, D.; Pielke, R.; Lee, T. J.
1995-01-01
The thermal-infrared (longwave) emission from a vegetated terrain is generally anisotropic, i.e., the emission temperature varies with the view direction. If a directional measurement of temperature is considered to be equal to the effective temperature of the hemispheric emission, then the estimate of the latter can be significantly in error. The view-direction (zenith angle theta(sub eq) at which the emission equivalence does hold is determined in our modeling study. In a two-temperature field-of-view (soil and plants), theta(sub eq) falls in a narrow range depending on plant density and canopy architecture. Theta(sub eq) does not depend on soil and (uniform) plant temperatures nor on their ratio, even though the pattern of emission vs. the view direction depends crucially on this ratio. For a sparse canopy represented as thin, vertical cylindrical stalks (or vertical blades uniformly distributed in azimuth) with horizontal facets, theta(sub eq) ranges from 48 to 53 deg depending on the optical density of the vertical elements alone. When plant elements are modeled as small spheres, theta(sub eq) lies between 53 to 57 deg (for the same values of the canopy optical density). Only for horizontal leaves (a truly planophile canopy) is the temperature measured from any direction equal to the temperature of the hemispheric emission. When the emission temperature changes with optical depth within the canopy at a specified rate, theta(sub eq) depends to some extent on that rate. For practically any sparsely vegetated surface, a directional measurement at the zenith angle of 50 deg offers an appropriate evaluation of the hemispheric emission, since the error in the estimate will, at most, only slightly exceed 1% (around 4 W/sq m). Estimates of the hemispheric emission through a nadir measurement, on the other hand, can be in error in some cases by about 10%, i.e., on the order of 40 W/sq m.
Turbulent heat transfer from a sparsely vegetated surface - Two-component representation
NASA Technical Reports Server (NTRS)
Otterman, J.; Novak, M. D.; Starr, D. O'C.
1993-01-01
The conventional calculation of heat fluxes from a vegetated surface involving the coefficient of turbulent heat transfer which increases logarithmically with surface roughness, is inappropriate such highly structured surfaces as desert scrub or open forest. An approach is developed here for computing sensible heat flux from sparsely vegetated surfaces, where the absorption of insolation and the transfer of absorbed heat to the atmosphere are calculated separately for the plants and for the soil. This approach is applied to a desert-scrub surface in the northern Sinai, for which the turbulent transfer coefficient of sensible heat flux from the plants is much larger than that from the soil below, as shown by an analysis of plant, soil, and air temperatures. The plant density is expressed as the sum of products (plant-height) x (plant-diameter) of plants per unit horizontal surface area. The solar heat absorbed by the plants is assumed to be transferred immediately to the airflow. The effective turbulent transfer coefficient k(g-eff) for sensible heat from the desert-scrub/soil surface computed under this assumption increases sharply with increasing solar zenith angle, as the plants absorb a greater fraction of the incoming irradiation. The surface absorptivity (the coalbedo) also increases sharply with increasing solar zenith angle, and thus the sensible heat flux from such complex surfaces is a much broader function of time of day than when computed under constant k(g-eff) and constant albedo assumptions.
STOMP, Subsurface Transport Over Multiple Phases, theory guide
White, M.D.; Oostrom, M.
1996-10-01
This guide describes the simulator`s governing equations, constitutive functions and numerical solution algorithms of the STOMP (Subsurface Transport Over Multiple Phases) simulator, a scientific tool for analyzing multiple phase subsurface flow and transport. The STOMP simulator`s fundamental purpose is to produce numerical predictions of thermal and hydrologic flow and transport phenomena in variably saturated subsurface environments, which are contaminated with volatile or nonvolatile organic compounds. Auxiliary applications include numerical predictions of solute transport processes including radioactive chain decay processes. In writing these guides for the STOMP simulator, the authors have assumed that the reader comprehends concepts and theories associated with multiple-phase hydrology, heat transfer, thermodynamics, radioactive chain decay, and nonhysteretic relative permeability, saturation-capillary pressure constitutive functions. The authors further assume that the reader is familiar with the computing environment on which they plan to compile and execute the STOMP simulator. The STOMP simulator requires an ANSI FORTRAN 77 compiler to generate an executable code. The memory requirements for executing the simulator are dependent on the complexity of physical system to be modeled and the size and dimensionality of the computational domain. Likewise execution speed depends on the problem complexity, size and dimensionality of the computational domain, and computer performance. One-dimensional problems of moderate complexity can be solved on conventional desktop computers, but multidimensional problems involving complex flow and transport phenomena typically require the power and memory capabilities of workstation or mainframe type computer systems.
NASA Astrophysics Data System (ADS)
Maltese, Antonino; Capodici, Fulvio; Ciraolo, Giuseppe; La Loggia, Goffredo
2013-01-01
A critical analysis of a thermal inertia approach to map surface soil water content on bare and sparsely vegetated soils by means of remotely sensed data is reported. The study area is an experimental field located in Barrax, Spain. In situ data were acquired within the Barrax 2011 research project. An advanced hyperspectral scanner airborne imager provides images in the visible/near-infrared and thermal infrared bands. Images were acquired both in day and night times by the Instituto Nacional de Técnica Aeroespacial between 12th and 13th of June 2011. The scene covers a corn irrigation pivot surrounded by bare soil, where a set of in situ data have been collected both previously and simultaneously to overpasses. To validate remotely sensed estimations, an ad hoc dataset has been produced by measuring spectra, radiometric temperatures, surface soil water content, and soil thermal properties. These data were collected on two transects covering bare and sparsely vegetated soils. This ground dataset was used (1) to verify if a thermal inertia method can be applied to map the water content on soil covered by sparse vegetation and (2) to quantify a correction factor accounting for solar radiation reduction due to sky cloudiness. The experiment intended to test a spatially constant and a spatially distributed approach to estimate the phase difference. Both methods were then applied to the airborne images collected during the following days to obtain the spatial distribution of surface soil water content. Results confirm that the thermal inertia method can be applied to sparsely vegetated soil characterized by low fractional cover if the solar radiation reaching the ground is accurately estimated. A spatially constant value of the phase difference allows a good assessment of thermal inertia, whereas the comparison with the three-temperature approach did not give conclusive responses. Results also show that clear sky, only at the time of the acquisition, does not provide
NASA Astrophysics Data System (ADS)
Maltese, Antonino; Capodici, Fulvio; Corbari, Chiara; Ciraolo, Giuseppe; La Loggia, Goffredo; Sobrino, José Antonio
2012-09-01
The paper reports a critical analysis of the thermal inertia approach to map surface soil water content on bare and sparsely vegetated soils by means of remotely sensed data. The study area is an experimental area located in Barrax (Spain). Field data were acquired within the Barrax 2011 research project. AHS airborne images including VIS/NIR and TIR bands were acquired both day and night time by the INTA (Instituto Nacional de Técnica Aeroespacial) between the 11th and 13rd of June 2011. Images cover a corn pivot surrounded by bare soil, where a set of in situ data have been collected previously and simultaneously to overpasses. To validate remotely sensed estimations, a preliminary proximity sensing set up has been arranged, measuring spectra and surface temperatures on transects by means of ASD hand-held spectroradiometer and an Everest Interscience radiometric thermometer respectively. These data were collected on two transects: the first one on bare soil and the second from bare to sparsely vegetated soil; soil water content in both transects ranged approximately between field and saturation values. Furthermore thermal inertia was measured using a KD2Pro probe, and surface water content of soil was measured using FDR and TDR probes. This ground dataset was used: 1) to verify if the thermal inertia method can be applied to map water content also on soil covered by sparse vegetation, and 2) to quantify a correction factor of the downwelling shortwave radiation taking into account sky cloudiness effects on thermal inertia assessment. The experiment tests both Xue and Cracknell approximation to retrieve the thermal inertia from a dumped value of the phase difference and the three-temperature approach of Sobrino to estimate the phase difference spatial distribution. Both methods were then applied on the remotely sensed airborne images collected during the following days, in order to obtain the spatial distribution of the surface soil moisture on bare soils and
NASA Astrophysics Data System (ADS)
Kumar, U.; Milesi, C.; Nemani, R. R.; Raja, S. Kumar; Ganguly, S.; Wang, W.
2015-06-01
In this paper, we explore the possibility of sparse regression, a new direction in unmixing, for vegetation and urban area classification. SUnSAL (Sparse unmixing via variable splitting and augmented Lagrangian) in both unconstrained and constrained forms (with the abundance non-negativity and abundance sum-to-one constraints) were used with a set of global endmembers (substrate, vegetation and dark objects) to unmix a set of computer simulated noise-free and noisy data (with Gaussian noise of different signal-to-noise ratio) in order to judge the robustness of the algorithm. The error in the fractional estimate was examined for varying noise power (variance): 2, 4, 8, 16, 32, 64, 128 and 256. In the second set of experiments, a spectrally diverse collection of 11 scenes of Level 1 terrain corrected, cloud free Landsat-5 TM data representing an agricultural setup in Fresno, California, USA were used. The corresponding ground data for validation were collected on the same days of satellite overpass. Finally in the third set of experiments, a clear sky Landsat-5 TM data for an area near the Golden Gate Bridge, San Francisco (an urbanized landscape), California, USA were used to assess the algorithm. The fractional estimates of the 30 m Landsat-5 TM data were compared with the fractional estimates of a high-resolution World View-2 data (2 m spatial resolution) obtained using a fully constrained least squares algorithm. The results were evaluated using descriptive statistics, correlation coefficient, RMSE, probability of success and bivariate distribution function, which showed that constrained model was better than unconstrained form.
STOMP Subsurface Transport Over Multiple Phases: User`s guide
White, M.D.; Oostrom, M.
1997-10-01
The U.S. Department of Energy, through the Office of Technology Development, has requested the demonstration of remediation technologies for the cleanup of volatile organic compounds and associated radionuclides within the soil and groundwater at arid sites. This demonstration program, called the VOC-Arid Soils Integrated Demonstration Program (Arid-ID), has been initially directed at a volume of unsaturated and saturated soil contaminated with carbon tetrachloride, on the Hanford Site near Richland, Washington. A principal subtask of the Arid-ID program involves the development of an integrated engineering simulator for evaluating the effectiveness and efficiency of various remediation technologies. The engineering simulator`s intended users include scientists and engineers who are investigating soil physics phenomena associated with remediation technologies. Principal design goals for the engineer simulator include broad applicability, verified algorithms, quality assurance controls, and validated simulations against laboratory and field-scale experiments. An important goal for the simulator development subtask involves the ability to scale laboratory and field-scale experiments to full-scale remediation technologies, and to transfer acquired technology to other arid sites. The STOMP (Subsurface Transport Over Multiple Phases) simulator has been developed by the Pacific Northwest National Laboratory for modeling remediation technologies. Information on the use, application, and theoretical basis of the STOMP simulator theory and discussions on the governing equations, constitutive relations, and numerical solution algorithms for the STOMP simulator.
STOMP Subsurface Transport Over Multiple Phases: Application guide
Nichols, W.E.; Aimo, N.J.; Oostrom, M.; White, M.D.
1997-09-01
The U.S. Department of Energy (DOE), through the Office of Technology Development, has requested the demonstration of remediation technologies for the cleanup of volatile organic compounds and associated radionuclides within the soil and ground water at arid sites. This demonstration program, called the VOC-Arid Soils Integrated Demonstrated Program (Arid-ID), has been initially directed at a volume of unsaturated and saturated soil contaminated with carbon tetrachloride on the Hanford Site near Richland, Washington. A principal subtask of the Arid-ID program involves the development of an integrated engineering simulator for evaluating the effectiveness and efficiency of various remediation technologies. The engineering simulator`s intended users include scientists and engineers who are investigating soil physics phenomena associated with remediation technologies. Principal design goals for the engineering simulator include broad applicability, verified algorithms, quality assurance controls, and validated simulations against laboratory and field-scale experiments. An important goal for the simulator development subtask involves the ability to scale laboratory and field-scale experiments to full-scale remediation technologies, and to transfer acquired technology to other arid sites. The STOMP (Subsurface Transport Over Multiple Phases) simulator has been developed by the Pacific Northwest Laboratory for modeling remediation technologies. Information on the use, application, and theoretical basis of the STOMP simulator are documented in three companion guide guides. This document, the Application Guide, provides a suite of example applications of the STOMP simulator.
NASA Astrophysics Data System (ADS)
Trautz, A.; Illangasekare, T. H.; Tilton, N.
2015-12-01
Soil moisture is a fundamental state variable that provides the water necessary for plant growth and evapotranspiration. Soil moisture has been extensively studied in the context of bare surface soils and root zones. Less attention has focused on the effects of sparse vegetation distributions, such as those typical of agricultural cropland and other natural surface environments, on soil moisture dynamics. The current study explores root zone, bulk soil, and near-surface atmosphere interactions in terms of soil moisture under different distributions of sparse vegetation using multi-scale laboratory experimentation and numerical simulation. This research is driven by the need to advance our fundamental understanding of soil moisture dynamics in the context of improving water conservation and next generation heat and mass transfer numerical models. Experimentation is performed in a two-dimensional 7.3 m long intermediate scale soil tank interfaced with a climate-controlled wind tunnel, both of which are outfitted with current sensor technologies for measuring atmospheric and soil variables. The soil tank is packed so that a sparsely vegetated soil is surrounded by bulk bare soil; the two regions are separated by porous membranes to isolate the root zone from the bulk soil. Results show that in the absence of vegetation, evaporation rates vary along the soil tank in response to longitudinal changes in humidity; soil dries fastest upstream where evaporation rates are highest. In the presence of vegetation, soil moisture in the bulk soil closest to a vegetated region decreases more rapidly than the bulk soil farther away. Evapotranspiration rates in this region are also higher than the bulk soil region. This study is the first step towards the development of more generalized models that account for non-uniformly distributed vegetation and land surfaces exhibiting micro-topology.
STOMP Subsurface Transport Over Multiple Phases, Version 4.0, User’s Guide
White, Mark D.; Oostrom, Martinus
2006-06-09
This guide describes the general use, input file formatting, compilation and execution of the STOMP (Subsurface Transport Over Multiple Phases) simulator, a scientific tool for analyzing single and multiple phase subsurface flow and transport. A description of the simulator’s governing equations, constitutive functions and numerical solution algorithms are provided in a companion theory guide. In writing these guides for the STOMP simulator, the authors have assumed that the reader comprehends concepts and theories associated with multiple-phase hydrology, heat transfer, thermodynamics, radioactive chain decay, and relative permeability-saturation-capillary pressure constitutive relations. The authors further assume that the reader is familiar with the computing environment on which they plan to compile and execute the STOMP simulator. Source codes for the sequential versions of the simulator are available in pure FORTRAN 77 or mixed FORTRAN 77/90 forms. The pure FORTRAN 77 source code form requires a parameters file to define the memory requirements for the array elements. The mixed FORTRAN 77/90 form of the source code uses dynamic memory allocation to define memory requirements, based on a FORTRAN 90 preprocessor STEP, that reads the input files. The simulator utilizes a variable source code configuration, which allows the execution memory and speed to be tailored to the problem specifics, and essentially requires that the source code be assembled and compiled through a software maintenance utility. The memory requirements for executing the simulator are dependent on the complexity of physical system to be modeled and the size and dimensionality of the computational domain. Likewise execution speed depends on the problem complexity, size and dimensionality of the computational domain, and computer performance. Selected operational modes of the STOMP simulator are available for scalable execution on multiple processor (i.e., parallel) computers. These versions
STOMP: A Software Architecture for the Design and Simulation UAV-Based Sensor Networks
Jones, E D; Roberts, R S; Hsia, T C S
2002-10-28
This paper presents the Simulation, Tactical Operations and Mission Planning (STOMP) software architecture and framework for simulating, controlling and communicating with unmanned air vehicles (UAVs) servicing large distributed sensor networks. STOMP provides hardware-in-the-loop capability enabling real UAVs and sensors to feedback state information, route data and receive command and control requests while interacting with other real or virtual objects thereby enhancing support for simulation of dynamic and complex events.
Dimitrov, Boyan D; Gadeva, Polina G; Benova, Donka K; Bineva, Maria V
2006-11-01
The genotoxicities of the herbicides Roundup (glyphosate), Stomp (pendimethaline) and Reglone (diquat), were compared in plant (Crepis capillaris L.) and mouse bone marrow test systems using chromosomal aberrations and micronuclei. Roundup did not induce chromosomal aberrations or micronuclei in either test system. Reglone also did not induce chromosomal aberrations in either test system; however, it increased micronucleus frequency in both plant cells and mouse bone marrow polychromatic erythrocytes (PCEs). The responses of the two test systems to Stomp were quite different. Stomp did not induce chromosomal aberrations in the plant cells, but increased their incidence in mouse cells; Stomp increased the frequency of micronuclei in both test systems. The induction of micronuclei in plant cells may have been due to the spindle-destroying effect of the herbicide, since all concentrations of Stomp produced C-mitoses. The increased chromosomal aberration frequency in mouse bone marrow cells observed at later sampling times after administration of Stomp into animals suggests that the induction of aberrations may be due to biosynthesis of genotoxic metabolites. This conclusion was supported by the coincidence between the frequencies of chromosomal aberrations and of micronucleated PCEs in mouse cells. These data indicate that plant and animal assays are differentially responsive to some pesticides, and these differences may be due to metabolism and their responses to mitotic spindle disruption. PMID:16998229
Lu, Zhiming; Meyer, D.J.
2002-01-01
We used interferometric methods on a pair of repeat-pass ERS-1 synthetic aperture radar (SAR) images to study soil moisture changes over sparsely vegetated targets. The intensity of the SAR image acquired at one time was higher than that of an image acquired at an earlier time. We used a correlation image computed from the SAR image pair to study the cause of the observed changes in SAR intensity. Because a reduction of correlation over areas with intensity changes was not observed, we interpreted the intensity changes as not being caused by changes in roughness/structure, but by a change in soil moisture owing to rainfall. An increase in soil moisture ranging from 5% to 20% is the most likely explanation for the increase of intensity. These analyses imply that both intensity and phase information should be used in SAR change detection applications.
STOMP-ECKEChem: An Engineering Perspective on Reactive Transport in Geologic Media
White, Mark D.; Fang, Yilin
2012-04-04
ECKEChem (Equilibrium, Conservation, Kinetic Equation Chemistry) is a reactive transport module for the STOMP suite of multifluid subsurface flow and transport simulators that was developed from an engineering perspective. STOMP comprises a suite of operational modes that are distinguished by the solved coupled conservation equations with capabilities for a variety of subsurface applications (e.g., environmental remediation and stewardship, geologic sequestration of greenhouse gases, gas hydrate production, and oil shale production). The ECKEChem module was designed to provide integrated reactive transport capabilities across the suite of STOMP simulator operational modes. The initial application for the ECKEChem module was in the simulation of the mineralization reactions that occurred with the injection of supercritical carbon dioxide into deep Columbia River basalt formations, where it was implemented in the STOMP-CO2 simulator. The STOMP-ECKEChem solution approach to modeling reactive transport in multifluid geologic media is founded on an engineering perspective: (1) sequential non-iterative coupling between the flow and reactive transport is sufficient, (2) reactive transport can be modeled by operator splitting with local geochemistry and global transport, (3) geochemistry can be expressed as a system of coupled nonlinear equilibrium, conservation and kinetic equations, (4) a limited number of kinetic equation forms are used in geochemical practice. This chapter describes the conceptual approach to converting a geochemical reaction network into a series of equilibrium, conservation and kinetic equations, the implementation of ECKEChem in STOMP, the numerical solution approach, and a demonstration of the simulator on a complex application involving desorption of uranium from contaminated field-textured sediments.
NASA Astrophysics Data System (ADS)
Powell, R. L.; Goulden, M.; Peterson, S.; Roberts, D. A.; Still, C. J.
2015-12-01
Temperature is a primary environmental control on biological systems and processes at a range of spatial and temporal scales, from controlling biochemical processes such as photosynthesis to influencing continental-scale species distribution. The Landsat satellite series provides a long record (since the mid-1980s) of relatively high spatial resolution thermal infrared (TIR) imagery, from which we derive land surface temperature (LST) grids. Here, we investigate fine spatial resolution factors that influence Landsat-derived LST over a spectrally and spatially heterogeneous landscape. We focus on paired sites (inside/outside a 1994 fire scar) within a pinyon-juniper scrubland in Southern California. The sites have nearly identical micro-meteorology and vegetation species composition, but distinctly different vegetation abundance and structure. The tower at the unburned site includes a number of in-situ imaging tools to quantify vegetation properties, including a thermal camera on a pan-tilt mount, allowing hourly characterization of landscape component temperatures (e.g., sunlit canopy, bare soil, leaf litter). We use these in-situ measurements to assess the impact of fine-scale landscape heterogeneity on estimates of LST, including sensitivity to (i) the relative abundance of component materials, (ii) directional effects due to solar and viewing geometry, (iii) duration of sunlit exposure for each compositional type, and (iv) air temperature. To scale these properties to Landsat spatial resolution (~100-m), we characterize the sub-pixel composition of landscape components (in addition to shade) by applying spectral mixture analysis (SMA) to the Landsat Operational Land Imager (OLI) spectral bands and test the sensitivity of the relationships established with the in-situ data at this coarser scale. The effects of vegetation abundance and cover height versus other controls on satellite-derived estimates of LST will be assessed by comparing estimates at the burned vs
NASA Astrophysics Data System (ADS)
Baghzouz, Malika
One of the most critical issues associated with using satellite data-based products to study and estimate surface energy fluxes and other ecosystem processes, has been the lack of frequent acquisition at a spatial scale equivalent to or finer than the footprint of field measurements. In this study, we incorporated continuous field measurements based on using Normalized difference vegetation index (NDVI) time series analysis of individual shrub species and transect measurements within 625 m2 size plots equivalent to the Landsat-5 Thematic Mapper spatial resolution. The NDVI system was a dual channel SKR-1800 radiometer that simultaneously measured incident solar radiation and upward reflectance in two broadband red and near-infrared channels comparable to Landsat-5 TM band 3 and band 4, respectively. The two study sites identified as Spring Valley 1 site (SV1) and Snake Valley 1 site (SNK1) were chosen for having different species composition, soil texture and percent canopy cover. NDVI time-series of greasewood (Sarcobatus vermiculatus) from the SV1 site allowed for clear distinction between the main phenological stages of the entire growing season during the period from January to November, 2007. Comparison of greasewood NDVI values between the two sites revealed a significant temporal difference associated with early canopy development and early dry down of greasewood at the SNK1 site. NDVI time series values were also significantly different between sagebrush (Artemisia tridentata ) and rabbitbrush (Chrysothamnus viscidiflorus) at SV1 as well as between the two bare soil types at the two sites, indicating the ability of the ground-based NDVI to distinguish between different plant species as well as between different desert soils based on their moisture level and color. The difference in phenological characteristics of greasewood between the two sites and between sagebrush, rabbitbrush and greasewood within the same site were not captured by the spatially
Moustafa, Gihan G.; Shaaban, F. E.; Hadeed, A. H. Abo; Elhady, Walaa M.
2016-01-01
Aim: The current study was directed to investigate the immunotoxic and oxidative stress effects of Roundup and Stomp herbicides and their combination on Nile catfish (Clarias gariepinus). Materials and Methods: The experiment was carried out on 120 fish that randomly divided into four equal groups with three replicates: The first group kept as control, the second group exposed to 1/2 96 h lethal concentration 50 (LC50) of Roundup, the third group exposed to 1/2 96 h LC50 of Stomp, and the fourth one exposed to a combination of Roundup and Stomp at previously-mentioned doses. The experiment was terminated after 15 days; blood samples were obtained at 1st, 8th, and 15th days of treatment where the sera were separated for estimation of antioxidant enzymes. Meanwhile, at 15th day of exposure part of blood was collected from all groups with an anticoagulant for evaluation of phagocytic activity, then the fish were sacrificed, and specimens from the liver of all groups were obtained for histopathological examination. Results: Our results indicated that both herbicides either individually or in combination elucidated significant decrease in phagocytic activity that was highly marked in group exposed to both herbicides. Furthermore, our data elicited an obvious elevation in the levels of superoxide dismutase (SOD), catalase (CAT), and glutathione peroxidase (GPx). Meanwhile, the data depicted reduction in levels of reduced glutathione (GSH) and glutathione-S-transferase (GST). Histopathological investigation of liver proved the aforementioned results. Conclusion: It could be concluded that either Roundup or Stomp alone cause significant deleterious effects on aquatic vertebrates. However, the use of their combination enhanced their toxic effects. Toxicity can end up in humans through the food chain. PMID:27397989
Using the STOMP (Seismic TOMography Program) Program for tomography with strong ray bending
Beatty, J.A.; Berryman, J.G.
1987-08-31
Accurate tomographic reconstructions of sound wave speed and attenuation are more difficult to obtain than are the corresponding reconstructions for x-rays or high frequency electromagnetic probes. The source of the difficulty is the common occurrence of large contrasts in acoustic or seismic wave speeds, leading to refraction and ray-bending effects. A new algorithm based on Fermat's principle has been developed to treat these problems. A description of the code STOMP (for Seismic TOMography Program) implementing the new algorithm is presented here together with a brief users manual for applications to borehole-to-borehole tomography. 3 refs., 3 figs.
Numerical Simulations of Urea Hydrolysis and Calcite Precipitation in Porous Media Using STOMP
Luanjing Guo; Hai Huang; Bill X. Hu
2010-11-01
Subsurface radionuclide and trace metal contaminants throughout the U.S. Department of Energy (DOE) complex pose one of DOE’s greatest challenges for long-term stewardship. One promising in situ immobilization approach of these contaminants is engineered mineral (co)precipitation of calcite driven by urea hydrolysis that is catalyzed by enzyme urease. The tight nonlinear coupling among flow, transport, reaction and reaction-induced property changes of media of this approach was studied by reactive transport simulations with systematically increasing level of complexities of reaction network and physical/chemical heterogeneities using a numerical simulator named STOMP. Sensitivity studies on the reaction rates of both urea hydrolysis and calcite precipitation are performed via controlling urease enzyme concentration and precipitation rate constant according to the rate models employed. We have found that the rate of ureolysis is a dominating factor in the amount of precipitated mineral; however, the spatial distribution of the precipitates depends on both rates of ureolysis and calcite precipitation. A maximum 5% reduction in the porosity was observed within the simulation time period of 6 pore volumes in our 1-dimensional (1D) column simulations. When a low permeability inclusion is considered in the 2D simulations, the altered flow fields redistribute mineral forming constituents, leading to a distorted precipitation reaction front. The simulations also indicate that mineral precipitation occurs along the boundary of the low permeability zone, which implies that contaminants in the low permeability zone could be encapsulated and isolated from the flow paths.
Sparse Regression as a Sparse Eigenvalue Problem
NASA Technical Reports Server (NTRS)
Moghaddam, Baback; Gruber, Amit; Weiss, Yair; Avidan, Shai
2008-01-01
We extend the l0-norm "subspectral" algorithms for sparse-LDA [5] and sparse-PCA [6] to general quadratic costs such as MSE in linear (kernel) regression. The resulting "Sparse Least Squares" (SLS) problem is also NP-hard, by way of its equivalence to a rank-1 sparse eigenvalue problem (e.g., binary sparse-LDA [7]). Specifically, for a general quadratic cost we use a highly-efficient technique for direct eigenvalue computation using partitioned matrix inverses which leads to dramatic x103 speed-ups over standard eigenvalue decomposition. This increased efficiency mitigates the O(n4) scaling behaviour that up to now has limited the previous algorithms' utility for high-dimensional learning problems. Moreover, the new computation prioritizes the role of the less-myopic backward elimination stage which becomes more efficient than forward selection. Similarly, branch-and-bound search for Exact Sparse Least Squares (ESLS) also benefits from partitioned matrix inverse techniques. Our Greedy Sparse Least Squares (GSLS) generalizes Natarajan's algorithm [9] also known as Order-Recursive Matching Pursuit (ORMP). Specifically, the forward half of GSLS is exactly equivalent to ORMP but more efficient. By including the backward pass, which only doubles the computation, we can achieve lower MSE than ORMP. Experimental comparisons to the state-of-the-art LARS algorithm [3] show forward-GSLS is faster, more accurate and more flexible in terms of choice of regularization
NASA Astrophysics Data System (ADS)
Carroll, K. C.; Nguyen, B. N.; Fang, Y.; Richmond, M. C.; Murray, C. J.
2011-12-01
Geomechanical alteration of porous media is generally ignored for most shallow subsurface applications, whereas CO2 injection, migration, and trapping in deep saline aquifers will be controlled by coupled multifluid flow, energy transfer, geomechanical, and geochemical processes. The accurate assessment of the risks associated with potential leakage of injected CO2 and the design of effective injection systems requires that we represent these coupled processes within numerical simulators. The objective of this study was to examine the coupling of hydraulic and geomechanical processes for simulation of CO2 injection into the subsurface for carbon sequestration. The impact of nonisothermal multifluid flow and porous media deformation mechanics on CO2 migration and storage was evaluated. We present a sequentially coupled approach for multifluid and geomechanical simulation using STOMP and ABAQUS that has been developed and validated through comparison to the solutions for benchmark problems that were solved with a coupled TOUGH-FLAC simulator. The poroelastic model was implemented with user-subroutines in ABAQUS. We also compare the STOMP-ABAQUS simulator to a new version of STOMP that includes the fully coupled poroelastic simulation within the multifluid flow and transport simulator. The poroelastic model computes stiffness, stresses, and strains using aqueous and gas pressures as well as saturations from STOMP output, and provides STOMP with the updated permeability, porosity, and capillary pressure over time during the simulation. The hydraulic only (uncoupled from mechanics) simulation and the hydrogeomechanical (coupled) simulation results using STOMP-ABAQUS were comparable to the previous results of a TOUGH-FLAC simulator. Results from the STOMP-ABAQUS coupled simulator were essentially identical to the fully coupled STOMP hydrogeomechanical simulator when the sequential coupling occurred at small time steps, and deviations between results increased with
A Modified Sparse Representation Method for Facial Expression Recognition
Wang, Wei; Xu, LiHong
2016-01-01
In this paper, we carry on research on a facial expression recognition method, which is based on modified sparse representation recognition (MSRR) method. On the first stage, we use Haar-like+LPP to extract feature and reduce dimension. On the second stage, we adopt LC-K-SVD (Label Consistent K-SVD) method to train the dictionary, instead of adopting directly the dictionary from samples, and add block dictionary training into the training process. On the third stage, stOMP (stagewise orthogonal matching pursuit) method is used to speed up the convergence of OMP (orthogonal matching pursuit). Besides, a dynamic regularization factor is added to iteration process to suppress noises and enhance accuracy. We verify the proposed method from the aspect of training samples, dimension, feature extraction and dimension reduction methods and noises in self-built database and Japan's JAFFE and CMU's CK database. Further, we compare this sparse method with classic SVM and RVM and analyze the recognition effect and time efficiency. The result of simulation experiment has shown that the coefficient of MSRR method contains classifying information, which is capable of improving the computing speed and achieving a satisfying recognition result. PMID:26880878
A Modified Sparse Representation Method for Facial Expression Recognition.
Wang, Wei; Xu, LiHong
2016-01-01
In this paper, we carry on research on a facial expression recognition method, which is based on modified sparse representation recognition (MSRR) method. On the first stage, we use Haar-like+LPP to extract feature and reduce dimension. On the second stage, we adopt LC-K-SVD (Label Consistent K-SVD) method to train the dictionary, instead of adopting directly the dictionary from samples, and add block dictionary training into the training process. On the third stage, stOMP (stagewise orthogonal matching pursuit) method is used to speed up the convergence of OMP (orthogonal matching pursuit). Besides, a dynamic regularization factor is added to iteration process to suppress noises and enhance accuracy. We verify the proposed method from the aspect of training samples, dimension, feature extraction and dimension reduction methods and noises in self-built database and Japan's JAFFE and CMU's CK database. Further, we compare this sparse method with classic SVM and RVM and analyze the recognition effect and time efficiency. The result of simulation experiment has shown that the coefficient of MSRR method contains classifying information, which is capable of improving the computing speed and achieving a satisfying recognition result. PMID:26880878
Ray, J.; Lee, J.; Yadav, V.; Lefantzi, S.; Michalak, A. M.; van Bloemen Waanders, B.
2014-08-20
We present a sparse reconstruction scheme that can also be used to ensure non-negativity when fitting wavelet-based random field models to limited observations in non-rectangular geometries. The method is relevant when multiresolution fields are estimated using linear inverse problems. Examples include the estimation of emission fields for many anthropogenic pollutants using atmospheric inversion or hydraulic conductivity in aquifers from flow measurements. The scheme is based on three new developments. Firstly, we extend an existing sparse reconstruction method, Stagewise Orthogonal Matching Pursuit (StOMP), to incorporate prior information on the target field. Secondly, we develop an iterative method that uses StOMP tomore » impose non-negativity on the estimated field. Finally, we devise a method, based on compressive sensing, to limit the estimated field within an irregularly shaped domain. We demonstrate the method on the estimation of fossil-fuel CO2 (ffCO2) emissions in the lower 48 states of the US. The application uses a recently developed multiresolution random field model and synthetic observations of ffCO2 concentrations from a limited set of measurement sites. We find that our method for limiting the estimated field within an irregularly shaped region is about a factor of 10 faster than conventional approaches. It also reduces the overall computational cost by a factor of two. Further, the sparse reconstruction scheme imposes non-negativity without introducing strong nonlinearities, such as those introduced by employing log-transformed fields, and thus reaps the benefits of simplicity and computational speed that are characteristic of linear inverse problems.« less
Sparse representation with kernels.
Gao, Shenghua; Tsang, Ivor Wai-Hung; Chia, Liang-Tien
2013-02-01
Recent research has shown the initial success of sparse coding (Sc) in solving many computer vision tasks. Motivated by the fact that kernel trick can capture the nonlinear similarity of features, which helps in finding a sparse representation of nonlinear features, we propose kernel sparse representation (KSR). Essentially, KSR is a sparse coding technique in a high dimensional feature space mapped by an implicit mapping function. We apply KSR to feature coding in image classification, face recognition, and kernel matrix approximation. More specifically, by incorporating KSR into spatial pyramid matching (SPM), we develop KSRSPM, which achieves a good performance for image classification. Moreover, KSR-based feature coding can be shown as a generalization of efficient match kernel and an extension of Sc-based SPM. We further show that our proposed KSR using a histogram intersection kernel (HIK) can be considered a soft assignment extension of HIK-based feature quantization in the feature coding process. Besides feature coding, comparing with sparse coding, KSR can learn more discriminative sparse codes and achieve higher accuracy for face recognition. Moreover, KSR can also be applied to kernel matrix approximation in large scale learning tasks, and it demonstrates its robustness to kernel matrix approximation, especially when a small fraction of the data is used. Extensive experimental results demonstrate promising results of KSR in image classification, face recognition, and kernel matrix approximation. All these applications prove the effectiveness of KSR in computer vision and machine learning tasks. PMID:23014744
sparse-msrf:A package for sparse modeling and estimation of fossil-fuel CO2 emission fields
2014-10-06
The software is used to fit models of emission fields (e.g., fossil-fuel CO2 emissions) to sparse measurements of gaseous concentrations. Its primary aim is to provide an implementation and a demonstration for the algorithms and models developed in J. Ray, V. Yadav, A. M. Michalak, B. van Bloemen Waanders and S. A. McKenna, "A multiresolution spatial parameterization for the estimation of fossil-fuel carbon dioxide emissions via atmospheric inversions", accepted, Geoscientific Model Development, 2014. The software can be used to estimate emissions of non-reactive gases such as fossil-fuel CO2, methane etc. The software uses a proxy of the emission field being estimated (e.g., for fossil-fuel CO2, a population density map is a good proxy) to construct a wavelet model for the emission field. It then uses a shrinkage regression algorithm called Stagewise Orthogonal Matching Pursuit (StOMP) to fit the wavelet model to concentration measurements, using an atmospheric transport model to relate emission and concentration fields. Algorithmic novelties described in the paper above (1) ensure that the estimated emission fields are non-negative, (2) allow the use of guesses for emission fields to accelerate the estimation processes and (3) ensure that under/overestimates in the guesses do not skew the estimation.
sparse-msrf:A package for sparse modeling and estimation of fossil-fuel CO2 emission fields
Energy Science and Technology Software Center (ESTSC)
2014-10-06
The software is used to fit models of emission fields (e.g., fossil-fuel CO2 emissions) to sparse measurements of gaseous concentrations. Its primary aim is to provide an implementation and a demonstration for the algorithms and models developed in J. Ray, V. Yadav, A. M. Michalak, B. van Bloemen Waanders and S. A. McKenna, "A multiresolution spatial parameterization for the estimation of fossil-fuel carbon dioxide emissions via atmospheric inversions", accepted, Geoscientific Model Development, 2014. The softwaremore » can be used to estimate emissions of non-reactive gases such as fossil-fuel CO2, methane etc. The software uses a proxy of the emission field being estimated (e.g., for fossil-fuel CO2, a population density map is a good proxy) to construct a wavelet model for the emission field. It then uses a shrinkage regression algorithm called Stagewise Orthogonal Matching Pursuit (StOMP) to fit the wavelet model to concentration measurements, using an atmospheric transport model to relate emission and concentration fields. Algorithmic novelties described in the paper above (1) ensure that the estimated emission fields are non-negative, (2) allow the use of guesses for emission fields to accelerate the estimation processes and (3) ensure that under/overestimates in the guesses do not skew the estimation.« less
Sparse distributed memory overview
NASA Technical Reports Server (NTRS)
Raugh, Mike
1990-01-01
The Sparse Distributed Memory (SDM) project is investigating the theory and applications of massively parallel computing architecture, called sparse distributed memory, that will support the storage and retrieval of sensory and motor patterns characteristic of autonomous systems. The immediate objectives of the project are centered in studies of the memory itself and in the use of the memory to solve problems in speech, vision, and robotics. Investigation of methods for encoding sensory data is an important part of the research. Examples of NASA missions that may benefit from this work are Space Station, planetary rovers, and solar exploration. Sparse distributed memory offers promising technology for systems that must learn through experience and be capable of adapting to new circumstances, and for operating any large complex system requiring automatic monitoring and control. Sparse distributed memory is a massively parallel architecture motivated by efforts to understand how the human brain works. Sparse distributed memory is an associative memory, able to retrieve information from cues that only partially match patterns stored in the memory. It is able to store long temporal sequences derived from the behavior of a complex system, such as progressive records of the system's sensory data and correlated records of the system's motor controls.
Grassmannian sparse representations
NASA Astrophysics Data System (ADS)
Azary, Sherif; Savakis, Andreas
2015-05-01
We present Grassmannian sparse representations (GSR), a sparse representation Grassmann learning framework for efficient classification. Sparse representation classification offers a powerful approach for recognition in a variety of contexts. However, a major drawback of sparse representation methods is their computational performance and memory utilization for high-dimensional data. A Grassmann manifold is a space that promotes smooth surfaces where points represent subspaces and the relationship between points is defined by the mapping of an orthogonal matrix. Grassmann manifolds are well suited for computer vision problems because they promote high between-class discrimination and within-class clustering, while offering computational advantages by mapping each subspace onto a single point. The GSR framework combines Grassmannian kernels and sparse representations, including regularized least squares and least angle regression, to improve high accuracy recognition while overcoming the drawbacks of performance and dependencies on high dimensional data distributions. The effectiveness of GSR is demonstrated on computationally intensive multiview action sequences, three-dimensional action sequences, and face recognition datasets.
Ray, J.; Lee, J.; Yadav, V.; Lefantzi, S.; Michalak, A. M.; van Bloemen Waanders, B.
2015-04-29
Atmospheric inversions are frequently used to estimate fluxes of atmospheric greenhouse gases (e.g., biospheric CO2 flux fields) at Earth's surface. These inversions typically assume that flux departures from a prior model are spatially smoothly varying, which are then modeled using a multi-variate Gaussian. When the field being estimated is spatially rough, multi-variate Gaussian models are difficult to construct and a wavelet-based field model may be more suitable. Unfortunately, such models are very high dimensional and are most conveniently used when the estimation method can simultaneously perform data-driven model simplification (removal of model parameters that cannot be reliably estimated) and fitting.more » Such sparse reconstruction methods are typically not used in atmospheric inversions. In this work, we devise a sparse reconstruction method, and illustrate it in an idealized atmospheric inversion problem for the estimation of fossil fuel CO2 (ffCO2) emissions in the lower 48 states of the USA. Our new method is based on stagewise orthogonal matching pursuit (StOMP), a method used to reconstruct compressively sensed images. Our adaptations bestow three properties to the sparse reconstruction procedure which are useful in atmospheric inversions. We have modified StOMP to incorporate prior information on the emission field being estimated and to enforce non-negativity on the estimated field. Finally, though based on wavelets, our method allows for the estimation of fields in non-rectangular geometries, e.g., emission fields inside geographical and political boundaries. Our idealized inversions use a recently developed multi-resolution (i.e., wavelet-based) random field model developed for ffCO2 emissions and synthetic observations of ffCO2 concentrations from a limited set of measurement sites. We find that our method for limiting the estimated field within an irregularly shaped region is about a factor of 10 faster than conventional approaches. It also
LOFAR sparse image reconstruction
NASA Astrophysics Data System (ADS)
Garsden, H.; Girard, J. N.; Starck, J. L.; Corbel, S.; Tasse, C.; Woiselle, A.; McKean, J. P.; van Amesfoort, A. S.; Anderson, J.; Avruch, I. M.; Beck, R.; Bentum, M. J.; Best, P.; Breitling, F.; Broderick, J.; Brüggen, M.; Butcher, H. R.; Ciardi, B.; de Gasperin, F.; de Geus, E.; de Vos, M.; Duscha, S.; Eislöffel, J.; Engels, D.; Falcke, H.; Fallows, R. A.; Fender, R.; Ferrari, C.; Frieswijk, W.; Garrett, M. A.; Grießmeier, J.; Gunst, A. W.; Hassall, T. E.; Heald, G.; Hoeft, M.; Hörandel, J.; van der Horst, A.; Juette, E.; Karastergiou, A.; Kondratiev, V. I.; Kramer, M.; Kuniyoshi, M.; Kuper, G.; Mann, G.; Markoff, S.; McFadden, R.; McKay-Bukowski, D.; Mulcahy, D. D.; Munk, H.; Norden, M. J.; Orru, E.; Paas, H.; Pandey-Pommier, M.; Pandey, V. N.; Pietka, G.; Pizzo, R.; Polatidis, A. G.; Renting, A.; Röttgering, H.; Rowlinson, A.; Schwarz, D.; Sluman, J.; Smirnov, O.; Stappers, B. W.; Steinmetz, M.; Stewart, A.; Swinbank, J.; Tagger, M.; Tang, Y.; Tasse, C.; Thoudam, S.; Toribio, C.; Vermeulen, R.; Vocks, C.; van Weeren, R. J.; Wijnholds, S. J.; Wise, M. W.; Wucknitz, O.; Yatawatta, S.; Zarka, P.; Zensus, A.
2015-03-01
Context. The LOw Frequency ARray (LOFAR) radio telescope is a giant digital phased array interferometer with multiple antennas distributed in Europe. It provides discrete sets of Fourier components of the sky brightness. Recovering the original brightness distribution with aperture synthesis forms an inverse problem that can be solved by various deconvolution and minimization methods. Aims: Recent papers have established a clear link between the discrete nature of radio interferometry measurement and the "compressed sensing" (CS) theory, which supports sparse reconstruction methods to form an image from the measured visibilities. Empowered by proximal theory, CS offers a sound framework for efficient global minimization and sparse data representation using fast algorithms. Combined with instrumental direction-dependent effects (DDE) in the scope of a real instrument, we developed and validated a new method based on this framework. Methods: We implemented a sparse reconstruction method in the standard LOFAR imaging tool and compared the photometric and resolution performance of this new imager with that of CLEAN-based methods (CLEAN and MS-CLEAN) with simulated and real LOFAR data. Results: We show that i) sparse reconstruction performs as well as CLEAN in recovering the flux of point sources; ii) performs much better on extended objects (the root mean square error is reduced by a factor of up to 10); and iii) provides a solution with an effective angular resolution 2-3 times better than the CLEAN images. Conclusions: Sparse recovery gives a correct photometry on high dynamic and wide-field images and improved realistic structures of extended sources (of simulated and real LOFAR datasets). This sparse reconstruction method is compatible with modern interferometric imagers that handle DDE corrections (A- and W-projections) required for current and future instruments such as LOFAR and SKA.
Banerjee, Abhirup; Maji, Pradipta
2015-12-01
The segmentation of brain MR images into different tissue classes is an important task for automatic image analysis technique, particularly due to the presence of intensity inhomogeneity artifact in MR images. In this regard, this paper presents a novel approach for simultaneous segmentation and bias field correction in brain MR images. It integrates judiciously the concept of rough sets and the merit of a novel probability distribution, called stomped normal (SN) distribution. The intensity distribution of a tissue class is represented by SN distribution, where each tissue class consists of a crisp lower approximation and a probabilistic boundary region. The intensity distribution of brain MR image is modeled as a mixture of finite number of SN distributions and one uniform distribution. The proposed method incorporates both the expectation-maximization and hidden Markov random field frameworks to provide an accurate and robust segmentation. The performance of the proposed approach, along with a comparison with related methods, is demonstrated on a set of synthetic and real brain MR images for different bias fields and noise levels. PMID:26462197
Structured Multifrontal Sparse Solver
Energy Science and Technology Software Center (ESTSC)
2014-05-01
StruMF is an algebraic structured preconditioner for the interative solution of large sparse linear systems. The preconditioner corresponds to a multifrontal variant of sparse LU factorization in which some dense blocks of the factors are approximated with low-rank matrices. It is algebraic in that it only requires the linear system itself, and the approximation threshold that determines the accuracy of individual low-rank approximations. Favourable rank properties are obtained using a block partitioning which is amore » refinement of the partitioning induced by nested dissection ordering.« less
Sparse inpainting and isotropy
Feeney, Stephen M.; McEwen, Jason D.; Peiris, Hiranya V.; Marinucci, Domenico; Cammarota, Valentina; Wandelt, Benjamin D. E-mail: marinucc@axp.mat.uniroma2.it E-mail: h.peiris@ucl.ac.uk E-mail: cammarot@axp.mat.uniroma2.it
2014-01-01
Sparse inpainting techniques are gaining in popularity as a tool for cosmological data analysis, in particular for handling data which present masked regions and missing observations. We investigate here the relationship between sparse inpainting techniques using the spherical harmonic basis as a dictionary and the isotropy properties of cosmological maps, as for instance those arising from cosmic microwave background (CMB) experiments. In particular, we investigate the possibility that inpainted maps may exhibit anisotropies in the behaviour of higher-order angular polyspectra. We provide analytic computations and simulations of inpainted maps for a Gaussian isotropic model of CMB data, suggesting that the resulting angular trispectrum may exhibit small but non-negligible deviations from isotropy.
Sparse matrix test collections
Duff, I.
1996-12-31
This workshop will discuss plans for coordinating and developing sets of test matrices for the comparison and testing of sparse linear algebra software. We will talk of plans for the next release (Release 2) of the Harwell-Boeing Collection and recent work on improving the accessibility of this Collection and others through the World Wide Web. There will only be three talks of about 15 to 20 minutes followed by a discussion from the floor.
NASA Technical Reports Server (NTRS)
Kanerva, Pentti
1988-01-01
Theoretical models of the human brain and proposed neural-network computers are developed analytically. Chapters are devoted to the mathematical foundations, background material from computer science, the theory of idealized neurons, neurons as address decoders, and the search of memory for the best match. Consideration is given to sparse memory, distributed storage, the storage and retrieval of sequences, the construction of distributed memory, and the organization of an autonomous learning system.
Kanerva, P.
1988-01-01
Theoretical models of the human brain and proposed neural-network computers are developed analytically. Chapters are devoted to the mathematical foundations, background material from computer science, the theory of idealized neurons, neurons as address decoders, and the search of memory for the best match. Consideration is given to sparse memory, distributed storage, the storage and retrieval of sequences, the construction of distributed memory, and the organization of an autonomous learning system. 63 refs.
Optical sparse aperture imaging.
Miller, Nicholas J; Dierking, Matthew P; Duncan, Bradley D
2007-08-10
The resolution of a conventional diffraction-limited imaging system is proportional to its pupil diameter. A primary goal of sparse aperture imaging is to enhance resolution while minimizing the total light collection area; the latter being desirable, in part, because of the cost of large, monolithic apertures. Performance metrics are defined and used to evaluate several sparse aperture arrays constructed from multiple, identical, circular subapertures. Subaperture piston and/or tilt effects on image quality are also considered. We selected arrays with compact nonredundant autocorrelations first described by Golay. We vary both the number of subapertures and their relative spacings to arrive at an optimized array. We report the results of an experiment in which we synthesized an image from multiple subaperture pupil fields by masking a large lens with a Golay array. For this experiment we imaged a slant edge feature of an ISO12233 resolution target in order to measure the modulation transfer function. We note the contrast reduction inherent in images formed through sparse aperture arrays and demonstrate the use of a Wiener-Helstrom filter to restore contrast in our experimental images. Finally, we describe a method to synthesize images from multiple subaperture focal plane intensity images using a phase retrieval algorithm to obtain estimates of subaperture pupil fields. Experimental results from synthesizing an image of a point object from multiple subaperture images are presented, and weaknesses of the phase retrieval method for this application are discussed. PMID:17694146
Energy Science and Technology Software Center (ESTSC)
2007-04-12
The Sparse Image Format (SIF) is a file format for storing spare raster images. It works by breaking an image down into tiles. Space is savid by only storing non-uniform tiles, i.e. tiles with at least two different pixel values. If a tile is completely uniform, its common pixel value is stored instead of the complete tile raster. The software is a library in the C language used for manipulating files in SIF format. Itmore » supports large files (> 2GB) and is designed to build in Windows and Linux environments.« less
Eads, Damian Ryan
2007-04-12
The Sparse Image Format (SIF) is a file format for storing spare raster images. It works by breaking an image down into tiles. Space is savid by only storing non-uniform tiles, i.e. tiles with at least two different pixel values. If a tile is completely uniform, its common pixel value is stored instead of the complete tile raster. The software is a library in the C language used for manipulating files in SIF format. It supports large files (> 2GB) and is designed to build in Windows and Linux environments.
and Drayton Munster, Miroslav Stoyanov
2013-09-20
Sparse Grids are the family of methods of choice for multidimensional integration and interpolation in low to moderate number of dimensions. The method is to select extend a one dimensional set of abscissas, weights and basis functions by taking a subset of all possible tensor products. The module provides the ability to create global and local approximations based on polynomials and wavelets. The software has three components, a library, a wrapper for the library that provides a command line interface via text files ad a MATLAB interface via the command line tool.
Energy Science and Technology Software Center (ESTSC)
2013-09-20
Sparse Grids are the family of methods of choice for multidimensional integration and interpolation in low to moderate number of dimensions. The method is to select extend a one dimensional set of abscissas, weights and basis functions by taking a subset of all possible tensor products. The module provides the ability to create global and local approximations based on polynomials and wavelets. The software has three components, a library, a wrapper for the library thatmore » provides a command line interface via text files ad a MATLAB interface via the command line tool.« less
Compressed Sampling of Spectrally Sparse Signals Using Sparse Circulant Matrices
NASA Astrophysics Data System (ADS)
Xu, Guangjie; Wang, Huali; Sun, Lei; Zeng, Weijun; Wang, Qingguo
2014-11-01
Circulant measurement matrices constructed by partial cyclically shifts of one generating sequence, are easier to be implemented in hardware than widely used random measurement matrices; however, the diminishment of randomness makes it more sensitive to signal noise. Selecting a deterministic sequence with optimal periodic autocorrelation property (PACP) as generating sequence, would enhance the noise robustness of circulant measurement matrix, but this kind of deterministic circulant matrices only exists in the fixed periodic length. Actually, the selection of generating sequence doesn't affect the compressive performance of circulant measurement matrix but the subspace energy in spectrally sparse signals. Sparse circulant matrices, whose generating sequence is a sparse sequence, could keep the energy balance of subspaces and have similar noise robustness to deterministic circulant matrices. In addition, sparse circulant matrices have no restriction on length and are more suitable for the compressed sampling of spectrally sparse signals at arbitrary dimensionality.
NASA Technical Reports Server (NTRS)
Denning, Peter J.
1989-01-01
Sparse distributed memory was proposed be Pentti Kanerva as a realizable architecture that could store large patterns and retrieve them based on partial matches with patterns representing current sensory inputs. This memory exhibits behaviors, both in theory and in experiment, that resemble those previously unapproached by machines - e.g., rapid recognition of faces or odors, discovery of new connections between seemingly unrelated ideas, continuation of a sequence of events when given a cue from the middle, knowing that one doesn't know, or getting stuck with an answer on the tip of one's tongue. These behaviors are now within reach of machines that can be incorporated into the computing systems of robots capable of seeing, talking, and manipulating. Kanerva's theory is a break with the Western rationalistic tradition, allowing a new interpretation of learning and cognition that respects biology and the mysteries of individual human beings.
Percolation on Sparse Networks
NASA Astrophysics Data System (ADS)
Karrer, Brian; Newman, M. E. J.; Zdeborová, Lenka
2014-11-01
We study percolation on networks, which is used as a model of the resilience of networked systems such as the Internet to attack or failure and as a simple model of the spread of disease over human contact networks. We reformulate percolation as a message passing process and demonstrate how the resulting equations can be used to calculate, among other things, the size of the percolating cluster and the average cluster size. The calculations are exact for sparse networks when the number of short loops in the network is small, but even on networks with many short loops we find them to be highly accurate when compared with direct numerical simulations. By considering the fixed points of the message passing process, we also show that the percolation threshold on a network with few loops is given by the inverse of the leading eigenvalue of the so-called nonbacktracking matrix.
White, Mark D.; McGrail, B. Peter
2005-12-01
flow and transport simulator, STOMP (Subsurface Transport Over Multiple Phases). Prior to these code development activities, the STOMP simulator included sequential and scalable implementations for numerically simulating the injection of supercritical CO2 into deep saline aquifers. Additionally, the sequential implementations included operational modes that considered nonisothermal conditions and kinetic dissolution of CO2 into the saline aqueous phase. This addendum documents the advancement of these numerical simulation capabilities to include reactive transport in the STOMP simulator through the inclusion of the recently PNNL developed batch geochemistry solution module ECKEChem (Equilibrium-Conservation-Kinetic Equation Chemistry). Potential geologic reservoirs for sequestering CO2 include deep saline aquifers, hydrate-bearing formations, depleted or partially depleted natural gas and petroleum reservoirs, and coal beds. The mechanisms for sequestering carbon dioxide in geologic reservoirs include physical trapping, dissolution in the reservoir fluids, hydraulic trapping (hysteretic entrapment of nonwetting fluids), and chemical reaction. This document and the associated code development and verification work are concerned with the chemistry of injecting CO2 into geologic reservoirs. As geologic sequestration of CO2 via chemical reaction, namely precipitation reactions, are most dominate in deep saline aquifers, the principal focus of this document is the numerical simulation of CO2 injection, migration, and geochemical reaction in deep saline aquifers. The ECKEChem batch chemistry module was developed in a fashion that would allow its implementation into all operational modes of the STOMP simulator, making it a more versatile chemistry component. Additionally, this approach allows for verification of the ECKEChem module against more classical reactive transport problems involving aqueous systems.
James, Gareth M.; Sabatti, Chiara; Zhou, Nengfeng; Zhu, Ji
2011-01-01
In many organisms the expression levels of each gene are controlled by the activation levels of known “Transcription Factors” (TF). A problem of considerable interest is that of estimating the “Transcription Regulation Networks” (TRN) relating the TFs and genes. While the expression levels of genes can be observed, the activation levels of the corresponding TFs are usually unknown, greatly increasing the difficulty of the problem. Based on previous experimental work, it is often the case that partial information about the TRN is available. For example, certain TFs may be known to regulate a given gene or in other cases a connection may be predicted with a certain probability. In general, the biology of the problem indicates there will be very few connections between TFs and genes. Several methods have been proposed for estimating TRNs. However, they all suffer from problems such as unrealistic assumptions about prior knowledge of the network structure or computational limitations. We propose a new approach that can directly utilize prior information about the network structure in conjunction with observed gene expression data to estimate the TRN. Our approach uses L1 penalties on the network to ensure a sparse structure. This has the advantage of being computationally efficient as well as making many fewer assumptions about the network structure. We use our methodology to construct the TRN for E. coli and show that the estimate is biologically sensible and compares favorably with previous estimates. PMID:21625366
Estimating sparse precision matrices
NASA Astrophysics Data System (ADS)
Padmanabhan, Nikhil; White, Martin; Zhou, Harrison H.; O'Connell, Ross
2016-08-01
We apply a method recently introduced to the statistical literature to directly estimate the precision matrix from an ensemble of samples drawn from a corresponding Gaussian distribution. Motivated by the observation that cosmological precision matrices are often approximately sparse, the method allows one to exploit this sparsity of the precision matrix to more quickly converge to an asymptotic 1/sqrt{N_sim} rate while simultaneously providing an error model for all of the terms. Such an estimate can be used as the starting point for further regularization efforts which can improve upon the 1/sqrt{N_sim} limit above, and incorporating such additional steps is straightforward within this framework. We demonstrate the technique with toy models and with an example motivated by large-scale structure two-point analysis, showing significant improvements in the rate of convergence. For the large-scale structure example, we find errors on the precision matrix which are factors of 5 smaller than for the sample precision matrix for thousands of simulations or, alternatively, convergence to the same error level with more than an order of magnitude fewer simulations.
Estimating sparse precision matrices
NASA Astrophysics Data System (ADS)
Padmanabhan, Nikhil; White, Martin; Zhou, Harrison H.; O'Connell, Ross
2016-05-01
We apply a method recently introduced to the statistical literature to directly estimate the precision matrix from an ensemble of samples drawn from a corresponding Gaussian distribution. Motivated by the observation that cosmological precision matrices are often approximately sparse, the method allows one to exploit this sparsity of the precision matrix to more quickly converge to an asymptotic 1/√{N_sim} rate while simultaneously providing an error model for all of the terms. Such an estimate can be used as the starting point for further regularization efforts which can improve upon the 1/√{N_sim} limit above, and incorporating such additional steps is straightforward within this framework. We demonstrate the technique with toy models and with an example motivated by large-scale structure two-point analysis, showing significant improvements in the rate of convergence. For the large-scale structure example we find errors on the precision matrix which are factors of 5 smaller than for the sample precision matrix for thousands of simulations or, alternatively, convergence to the same error level with more than an order of magnitude fewer simulations.
Estimating sparse precision matrices
NASA Astrophysics Data System (ADS)
Padmanabhan, Nikhil; White, Martin; Zhou, Harrison H.; O'Connell, Ross
2016-08-01
We apply a method recently introduced to the statistical literature to directly estimate the precision matrix from an ensemble of samples drawn from a corresponding Gaussian distribution. Motivated by the observation that cosmological precision matrices are often approximately sparse, the method allows one to exploit this sparsity of the precision matrix to more quickly converge to an asymptotic 1/√{N_sim} rate while simultaneously providing an error model for all of the terms. Such an estimate can be used as the starting point for further regularization efforts which can improve upon the 1/√{N_sim} limit above, and incorporating such additional steps is straightforward within this framework. We demonstrate the technique with toy models and with an example motivated by large-scale structure two-point analysis, showing significant improvements in the rate of convergence. For the large-scale structure example, we find errors on the precision matrix which are factors of 5 smaller than for the sample precision matrix for thousands of simulations or, alternatively, convergence to the same error level with more than an order of magnitude fewer simulations.
Completeness for sparse potential scattering
Shen, Zhongwei
2014-01-15
The present paper is devoted to the scattering theory of a class of continuum Schrödinger operators with deterministic sparse potentials. We first establish the limiting absorption principle for both modified free resolvents and modified perturbed resolvents. This actually is a weak form of the classical limiting absorption principle. We then prove the existence and completeness of local wave operators, which, in particular, imply the existence of wave operators. Under additional assumptions on the sparse potential, we prove the completeness of wave operators. In the context of continuum Schrödinger operators with sparse potentials, this paper gives the first proof of the completeness of wave operators.
Threaded Operations on Sparse Matrices
Sneed, Brett
2015-09-01
We investigate the use of sparse matrices and OpenMP multi-threading on linear algebra operations involving them. Several sparse matrix data structures are presented. Implementation of the multi- threading primarily occurs in the level one and two BLAS functions used within the four algorithms investigated{the Power Method, Conjugate Gradient, Biconjugate Gradient, and Jacobi's Method. The bene ts of launching threads once per high level algorithm are explored.
Learning Sparse Representations of Depth
NASA Astrophysics Data System (ADS)
Tosic, Ivana; Olshausen, Bruno A.; Culpepper, Benjamin J.
2011-09-01
This paper introduces a new method for learning and inferring sparse representations of depth (disparity) maps. The proposed algorithm relaxes the usual assumption of the stationary noise model in sparse coding. This enables learning from data corrupted with spatially varying noise or uncertainty, typically obtained by laser range scanners or structured light depth cameras. Sparse representations are learned from the Middlebury database disparity maps and then exploited in a two-layer graphical model for inferring depth from stereo, by including a sparsity prior on the learned features. Since they capture higher-order dependencies in the depth structure, these priors can complement smoothness priors commonly used in depth inference based on Markov Random Field (MRF) models. Inference on the proposed graph is achieved using an alternating iterative optimization technique, where the first layer is solved using an existing MRF-based stereo matching algorithm, then held fixed as the second layer is solved using the proposed non-stationary sparse coding algorithm. This leads to a general method for improving solutions of state of the art MRF-based depth estimation algorithms. Our experimental results first show that depth inference using learned representations leads to state of the art denoising of depth maps obtained from laser range scanners and a time of flight camera. Furthermore, we show that adding sparse priors improves the results of two depth estimation methods: the classical graph cut algorithm by Boykov et al. and the more recent algorithm of Woodford et al.
Wavelet Sparse Approximate Inverse Preconditioners
NASA Technical Reports Server (NTRS)
Chan, Tony F.; Tang, W.-P.; Wan, W. L.
1996-01-01
There is an increasing interest in using sparse approximate inverses as preconditioners for Krylov subspace iterative methods. Recent studies of Grote and Huckle and Chow and Saad also show that sparse approximate inverse preconditioner can be effective for a variety of matrices, e.g. Harwell-Boeing collections. Nonetheless a drawback is that it requires rapid decay of the inverse entries so that sparse approximate inverse is possible. However, for the class of matrices that, come from elliptic PDE problems, this assumption may not necessarily hold. Our main idea is to look for a basis, other than the standard one, such that a sparse representation of the inverse is feasible. A crucial observation is that the kind of matrices we are interested in typically have a piecewise smooth inverse. We exploit this fact, by applying wavelet techniques to construct a better sparse approximate inverse in the wavelet basis. We shall justify theoretically and numerically that our approach is effective for matrices with smooth inverse. We emphasize that in this paper we have only presented the idea of wavelet approximate inverses and demonstrated its potential but have not yet developed a highly refined and efficient algorithm.
A scalable 2-D parallel sparse solver
Kothari, S.C.; Mitra, S.
1995-12-01
Scalability beyond a small number of processors, typically 32 or less, is known to be a problem for existing parallel general sparse (PGS) direct solvers. This paper presents a parallel general sparse PGS direct solver for general sparse linear systems on distributed memory machines. The algorithm is based on the well-known sequential sparse algorithm Y12M. To achieve efficient parallelization, a 2-D scattered decomposition of the sparse matrix is used. The proposed algorithm is more scalable than existing parallel sparse direct solvers. Its scalability is evaluated on a 256 processor nCUBE2s machine using Boeing/Harwell benchmark matrices.
Amesos2 Templated Direct Sparse Solver Package
Energy Science and Technology Software Center (ESTSC)
2011-05-24
Amesos2 is a templated direct sparse solver package. Amesos2 provides interfaces to direct sparse solvers, rather than providing native solver capabilities. Amesos2 is a derivative work of the Trilinos package Amesos.
SparsePZ: Sparse Representation of Photometric Redshift PDFs
NASA Astrophysics Data System (ADS)
Carrasco Kind, Matias; Brunner, R. J.
2015-11-01
SparsePZ uses sparse basis representation to fully represent individual photometric redshift probability density functions (PDFs). This approach requires approximately half the parameters for the same multi-Gaussian fitting accuracy, and has the additional advantage that an entire PDF can be stored by using a 4-byte integer per basis function. Only 10-20 points per galaxy are needed to reconstruct both the individual PDFs and the ensemble redshift distribution, N(z), to an accuracy of 99.9 per cent when compared to the one built using the original PDFs computed with a resolution of δz = 0.01, reducing the required storage of 200 original values by a factor of 10-20. This basis representation can be directly extended to a cosmological analysis, thereby increasing computational performance without losing resolution or accuracy.
Sparse Biclustering of Transposable Data
Tan, Kean Ming
2013-01-01
We consider the task of simultaneously clustering the rows and columns of a large transposable data matrix. We assume that the matrix elements are normally distributed with a bicluster-specific mean term and a common variance, and perform biclustering by maximizing the corresponding log likelihood. We apply an ℓ1 penalty to the means of the biclusters in order to obtain sparse and interpretable biclusters. Our proposal amounts to a sparse, symmetrized version of k-means clustering. We show that k-means clustering of the rows and of the columns of a data matrix can be seen as special cases of our proposal, and that a relaxation of our proposal yields the singular value decomposition. In addition, we propose a framework for bi-clustering based on the matrix-variate normal distribution. The performances of our proposals are demonstrated in a simulation study and on a gene expression data set. This article has supplementary material online. PMID:25364221
Sparse Biclustering of Transposable Data.
Tan, Kean Ming; Witten, Daniela M
2014-01-01
We consider the task of simultaneously clustering the rows and columns of a large transposable data matrix. We assume that the matrix elements are normally distributed with a bicluster-specific mean term and a common variance, and perform biclustering by maximizing the corresponding log likelihood. We apply an ℓ1 penalty to the means of the biclusters in order to obtain sparse and interpretable biclusters. Our proposal amounts to a sparse, symmetrized version of k-means clustering. We show that k-means clustering of the rows and of the columns of a data matrix can be seen as special cases of our proposal, and that a relaxation of our proposal yields the singular value decomposition. In addition, we propose a framework for bi-clustering based on the matrix-variate normal distribution. The performances of our proposals are demonstrated in a simulation study and on a gene expression data set. This article has supplementary material online. PMID:25364221
Finding communities in sparse networks
Singh, Abhinav; Humphries, Mark D.
2015-01-01
Spectral algorithms based on matrix representations of networks are often used to detect communities, but classic spectral methods based on the adjacency matrix and its variants fail in sparse networks. New spectral methods based on non-backtracking random walks have recently been introduced that successfully detect communities in many sparse networks. However, the spectrum of non-backtracking random walks ignores hanging trees in networks that can contain information about their community structure. We introduce the reluctant backtracking operators that explicitly account for hanging trees as they admit a small probability of returning to the immediately previous node, unlike the non-backtracking operators that forbid an immediate return. We show that the reluctant backtracking operators can detect communities in certain sparse networks where the non-backtracking operators cannot, while performing comparably on benchmark stochastic block model networks and real world networks. We also show that the spectrum of the reluctant backtracking operator approximately optimises the standard modularity function. Interestingly, for this family of non- and reluctant-backtracking operators the main determinant of performance on real-world networks is whether or not they are normalised to conserve probability at each node. PMID:25742951
Highly parallel sparse Cholesky factorization
NASA Technical Reports Server (NTRS)
Gilbert, John R.; Schreiber, Robert
1990-01-01
Several fine grained parallel algorithms were developed and compared to compute the Cholesky factorization of a sparse matrix. The experimental implementations are on the Connection Machine, a distributed memory SIMD machine whose programming model conceptually supplies one processor per data element. In contrast to special purpose algorithms in which the matrix structure conforms to the connection structure of the machine, the focus is on matrices with arbitrary sparsity structure. The most promising algorithm is one whose inner loop performs several dense factorizations simultaneously on a 2-D grid of processors. Virtually any massively parallel dense factorization algorithm can be used as the key subroutine. The sparse code attains execution rates comparable to those of the dense subroutine. Although at present architectural limitations prevent the dense factorization from realizing its potential efficiency, it is concluded that a regular data parallel architecture can be used efficiently to solve arbitrarily structured sparse problems. A performance model is also presented and it is used to analyze the algorithms.
Sparse Matrices in MATLAB: Design and Implementation
NASA Technical Reports Server (NTRS)
Gilbert, John R.; Moler, Cleve; Schreiber, Robert
1992-01-01
The matrix computation language and environment MATLAB is extended to include sparse matrix storage and operations. The only change to the outward appearance of the MATLAB language is a pair of commands to create full or sparse matrices. Nearly all the operations of MATLAB now apply equally to full or sparse matrices, without any explicit action by the user. The sparse data structure represents a matrix in space proportional to the number of nonzero entries, and most of the operations compute sparse results in time proportional to the number of arithmetic operations on nonzeros.
Sparse Coding for Alpha Matting
NASA Astrophysics Data System (ADS)
Johnson, Jubin; Varnousfaderani, Ehsan Shahrian; Cholakkal, Hisham; Rajan, Deepu
2016-07-01
Existing color sampling based alpha matting methods use the compositing equation to estimate alpha at a pixel from pairs of foreground (F) and background (B) samples. The quality of the matte depends on the selected (F,B) pairs. In this paper, the matting problem is reinterpreted as a sparse coding of pixel features, wherein the sum of the codes gives the estimate of the alpha matte from a set of unpaired F and B samples. A non-parametric probabilistic segmentation provides a certainty measure on the pixel belonging to foreground or background, based on which a dictionary is formed for use in sparse coding. By removing the restriction to conform to (F,B) pairs, this method allows for better alpha estimation from multiple F and B samples. The same framework is extended to videos, where the requirement of temporal coherence is handled effectively. Here, the dictionary is formed by samples from multiple frames. A multi-frame graph model, as opposed to a single image as for image matting, is proposed that can be solved efficiently in closed form. Quantitative and qualitative evaluations on a benchmark dataset are provided to show that the proposed method outperforms current state-of-the-art in image and video matting.
Image fusion using sparse overcomplete feature dictionaries
Brumby, Steven P.; Bettencourt, Luis; Kenyon, Garrett T.; Chartrand, Rick; Wohlberg, Brendt
2015-10-06
Approaches for deciding what individuals in a population of visual system "neurons" are looking for using sparse overcomplete feature dictionaries are provided. A sparse overcomplete feature dictionary may be learned for an image dataset and a local sparse representation of the image dataset may be built using the learned feature dictionary. A local maximum pooling operation may be applied on the local sparse representation to produce a translation-tolerant representation of the image dataset. An object may then be classified and/or clustered within the translation-tolerant representation of the image dataset using a supervised classification algorithm and/or an unsupervised clustering algorithm.
Sparse representation for vehicle recognition
NASA Astrophysics Data System (ADS)
Monnig, Nathan D.; Sakla, Wesam
2014-06-01
The Sparse Representation for Classification (SRC) algorithm has been demonstrated to be a state-of-the-art algorithm for facial recognition applications. Wright et al. demonstrate that under certain conditions, the SRC algorithm classification performance is agnostic to choice of linear feature space and highly resilient to image corruption. In this work, we examined the SRC algorithm performance on the vehicle recognition application, using images from the semi-synthetic vehicle database generated by the Air Force Research Laboratory. To represent modern operating conditions, vehicle images were corrupted with noise, blurring, and occlusion, with representation of varying pose and lighting conditions. Experiments suggest that linear feature space selection is important, particularly in the cases involving corrupted images. Overall, the SRC algorithm consistently outperforms a standard k nearest neighbor classifier on the vehicle recognition task.
Sparse and stable Markowitz portfolios
Brodie, Joshua; Daubechies, Ingrid; De Mol, Christine; Giannone, Domenico; Loris, Ignace
2009-01-01
We consider the problem of portfolio selection within the classical Markowitz mean-variance framework, reformulated as a constrained least-squares regression problem. We propose to add to the objective function a penalty proportional to the sum of the absolute values of the portfolio weights. This penalty regularizes (stabilizes) the optimization problem, encourages sparse portfolios (i.e., portfolios with only few active positions), and allows accounting for transaction costs. Our approach recovers as special cases the no-short-positions portfolios, but does allow for short positions in limited number. We implement this methodology on two benchmark data sets constructed by Fama and French. Using only a modest amount of training data, we construct portfolios whose out-of-sample performance, as measured by Sharpe ratio, is consistently and significantly better than that of the naïve evenly weighted portfolio. PMID:19617537
Sparse representation for the ISAR image reconstruction
NASA Astrophysics Data System (ADS)
Hu, Mengqi; Montalbo, John; Li, Shuxia; Sun, Ligang; Qiao, Zhijun G.
2016-05-01
In this paper, a sparse representation of the data for an inverse synthetic aperture radar (ISAR) system is provided in two dimensions. The proposed sparse representation motivates the use a of a Convex Optimization that recovers the image with far less samples, which is required by Nyquist-Shannon sampling theorem to increases the efficiency and decrease the cost of calculation in radar imaging.
Approximate Orthogonal Sparse Embedding for Dimensionality Reduction.
Lai, Zhihui; Wong, Wai Keung; Xu, Yong; Yang, Jian; Zhang, David
2016-04-01
Locally linear embedding (LLE) is one of the most well-known manifold learning methods. As the representative linear extension of LLE, orthogonal neighborhood preserving projection (ONPP) has attracted widespread attention in the field of dimensionality reduction. In this paper, a unified sparse learning framework is proposed by introducing the sparsity or L1-norm learning, which further extends the LLE-based methods to sparse cases. Theoretical connections between the ONPP and the proposed sparse linear embedding are discovered. The optimal sparse embeddings derived from the proposed framework can be computed by iterating the modified elastic net and singular value decomposition. We also show that the proposed model can be viewed as a general model for sparse linear and nonlinear (kernel) subspace learning. Based on this general model, sparse kernel embedding is also proposed for nonlinear sparse feature extraction. Extensive experiments on five databases demonstrate that the proposed sparse learning framework performs better than the existing subspace learning algorithm, particularly in the cases of small sample sizes. PMID:25955995
Approximate inverse preconditioners for general sparse matrices
Chow, E.; Saad, Y.
1994-12-31
Preconditioned Krylov subspace methods are often very efficient in solving sparse linear matrices that arise from the discretization of elliptic partial differential equations. However, for general sparse indifinite matrices, the usual ILU preconditioners fail, often because of the fact that the resulting factors L and U give rise to unstable forward and backward sweeps. In such cases, alternative preconditioners based on approximate inverses may be attractive. We are currently developing a number of such preconditioners based on iterating on each column to get the approximate inverse. For this approach to be efficient, the iteration must be done in sparse mode, i.e., we must use sparse-matrix by sparse-vector type operatoins. We will discuss a few options and compare their performance on standard problems from the Harwell-Boeing collection.
Large-scale sparse singular value computations
NASA Technical Reports Server (NTRS)
Berry, Michael W.
1992-01-01
Four numerical methods for computing the singular value decomposition (SVD) of large sparse matrices on a multiprocessor architecture are presented. Lanczos and subspace iteration-based methods for determining several of the largest singular triplets (singular values and corresponding left and right-singular vectors) for sparse matrices arising from two practical applications: information retrieval and seismic reflection tomography are emphasized. The target architectures for implementations are the CRAY-2S/4-128 and Alliant FX/80. The sparse SVD problem is well motivated by recent information-retrieval techniques in which dominant singular values and their corresponding singular vectors of large sparse term-document matrices are desired, and by nonlinear inverse problems from seismic tomography applications which require approximate pseudo-inverses of large sparse Jacobian matrices.
Resistant multiple sparse canonical correlation.
Coleman, Jacob; Replogle, Joseph; Chandler, Gabriel; Hardin, Johanna
2016-04-01
Canonical correlation analysis (CCA) is a multivariate technique that takes two datasets and forms the most highly correlated possible pairs of linear combinations between them. Each subsequent pair of linear combinations is orthogonal to the preceding pair, meaning that new information is gleaned from each pair. By looking at the magnitude of coefficient values, we can find out which variables can be grouped together, thus better understanding multiple interactions that are otherwise difficult to compute or grasp intuitively. CCA appears to have quite powerful applications to high-throughput data, as we can use it to discover, for example, relationships between gene expression and gene copy number variation. One of the biggest problems of CCA is that the number of variables (often upwards of 10,000) makes biological interpretation of linear combinations nearly impossible. To limit variable output, we have employed a method known as sparse canonical correlation analysis (SCCA), while adding estimation which is resistant to extreme observations or other types of deviant data. In this paper, we have demonstrated the success of resistant estimation in variable selection using SCCA. Additionally, we have used SCCA to find multiple canonical pairs for extended knowledge about the datasets at hand. Again, using resistant estimators provided more accurate estimates than standard estimators in the multiple canonical correlation setting. R code is available and documented at https://github.com/hardin47/rmscca. PMID:26963062
Sparse Spectrotemporal Coding of Sounds
NASA Astrophysics Data System (ADS)
Klein, David J.; König, Peter; Körding, Konrad P.
2003-12-01
Recent studies of biological auditory processing have revealed that sophisticated spectrotemporal analyses are performed by central auditory systems of various animals. The analysis is typically well matched with the statistics of relevant natural sounds, suggesting that it produces an optimal representation of the animal's acoustic biotope. We address this topic using simulated neurons that learn an optimal representation of a speech corpus. As input, the neurons receive a spectrographic representation of sound produced by a peripheral auditory model. The output representation is deemed optimal when the responses of the neurons are maximally sparse. Following optimization, the simulated neurons are similar to real neurons in many respects. Most notably, a given neuron only analyzes the input over a localized region of time and frequency. In addition, multiple subregions either excite or inhibit the neuron, together producing selectivity to spectral and temporal modulation patterns. This suggests that the brain's solution is particularly well suited for coding natural sound; therefore, it may prove useful in the design of new computational methods for processing speech.
Sparse Bayesian infinite factor models
Bhattacharya, A.; Dunson, D. B.
2011-01-01
We focus on sparse modelling of high-dimensional covariance matrices using Bayesian latent factor models. We propose a multiplicative gamma process shrinkage prior on the factor loadings which allows introduction of infinitely many factors, with the loadings increasingly shrunk towards zero as the column index increases. We use our prior on a parameter-expanded loading matrix to avoid the order dependence typical in factor analysis models and develop an efficient Gibbs sampler that scales well as data dimensionality increases. The gain in efficiency is achieved by the joint conjugacy property of the proposed prior, which allows block updating of the loadings matrix. We propose an adaptive Gibbs sampler for automatically truncating the infinite loading matrix through selection of the number of important factors. Theoretical results are provided on the support of the prior and truncation approximation bounds. A fast algorithm is proposed to produce approximate Bayes estimates. Latent factor regression methods are developed for prediction and variable selection in applications with high-dimensional correlated predictors. Operating characteristics are assessed through simulation studies, and the approach is applied to predict survival times from gene expression data. PMID:23049129
Enhancing Scalability of Sparse Direct Methods
Li, Xiaoye S.; Demmel, James; Grigori, Laura; Gu, Ming; Xia,Jianlin; Jardin, Steve; Sovinec, Carl; Lee, Lie-Quan
2007-07-23
TOPS is providing high-performance, scalable sparse direct solvers, which have had significant impacts on the SciDAC applications, including fusion simulation (CEMM), accelerator modeling (COMPASS), as well as many other mission-critical applications in DOE and elsewhere. Our recent developments have been focusing on new techniques to overcome scalability bottleneck of direct methods, in both time and memory. These include parallelizing symbolic analysis phase and developing linear-complexity sparse factorization methods. The new techniques will make sparse direct methods more widely usable in large 3D simulations on highly-parallel petascale computers.
Sparse High Dimensional Models in Economics
Fan, Jianqing; Lv, Jinchi; Qi, Lei
2010-01-01
This paper reviews the literature on sparse high dimensional models and discusses some applications in economics and finance. Recent developments of theory, methods, and implementations in penalized least squares and penalized likelihood methods are highlighted. These variable selection methods are proved to be effective in high dimensional sparse modeling. The limits of dimensionality that regularization methods can handle, the role of penalty functions, and their statistical properties are detailed. Some recent advances in ultra-high dimensional sparse modeling are also briefly discussed. PMID:22022635
Imaging correlography with sparse collecting apertures
NASA Astrophysics Data System (ADS)
Idell, Paul S.; Fienup, J. R.
1987-01-01
This paper investigates the possibility of implementing an imaging correlography system with sparse arrays of intensity detectors. The theory underlying the image formation process for imaging correlography is reviewed, emphasizing the spatial filtering effects that sparse collecting apertures have on the reconstructed imagery. Image recovery with sparse arrays of intensity detectors through the use of computer experiments in which laser speckle measurements are digitally simulated is then demonstrated. It is shown that the quality of imagery reconstructed using this technique is visibly enhanced when appropriate filtering techniques are applied. A performance tradeoff between collecting array redundancy and the number of speckle pattern measurements is briefly discussed.
Sparse principal component analysis in cancer research
Hsu, Ying-Lin; Huang, Po-Yu; Chen, Dung-Tsa
2015-01-01
A critical challenging component in analyzing high-dimensional data in cancer research is how to reduce the dimension of data and how to extract relevant features. Sparse principal component analysis (PCA) is a powerful statistical tool that could help reduce data dimension and select important variables simultaneously. In this paper, we review several approaches for sparse PCA, including variance maximization (VM), reconstruction error minimization (REM), singular value decomposition (SVD), and probabilistic modeling (PM) approaches. A simulation study is conducted to compare PCA and the sparse PCAs. An example using a published gene signature in a lung cancer dataset is used to illustrate the potential application of sparse PCAs in cancer research. PMID:26719835
Finding Nonoverlapping Substructures of a Sparse Matrix
Pinar, Ali; Vassilevska, Virginia
2005-08-11
Many applications of scientific computing rely on computations on sparse matrices. The design of efficient implementations of sparse matrix kernels is crucial for the overall efficiency of these applications. Due to the high compute-to-memory ratio and irregular memory access patterns, the performance of sparse matrix kernels is often far away from the peak performance on a modern processor. Alternative data structures have been proposed, which split the original matrix A into A{sub d} and A{sub s}, so that A{sub d} contains all dense blocks of a specified size in the matrix, and A{sub s} contains the remaining entries. This enables the use of dense matrix kernels on the entries of A{sub d} producing better memory performance. In this work, we study the problem of finding a maximum number of nonoverlapping dense blocks in a sparse matrix, which is previously not studied in the sparse matrix community. We show that the maximum nonoverlapping dense blocks problem is NP-complete by using a reduction from the maximum independent set problem on cubic planar graphs. We also propose a 2/3-approximation algorithm that runs in linear time in the number of nonzeros in the matrix. This extended abstract focuses on our results for 2x2 dense blocks. However we show that our results can be generalized to arbitrary sized dense blocks, and many other oriented substructures, which can be exploited to improve the memory performance of sparse matrix operations.
Sparse subspace clustering: algorithm, theory, and applications.
Elhamifar, Ehsan; Vidal, René
2013-11-01
Many real-world problems deal with collections of high-dimensional data, such as images, videos, text, and web documents, DNA microarray data, and more. Often, such high-dimensional data lie close to low-dimensional structures corresponding to several classes or categories to which the data belong. In this paper, we propose and study an algorithm, called sparse subspace clustering, to cluster data points that lie in a union of low-dimensional subspaces. The key idea is that, among the infinitely many possible representations of a data point in terms of other points, a sparse representation corresponds to selecting a few points from the same subspace. This motivates solving a sparse optimization program whose solution is used in a spectral clustering framework to infer the clustering of the data into subspaces. Since solving the sparse optimization program is in general NP-hard, we consider a convex relaxation and show that, under appropriate conditions on the arrangement of the subspaces and the distribution of the data, the proposed minimization program succeeds in recovering the desired sparse representations. The proposed algorithm is efficient and can handle data points near the intersections of subspaces. Another key advantage of the proposed algorithm with respect to the state of the art is that it can deal directly with data nuisances, such as noise, sparse outlying entries, and missing entries, by incorporating the model of the data into the sparse optimization program. We demonstrate the effectiveness of the proposed algorithm through experiments on synthetic data as well as the two real-world problems of motion segmentation and face clustering. PMID:24051734
Wavelet-based reconstruction of fossil-fuel CO2 emissions from sparse measurements
NASA Astrophysics Data System (ADS)
McKenna, S. A.; Ray, J.; Yadav, V.; Van Bloemen Waanders, B.; Michalak, A. M.
2012-12-01
We present a method to estimate spatially resolved fossil-fuel CO2 (ffCO2) emissions from sparse measurements of time-varying CO2 concentrations. It is based on the wavelet-modeling of the strongly non-stationary spatial distribution of ffCO2 emissions. The dimensionality of the wavelet model is first reduced using images of nightlights, which identify regions of human habitation. Since wavelets are a multiresolution basis set, most of the reduction is accomplished by removing fine-scale wavelets, in the regions with low nightlight radiances. The (reduced) wavelet model of emissions is propagated through an atmospheric transport model (WRF) to predict CO2 concentrations at a handful of measurement sites. The estimation of the wavelet model of emissions i.e., inferring the wavelet weights, is performed by fitting to observations at the measurement sites. This is done using Staggered Orthogonal Matching Pursuit (StOMP), which first identifies (and sets to zero) the wavelet coefficients that cannot be estimated from the observations, before estimating the remaining coefficients. This model sparsification and fitting is performed simultaneously, allowing us to explore multiple wavelet-models of differing complexity. This technique is borrowed from the field of compressive sensing, and is generally used in image and video processing. We test this approach using synthetic observations generated from emissions from the Vulcan database. 35 sensor sites are chosen over the USA. FfCO2 emissions, averaged over 8-day periods, are estimated, at a 1 degree spatial resolutions. We find that only about 40% of the wavelets in emission model can be estimated from the data; however the mix of coefficients that are estimated changes with time. Total US emission can be reconstructed with about ~5% errors. The inferred emissions, if aggregated monthly, have a correlation of 0.9 with Vulcan fluxes. We find that the estimated emissions in the Northeast US are the most accurate. Sandia
Finding nonoverlapping substructures of a sparse matrix
Pinar, Ali; Vassilevska, Virginia
2004-08-09
Many applications of scientific computing rely on computations on sparse matrices, thus the design of efficient implementations of sparse matrix kernels is crucial for the overall efficiency of these applications. Due to the high compute-to-memory ratio and irregular memory access patterns, the performance of sparse matrix kernels is often far away from the peak performance on a modern processor. Alternative data structures have been proposed, which split the original matrix A into A{sub d} and A{sub s}, so that A{sub d} contains all dense blocks of a specified size in the matrix, and A{sub s} contains the remaining entries. This enables the use of dense matrix kernels on the entries of A{sub d} producing better memory performance. In this work, we study the problem of finding a maximum number of non overlapping rectangular dense blocks in a sparse matrix, which has not been studied in the sparse matrix community. We show that the maximum non overlapping dense blocks problem is NP-complete by using a reduction from the maximum independent set problem on cubic planar graphs. We also propose a 2/3-approximation algorithm for 2 times 2 blocks that runs in linear time in the number of nonzeros in the matrix. We discuss alternatives to rectangular blocks such as diagonal blocks and cross blocks and present complexity analysis and approximation algorithms.
Removing sparse noise from hyperspectral images with sparse and low-rank penalties
NASA Astrophysics Data System (ADS)
Tariyal, Snigdha; Aggarwal, Hemant Kumar; Majumdar, Angshul
2016-03-01
In diffraction grating, at times, there are defective pixels on the focal plane array; this results in horizontal lines of corrupted pixels in some channels. Since only a few such pixels exist, the corruption/noise is sparse. Studies on sparse noise removal from hyperspectral noise are parsimonious. To remove such sparse noise, a prior work exploited the interband spectral correlation along with intraband spatial redundancy to yield a sparse representation in transform domains. We improve upon the prior technique. The intraband spatial redundancy is modeled as a sparse set of transform coefficients and the interband spectral correlation is modeled as a rank deficient matrix. The resulting optimization problem is solved using the split Bregman technique. Comparative experimental results show that our proposed approach is better than the previous one.
Fast wavelet based sparse approximate inverse preconditioner
Wan, W.L.
1996-12-31
Incomplete LU factorization is a robust preconditioner for both general and PDE problems but unfortunately not easy to parallelize. Recent study of Huckle and Grote and Chow and Saad showed that sparse approximate inverse could be a potential alternative while readily parallelizable. However, for special class of matrix A that comes from elliptic PDE problems, their preconditioners are not optimal in the sense that independent of mesh size. A reason may be that no good sparse approximate inverse exists for the dense inverse matrix. Our observation is that for this kind of matrices, its inverse entries typically have piecewise smooth changes. We can take advantage of this fact and use wavelet compression techniques to construct a better sparse approximate inverse preconditioner. We shall show numerically that our approach is effective for this kind of matrices.
A unified approach to sparse signal processing
NASA Astrophysics Data System (ADS)
Marvasti, Farokh; Amini, Arash; Haddadi, Farzan; Soltanolkotabi, Mahdi; Khalaj, Babak Hossein; Aldroubi, Akram; Sanei, Saeid; Chambers, Janathon
2012-12-01
A unified view of the area of sparse signal processing is presented in tutorial form by bringing together various fields in which the property of sparsity has been successfully exploited. For each of these fields, various algorithms and techniques, which have been developed to leverage sparsity, are described succinctly. The common potential benefits of significant reduction in sampling rate and processing manipulations through sparse signal processing are revealed. The key application domains of sparse signal processing are sampling, coding, spectral estimation, array processing, component analysis, and multipath channel estimation. In terms of the sampling process and reconstruction algorithms, linkages are made with random sampling, compressed sensing, and rate of innovation. The redundancy introduced by channel coding in finite and real Galois fields is then related to over-sampling with similar reconstruction algorithms. The error locator polynomial (ELP) and iterative methods are shown to work quite effectively for both sampling and coding applications. The methods of Prony, Pisarenko, and MUltiple SIgnal Classification (MUSIC) are next shown to be targeted at analyzing signals with sparse frequency domain representations. Specifically, the relations of the approach of Prony to an annihilating filter in rate of innovation and ELP in coding are emphasized; the Pisarenko and MUSIC methods are further improvements of the Prony method under noisy environments. The iterative methods developed for sampling and coding applications are shown to be powerful tools in spectral estimation. Such narrowband spectral estimation is then related to multi-source location and direction of arrival estimation in array processing. Sparsity in unobservable source signals is also shown to facilitate source separation in sparse component analysis; the algorithms developed in this area such as linear programming and matching pursuit are also widely used in compressed sensing. Finally
Sparse regularization for force identification using dictionaries
NASA Astrophysics Data System (ADS)
Qiao, Baijie; Zhang, Xingwu; Wang, Chenxi; Zhang, Hang; Chen, Xuefeng
2016-04-01
The classical function expansion method based on minimizing l2-norm of the response residual employs various basis functions to represent the unknown force. Its difficulty lies in determining the optimum number of basis functions. Considering the sparsity of force in the time domain or in other basis space, we develop a general sparse regularization method based on minimizing l1-norm of the coefficient vector of basis functions. The number of basis functions is adaptively determined by minimizing the number of nonzero components in the coefficient vector during the sparse regularization process. First, according to the profile of the unknown force, the dictionary composed of basis functions is determined. Second, a sparsity convex optimization model for force identification is constructed. Third, given the transfer function and the operational response, Sparse reconstruction by separable approximation (SpaRSA) is developed to solve the sparse regularization problem of force identification. Finally, experiments including identification of impact and harmonic forces are conducted on a cantilever thin plate structure to illustrate the effectiveness and applicability of SpaRSA. Besides the Dirac dictionary, other three sparse dictionaries including Db6 wavelets, Sym4 wavelets and cubic B-spline functions can also accurately identify both the single and double impact forces from highly noisy responses in a sparse representation frame. The discrete cosine functions can also successfully reconstruct the harmonic forces including the sinusoidal, square and triangular forces. Conversely, the traditional Tikhonov regularization method with the L-curve criterion fails to identify both the impact and harmonic forces in these cases.
Sparse representation of complex MRI images.
Nandakumar, Hari Prasad; Ji, Jim
2008-01-01
Sparse representation of images acquired from Magnet Resonance Imaging (MRI) has several potential applications. MRI is unique in that the raw images are complex. Complex wavelet transforms (CWT) can be used to produce flexible signal representations when compared to Discrete Wavelet Transform (DWT). In this work, five different schemes using CWT or DWT are tested for sparse representation of MRI images which are in the form of complex values, separate real/imaginary, or separate magnitude/phase. The experimental results on real in-vivo MRI images show that appropriate CWT, e.g., dual-tree CWT (DTCWT), can achieve sparsity better than DWT with similar Mean Square Error. PMID:19162677
Native ultrametricity of sparse random ensembles
NASA Astrophysics Data System (ADS)
Avetisov, V.; Krapivsky, P. L.; Nechaev, S.
2016-01-01
We investigate the eigenvalue density in ensembles of large sparse Bernoulli random matrices. Analyzing in detail the spectral density of ensembles of linear subgraphs, we discuss its ultrametric nature and show that near the spectrum boundary, the tails of the spectral density exhibit a Lifshitz singularity typical for Anderson localization. We pay attention to an intriguing connection of the spectral density to the Dedekind η-function. We conjecture that ultrametricity emerges in rare-event statistics and is inherit to generic complex sparse systems.
Analog system for computing sparse codes
Rozell, Christopher John; Johnson, Don Herrick; Baraniuk, Richard Gordon; Olshausen, Bruno A.; Ortman, Robert Lowell
2010-08-24
A parallel dynamical system for computing sparse representations of data, i.e., where the data can be fully represented in terms of a small number of non-zero code elements, and for reconstructing compressively sensed images. The system is based on the principles of thresholding and local competition that solves a family of sparse approximation problems corresponding to various sparsity metrics. The system utilizes Locally Competitive Algorithms (LCAs), nodes in a population continually compete with neighboring units using (usually one-way) lateral inhibition to calculate coefficients representing an input in an over complete dictionary.
Sparse representation in speech signal processing
NASA Astrophysics Data System (ADS)
Lee, Te-Won; Jang, Gil-Jin; Kwon, Oh-Wook
2003-11-01
We review the sparse representation principle for processing speech signals. A transformation for encoding the speech signals is learned such that the resulting coefficients are as independent as possible. We use independent component analysis with an exponential prior to learn a statistical representation for speech signals. This representation leads to extremely sparse priors that can be used for encoding speech signals for a variety of purposes. We review applications of this method for speech feature extraction, automatic speech recognition and speaker identification. Furthermore, this method is also suited for tackling the difficult problem of separating two sounds given only a single microphone.
Tensor methods for large, sparse unconstrained optimization
Bouaricha, A.
1996-11-01
Tensor methods for unconstrained optimization were first introduced by Schnabel and Chow [SIAM J. Optimization, 1 (1991), pp. 293-315], who describe these methods for small to moderate size problems. This paper extends these methods to large, sparse unconstrained optimization problems. This requires an entirely new way of solving the tensor model that makes the methods suitable for solving large, sparse optimization problems efficiently. We present test results for sets of problems where the Hessian at the minimizer is nonsingular and where it is singular. These results show that tensor methods are significantly more efficient and more reliable than standard methods based on Newton`s method.
Infrared image recognition based on structure sparse and atomic sparse parallel
NASA Astrophysics Data System (ADS)
Wu, Yalu; Li, Ruilong; Xu, Yi; Wang, Liping
2015-12-01
Use the redundancy of the super complete dictionary can capture the structural features of the image effectively, can achieving the effective representation of the image. However, the commonly used atomic sparse representation without regard the structure of the dictionary and the unrelated non-zero-term in the process of the computation, though structure sparse consider the structure feature of dictionary, the majority coefficients of the blocks maybe are non-zero, it may affect the identification efficiency. For the disadvantages of these two sparse expressions, a weighted parallel atomic sparse and sparse structure is proposed, and the recognition efficiency is improved by the adaptive computation of the optimal weights. The atomic sparse expression and structure sparse expression are respectively, and the optimal weights are calculated by the adaptive method. Methods are as follows: training by using the less part of the identification sample, the recognition rate is calculated by the increase of the certain step size and t the constraint between weight. The recognition rate as the Z axis, two weight values respectively as X, Y axis, the resulting points can be connected in a straight line in the 3 dimensional coordinate system, by solving the highest recognition rate, the optimal weights can be obtained. Through simulation experiments can be known, the optimal weights based on adaptive method are better in the recognition rate, weights obtained by adaptive computation of a few samples, suitable for parallel recognition calculation, can effectively improve the recognition rate of infrared images.
Facial expression recognition with facial parts based sparse representation classifier
NASA Astrophysics Data System (ADS)
Zhi, Ruicong; Ruan, Qiuqi
2009-10-01
Facial expressions play important role in human communication. The understanding of facial expression is a basic requirement in the development of next generation human computer interaction systems. Researches show that the intrinsic facial features always hide in low dimensional facial subspaces. This paper presents facial parts based facial expression recognition system with sparse representation classifier. Sparse representation classifier exploits sparse representation to select face features and classify facial expressions. The sparse solution is obtained by solving l1 -norm minimization problem with constraint of linear combination equation. Experimental results show that sparse representation is efficient for facial expression recognition and sparse representation classifier obtain much higher recognition accuracies than other compared methods.
Second SIAM conference on sparse matrices: Abstracts. Final technical report
1996-12-31
This report contains abstracts on the following topics: invited and long presentations (IP1 & LP1); sparse matrix reordering & graph theory I; sparse matrix tools & environments I; eigenvalue computations I; iterative methods & acceleration techniques I; applications I; parallel algorithms I; sparse matrix reordering & graphy theory II; sparse matrix tool & environments II; least squares & optimization I; iterative methods & acceleration techniques II; applications II; eigenvalue computations II; least squares & optimization II; parallel algorithms II; sparse direct methods; iterative methods & acceleration techniques III; eigenvalue computations III; and sparse matrix reordering & graph theory III.
Multilevel sparse functional principal component analysis.
Di, Chongzhi; Crainiceanu, Ciprian M; Jank, Wolfgang S
2014-01-29
We consider analysis of sparsely sampled multilevel functional data, where the basic observational unit is a function and data have a natural hierarchy of basic units. An example is when functions are recorded at multiple visits for each subject. Multilevel functional principal component analysis (MFPCA; Di et al. 2009) was proposed for such data when functions are densely recorded. Here we consider the case when functions are sparsely sampled and may contain only a few observations per function. We exploit the multilevel structure of covariance operators and achieve data reduction by principal component decompositions at both between and within subject levels. We address inherent methodological differences in the sparse sampling context to: 1) estimate the covariance operators; 2) estimate the functional principal component scores; 3) predict the underlying curves. Through simulations the proposed method is able to discover dominating modes of variations and reconstruct underlying curves well even in sparse settings. Our approach is illustrated by two applications, the Sleep Heart Health Study and eBay auctions. PMID:24872597
Multilevel sparse functional principal component analysis
Di, Chongzhi; Crainiceanu, Ciprian M.; Jank, Wolfgang S.
2014-01-01
We consider analysis of sparsely sampled multilevel functional data, where the basic observational unit is a function and data have a natural hierarchy of basic units. An example is when functions are recorded at multiple visits for each subject. Multilevel functional principal component analysis (MFPCA; Di et al. 2009) was proposed for such data when functions are densely recorded. Here we consider the case when functions are sparsely sampled and may contain only a few observations per function. We exploit the multilevel structure of covariance operators and achieve data reduction by principal component decompositions at both between and within subject levels. We address inherent methodological differences in the sparse sampling context to: 1) estimate the covariance operators; 2) estimate the functional principal component scores; 3) predict the underlying curves. Through simulations the proposed method is able to discover dominating modes of variations and reconstruct underlying curves well even in sparse settings. Our approach is illustrated by two applications, the Sleep Heart Health Study and eBay auctions. PMID:24872597
Structured Sparse Method for Hyperspectral Unmixing
NASA Astrophysics Data System (ADS)
Zhu, Feiyun; Wang, Ying; Xiang, Shiming; Fan, Bin; Pan, Chunhong
2014-02-01
Hyperspectral Unmixing (HU) has received increasing attention in the past decades due to its ability of unveiling information latent in hyperspectral data. Unfortunately, most existing methods fail to take advantage of the spatial information in data. To overcome this limitation, we propose a Structured Sparse regularized Nonnegative Matrix Factorization (SS-NMF) method based on the following two aspects. First, we incorporate a graph Laplacian to encode the manifold structures embedded in the hyperspectral data space. In this way, the highly similar neighboring pixels can be grouped together. Second, the lasso penalty is employed in SS-NMF for the fact that pixels in the same manifold structure are sparsely mixed by a common set of relevant bases. These two factors act as a new structured sparse constraint. With this constraint, our method can learn a compact space, where highly similar pixels are grouped to share correlated sparse representations. Experiments on real hyperspectral data sets with different noise levels demonstrate that our method outperforms the state-of-the-art methods significantly.
A Comparative Study of Sparse Associative Memories
NASA Astrophysics Data System (ADS)
Gripon, Vincent; Heusel, Judith; Löwe, Matthias; Vermet, Franck
2016-05-01
We study various models of associative memories with sparse information, i.e. a pattern to be stored is a random string of 0s and 1s with about log N 1s, only. We compare different synaptic weights, architectures and retrieval mechanisms to shed light on the influence of the various parameters on the storage capacity.
Self-Control in Sparsely Coded Networks
NASA Astrophysics Data System (ADS)
Dominguez, D. R. C.; Bollé, D.
1998-03-01
A complete self-control mechanism is proposed in the dynamics of neural networks through the introduction of a time-dependent threshold, determined in function of both the noise and the pattern activity in the network. Especially for sparsely coded models this mechanism is shown to considerably improve the storage capacity, the basins of attraction, and the mutual information content.
Sparse matrix orderings for factorized inverse preconditioners
Benzi, M.; Tuama, M.
1998-09-01
The effect of reorderings on the performance of factorized sparse approximate inverse preconditioners is considered. It is shown that certain reorderings can be very beneficial both in the preconditioner construction phase and in terms of the rate of convergence of the preconditioned iteration.
A Comparative Study of Sparse Associative Memories
NASA Astrophysics Data System (ADS)
Gripon, Vincent; Heusel, Judith; Löwe, Matthias; Vermet, Franck
2016-07-01
We study various models of associative memories with sparse information, i.e. a pattern to be stored is a random string of 0s and 1s with about log N 1s, only. We compare different synaptic weights, architectures and retrieval mechanisms to shed light on the influence of the various parameters on the storage capacity.
Automatic anatomy recognition of sparse objects
NASA Astrophysics Data System (ADS)
Zhao, Liming; Udupa, Jayaram K.; Odhner, Dewey; Wang, Huiqian; Tong, Yubing; Torigian, Drew A.
2015-03-01
A general body-wide automatic anatomy recognition (AAR) methodology was proposed in our previous work based on hierarchical fuzzy models of multitudes of objects which was not tied to any specific organ system, body region, or image modality. That work revealed the challenges encountered in modeling, recognizing, and delineating sparse objects throughout the body (compared to their non-sparse counterparts) if the models are based on the object's exact geometric representations. The challenges stem mainly from the variation in sparse objects in their shape, topology, geographic layout, and relationship to other objects. That led to the idea of modeling sparse objects not from the precise geometric representations of their samples but by using a properly designed optimal super form. This paper presents the underlying improved methodology which includes 5 steps: (a) Collecting image data from a specific population group G and body region Β and delineating in these images the objects in Β to be modeled; (b) Building a super form, S-form, for each object O in Β; (c) Refining the S-form of O to construct an optimal (minimal) super form, S*-form, which constitutes the (fuzzy) model of O; (d) Recognizing objects in Β using the S*-form; (e) Defining confounding and background objects in each S*-form for each object and performing optimal delineation. Our evaluations based on 50 3D computed tomography (CT) image sets in the thorax on four sparse objects indicate that substantially improved performance (FPVF~2%, FNVF~10%, and success where the previous approach failed) can be achieved using the new approach.
Technology Transfer Automated Retrieval System (TEKTRAN)
This chapter is organized into several sections. The first has information on the history of vegetable fermentation research in the US, dating back to the late 1880s. A overview of commercial cucumber and sauerkraut fermentation practices follows, focusing on the US market, although there is some me...
Technology Transfer Automated Retrieval System (TEKTRAN)
The wide variety of fermented foods of the world can be classified by the materials obtained from the fermentation, such as alcohol (beer, wine), organic acid such as lactic acid and acetic acid (vegetables, dairy), carbon dioxide (bread), and amino acids or peptides from protein (fish fermentations...
The relationship between phytomass, NDVI and vegetation communities on Svalbard
NASA Astrophysics Data System (ADS)
Johansen, Bernt; Tømmervik, Hans
2014-04-01
Several studies have shown a close relationship between vegetation fertility and different vegetation indices extracted from satellite data. The vegetation fertility in Arctic is at overall scales highly related to temperature. At lower scales surface material, snow cover, hydrology and anthropogenic effects (geese, reindeer) are determinant in constituting the different vegetation communities. The extent and occurrence of different vegetation communities are expressed in vegetation maps. On Svalbard a vegetation map covering the entire archipelago has recently been developed. The map is differentiated into 18 map units showing large areas of non- and sparsely vegetated ground. The most favorable vegetation is seen as productive marshes and moss tundra communities in the lowland. Various mathematical combinations of spectral channels in satellite images have been applied as sensitive indicators of the presence and condition of green vegetation. Today the normalized difference vegetation index (NDVI) is mostly used to display this information. NDVI is an indicator of the density of chlorophyll in leaf tissue calculated from the red and near infrared bands: NDVI = (NIR - RED)/(NIR + RED). NDVI gives values between -1 and +1 where vegetated areas in general yield high positive values, while non-vegetated ground is found on the negative side.
NASA Astrophysics Data System (ADS)
Qiu, G. Y.; Feng, Y.
2011-12-01
Separately estimation of soil evaporation (E) and vegetation transpiration (T) is important for water management. In this study, a methodology to estimate the ratio of vegetation cover, soil evaporation and vegetation transpiration in a sparse steppe is developed based on a previous published model, the three-temperatures (3T) model. The input parameters of the model includes the surface temperatures of soil and vegetation (from thermal image), net radiation (estimated from surface temperature and solar radiation), and air temperature. The approach of unsupervised classification was used to separate the bare soil and vegetation pixels from the images. The areas with higher temperature could be regarded as the bare soil and E was estimated by the evaporation sub-model in the 3T model; while the areas with lower temperature could be regarded as pure vegetation and T was estimated by the transpiration sub-model in the 3T model. Afterward, the estimated E and T were converted into daily values and compared with the measured E and T by using Bowen Ratio and micro-lysimeter methods. Results show that the proposed approach is a useful way to separately estimated E and T in sparse steppe.
Solving large sparse eigenvalue problems on supercomputers
NASA Technical Reports Server (NTRS)
Philippe, Bernard; Saad, Youcef
1988-01-01
An important problem in scientific computing consists in finding a few eigenvalues and corresponding eigenvectors of a very large and sparse matrix. The most popular methods to solve these problems are based on projection techniques on appropriate subspaces. The main attraction of these methods is that they only require the use of the matrix in the form of matrix by vector multiplications. The implementations on supercomputers of two such methods for symmetric matrices, namely Lanczos' method and Davidson's method are compared. Since one of the most important operations in these two methods is the multiplication of vectors by the sparse matrix, methods of performing this operation efficiently are discussed. The advantages and the disadvantages of each method are compared and implementation aspects are discussed. Numerical experiments on a one processor CRAY 2 and CRAY X-MP are reported. Possible parallel implementations are also discussed.
Sparse decomposition learning based dynamic MRI reconstruction
NASA Astrophysics Data System (ADS)
Zhu, Peifei; Zhang, Qieshi; Kamata, Sei-ichiro
2015-02-01
Dynamic MRI is widely used for many clinical exams but slow data acquisition becomes a serious problem. The application of Compressed Sensing (CS) demonstrated great potential to increase imaging speed. However, the performance of CS is largely depending on the sparsity of image sequence in the transform domain, where there are still a lot to be improved. In this work, the sparsity is exploited by proposed Sparse Decomposition Learning (SDL) algorithm, which is a combination of low-rank plus sparsity and Blind Compressed Sensing (BCS). With this decomposition, only sparsity component is modeled as a sparse linear combination of temporal basis functions. This enables coefficients to be sparser and remain more details of dynamic components comparing learning the whole images. A reconstruction is performed on the undersampled data where joint multicoil data consistency is enforced by combing Parallel Imaging (PI). The experimental results show the proposed methods decrease about 15~20% of Mean Square Error (MSE) compared to other existing methods.
Sparse brain network using penalized linear regression
NASA Astrophysics Data System (ADS)
Lee, Hyekyoung; Lee, Dong Soo; Kang, Hyejin; Kim, Boong-Nyun; Chung, Moo K.
2011-03-01
Sparse partial correlation is a useful connectivity measure for brain networks when it is difficult to compute the exact partial correlation in the small-n large-p setting. In this paper, we formulate the problem of estimating partial correlation as a sparse linear regression with a l1-norm penalty. The method is applied to brain network consisting of parcellated regions of interest (ROIs), which are obtained from FDG-PET images of the autism spectrum disorder (ASD) children and the pediatric control (PedCon) subjects. To validate the results, we check their reproducibilities of the obtained brain networks by the leave-one-out cross validation and compare the clustered structures derived from the brain networks of ASD and PedCon.
Statistical prediction with Kanerva's sparse distributed memory
NASA Technical Reports Server (NTRS)
Rogers, David
1989-01-01
A new viewpoint of the processing performed by Kanerva's sparse distributed memory (SDM) is presented. In conditions of near- or over-capacity, where the associative-memory behavior of the model breaks down, the processing performed by the model can be interpreted as that of a statistical predictor. Mathematical results are presented which serve as the framework for a new statistical viewpoint of sparse distributed memory and for which the standard formulation of SDM is a special case. This viewpoint suggests possible enhancements to the SDM model, including a procedure for improving the predictiveness of the system based on Holland's work with genetic algorithms, and a method for improving the capacity of SDM even when used as an associative memory.
Sparse representation for color image restoration.
Mairal, Julien; Elad, Michael; Sapiro, Guillermo
2008-01-01
Sparse representations of signals have drawn considerable interest in recent years. The assumption that natural signals, such as images, admit a sparse decomposition over a redundant dictionary leads to efficient algorithms for handling such sources of data. In particular, the design of well adapted dictionaries for images has been a major challenge. The K-SVD has been recently proposed for this task and shown to perform very well for various grayscale image processing tasks. In this paper, we address the problem of learning dictionaries for color images and extend the K-SVD-based grayscale image denoising algorithm that appears in. This work puts forward ways for handling nonhomogeneous noise and missing information, paving the way to state-of-the-art results in applications such as color image denoising, demosaicing, and inpainting, as demonstrated in this paper. PMID:18229804
Color demosaicking via robust adaptive sparse representation
NASA Astrophysics Data System (ADS)
Huang, Lili; Xiao, Liang; Chen, Qinghua; Wang, Kai
2015-09-01
A single sensor camera can capture scenes by means of a color filter array. Each pixel samples only one of the three primary colors. We use a color demosaicking (CDM) technique to produce full color images and propose a robust adaptive sparse representation model for high quality CDM. The data fidelity term is characterized by l1 norm to suppress the heavy-tailed visual artifacts with an adaptively learned dictionary, while the regularization term is encouraged to seek sparsity by forcing sparse coding close to its nonlocal means to reduce coding errors. Based on the classical quadratic penalty function technique in optimization and an operator splitting method in convex analysis, we further present an effective iterative algorithm to solve the variational problem. The efficiency of the proposed method is demonstrated by experimental results with simulated and real camera data.
Inpainting with sparse linear combinations of exemplars
Wohlberg, Brendt
2008-01-01
We introduce a new exemplar-based inpainting algorithm based on representing the region to be inpainted as a sparse linear combination of blocks extracted from similar parts of the image being inpainted. This method is conceptually simple, being computed by functional minimization, and avoids the complexity of correctly ordering the filling in of missing regions of other exemplar-based methods. Initial performance comparisons on small inpainting regions indicate that this method provides similar or better performance than other recent methods.
Dictionary Learning Algorithms for Sparse Representation
Kreutz-Delgado, Kenneth; Murray, Joseph F.; Rao, Bhaskar D.; Engan, Kjersti; Lee, Te-Won; Sejnowski, Terrence J.
2010-01-01
Algorithms for data-driven learning of domain-specific overcomplete dictionaries are developed to obtain maximum likelihood and maximum a posteriori dictionary estimates based on the use of Bayesian models with concave/Schur-concave (CSC) negative log priors. Such priors are appropriate for obtaining sparse representations of environmental signals within an appropriately chosen (environmentally matched) dictionary. The elements of the dictionary can be interpreted as concepts, features, or words capable of succinct expression of events encountered in the environment (the source of the measured signals). This is a generalization of vector quantization in that one is interested in a description involving a few dictionary entries (the proverbial “25 words or less”), but not necessarily as succinct as one entry. To learn an environmentally adapted dictionary capable of concise expression of signals generated by the environment, we develop algorithms that iterate between a representative set of sparse representations found by variants of FOCUSS and an update of the dictionary using these sparse representations. Experiments were performed using synthetic data and natural images. For complete dictionaries, we demonstrate that our algorithms have improved performance over other independent component analysis (ICA) methods, measured in terms of signal-to-noise ratios of separated sources. In the overcomplete case, we show that the true underlying dictionary and sparse sources can be accurately recovered. In tests with natural images, learned overcomplete dictionaries are shown to have higher coding efficiency than complete dictionaries; that is, images encoded with an over-complete dictionary have both higher compression (fewer bits per pixel) and higher accuracy (lower mean square error). PMID:12590811
Notes on implementation of sparsely distributed memory
NASA Technical Reports Server (NTRS)
Keeler, J. D.; Denning, P. J.
1986-01-01
The Sparsely Distributed Memory (SDM) developed by Kanerva is an unconventional memory design with very interesting and desirable properties. The memory works in a manner that is closely related to modern theories of human memory. The SDM model is discussed in terms of its implementation in hardware. Two appendices discuss the unconventional approaches of the SDM: Appendix A treats a resistive circuit for fast, parallel address decoding; and Appendix B treats a systolic array for high throughput read and write operations.
Imaging black holes with sparse modeling
NASA Astrophysics Data System (ADS)
Honma, Mareki; Akiyama, Kazunori; Tazaki, Fumie; Kuramochi, Kazuki; Ikeda, Shiro; Hada, Kazuhiro; Uemura, Makoto
2016-03-01
We introduce a new imaging method for radio interferometry based on sparse- modeling. The direct observables in radio interferometry are visibilities, which are Fourier transformation of an astronomical image on the sky-plane, and incomplete sampling of visibilities in the spatial frequency domain results in an under-determined problem, which has been usually solved with 0 filling to un-sampled grids. In this paper we propose to directly solve this under-determined problem using sparse modeling without 0 filling, which realizes super resolution, i.e., resolution higher than the standard refraction limit. We show simulation results of sparse modeling for the Event Horizon Telescope (EHT) observations of super-massive black holes and demonstrate that our approach has significant merit in observations of black hole shadows expected to be realized in near future. We also present some results with the method applied to real data, and also discuss more advanced techniques for practical observations such as imaging with closure phase as well as treating the effect of interstellar scattering effect.
Modified sparse regularization for electrical impedance tomography.
Fan, Wenru; Wang, Huaxiang; Xue, Qian; Cui, Ziqiang; Sun, Benyuan; Wang, Qi
2016-03-01
Electrical impedance tomography (EIT) aims to estimate the electrical properties at the interior of an object from current-voltage measurements on its boundary. It has been widely investigated due to its advantages of low cost, non-radiation, non-invasiveness, and high speed. Image reconstruction of EIT is a nonlinear and ill-posed inverse problem. Therefore, regularization techniques like Tikhonov regularization are used to solve the inverse problem. A sparse regularization based on L1 norm exhibits superiority in preserving boundary information at sharp changes or discontinuous areas in the image. However, the limitation of sparse regularization lies in the time consumption for solving the problem. In order to further improve the calculation speed of sparse regularization, a modified method based on separable approximation algorithm is proposed by using adaptive step-size and preconditioning technique. Both simulation and experimental results show the effectiveness of the proposed method in improving the image quality and real-time performance in the presence of different noise intensities and conductivity contrasts. PMID:27036798
Aerial Scene Recognition using Efficient Sparse Representation
Cheriyadat, Anil M
2012-01-01
Advanced scene recognition systems for processing large volumes of high-resolution aerial image data are in great demand today. However, automated scene recognition remains a challenging problem. Efficient encoding and representation of spatial and structural patterns in the imagery are key in developing automated scene recognition algorithms. We describe an image representation approach that uses simple and computationally efficient sparse code computation to generate accurate features capable of producing excellent classification performance using linear SVM kernels. Our method exploits unlabeled low-level image feature measurements to learn a set of basis vectors. We project the low-level features onto the basis vectors and use simple soft threshold activation function to derive the sparse features. The proposed technique generates sparse features at a significantly lower computational cost than other methods~\\cite{Yang10, newsam11}, yet it produces comparable or better classification accuracy. We apply our technique to high-resolution aerial image datasets to quantify the aerial scene classification performance. We demonstrate that the dense feature extraction and representation methods are highly effective for automatic large-facility detection on wide area high-resolution aerial imagery.
Learning joint intensity-depth sparse representations.
Tosic, Ivana; Drewes, Sarah
2014-05-01
This paper presents a method for learning overcomplete dictionaries of atoms composed of two modalities that describe a 3D scene: 1) image intensity and 2) scene depth. We propose a novel joint basis pursuit (JBP) algorithm that finds related sparse features in two modalities using conic programming and we integrate it into a two-step dictionary learning algorithm. The JBP differs from related convex algorithms because it finds joint sparsity models with different atoms and different coefficient values for intensity and depth. This is crucial for recovering generative models where the same sparse underlying causes (3D features) give rise to different signals (intensity and depth). We give a bound for recovery error of sparse coefficients obtained by JBP, and show numerically that JBP is superior to the group lasso algorithm. When applied to the Middlebury depth-intensity database, our learning algorithm converges to a set of related features, such as pairs of depth and intensity edges or image textures and depth slants. Finally, we show that JBP outperforms state of the art methods on depth inpainting for time-of-flight and Microsoft Kinect 3D data. PMID:24723574
Automatic target recognition via sparse representations
NASA Astrophysics Data System (ADS)
Estabridis, Katia
2010-04-01
Automatic target recognition (ATR) based on the emerging technology of Compressed Sensing (CS) can considerably improve accuracy, speed and cost associated with these types of systems. An image based ATR algorithm has been built upon this new theory, which can perform target detection and recognition in a low dimensional space. Compressed dictionaries (A) are formed to include rotational information for a scale of interest. The algorithm seeks to identify y(test sample) as a linear combination of the dictionary elements : y=Ax, where A ∈ Rnxm(n<
Robust Fringe Projection Profilometry via Sparse Representation.
Budianto; Lun, Daniel P K
2016-04-01
In this paper, a robust fringe projection profilometry (FPP) algorithm using the sparse dictionary learning and sparse coding techniques is proposed. When reconstructing the 3D model of objects, traditional FPP systems often fail to perform if the captured fringe images have a complex scene, such as having multiple and occluded objects. It introduces great difficulty to the phase unwrapping process of an FPP system that can result in serious distortion in the final reconstructed 3D model. For the proposed algorithm, it encodes the period order information, which is essential to phase unwrapping, into some texture patterns and embeds them to the projected fringe patterns. When the encoded fringe image is captured, a modified morphological component analysis and a sparse classification procedure are performed to decode and identify the embedded period order information. It is then used to assist the phase unwrapping process to deal with the different artifacts in the fringe images. Experimental results show that the proposed algorithm can significantly improve the robustness of an FPP system. It performs equally well no matter the fringe images have a simple or complex scene, or are affected due to the ambient lighting of the working environment. PMID:26890867
SAR Image despeckling via sparse representation
NASA Astrophysics Data System (ADS)
Wang, Zhongmei; Yang, Xiaomei; Zheng, Liang
2014-11-01
SAR image despeckling is an active research area in image processing due to its importance in improving the quality of image for object detection and classification.In this paper, a new approach is proposed for multiplicative noise in SAR image removal based on nonlocal sparse representation by dictionary learning and collaborative filtering. First, a image is divided into many patches, and then a cluster is formed by clustering log-similar image patches using Fuzzy C-means (FCM). For each cluster, an over-complete dictionary is computed using the K-SVD method that iteratively updates the dictionary and the sparse coefficients. The patches belonging to the same cluster are then reconstructed by a sparse combination of the corresponding dictionary atoms. The reconstructed patches are finally collaboratively aggregated to build the denoised image. The experimental results show that the proposed method achieves much better results than many state-of-the-art algorithms in terms of both objective evaluation index (PSNR and ENL) and subjective visual perception.
Mean-field sparse optimal control
Fornasier, Massimo; Piccoli, Benedetto; Rossi, Francesco
2014-01-01
We introduce the rigorous limit process connecting finite dimensional sparse optimal control problems with ODE constraints, modelling parsimonious interventions on the dynamics of a moving population divided into leaders and followers, to an infinite dimensional optimal control problem with a constraint given by a system of ODE for the leaders coupled with a PDE of Vlasov-type, governing the dynamics of the probability distribution of the followers. In the classical mean-field theory, one studies the behaviour of a large number of small individuals freely interacting with each other, by simplifying the effect of all the other individuals on any given individual by a single averaged effect. In this paper, we address instead the situation where the leaders are actually influenced also by an external policy maker, and we propagate its effect for the number N of followers going to infinity. The technical derivation of the sparse mean-field optimal control is realized by the simultaneous development of the mean-field limit of the equations governing the followers dynamics together with the Γ-limit of the finite dimensional sparse optimal control problems. PMID:25288818
Sparseness- and continuity-constrained seismic imaging
NASA Astrophysics Data System (ADS)
Herrmann, Felix J.
2005-04-01
Non-linear solution strategies to the least-squares seismic inverse-scattering problem with sparseness and continuity constraints are proposed. Our approach is designed to (i) deal with substantial amounts of additive noise (SNR < 0 dB); (ii) use the sparseness and locality (both in position and angle) of directional basis functions (such as curvelets and contourlets) on the model: the reflectivity; and (iii) exploit the near invariance of these basis functions under the normal operator, i.e., the scattering-followed-by-imaging operator. Signal-to-noise ratio and the continuity along the imaged reflectors are significantly enhanced by formulating the solution of the seismic inverse problem in terms of an optimization problem. During the optimization, sparseness on the basis and continuity along the reflectors are imposed by jointly minimizing the l1- and anisotropic diffusion/total-variation norms on the coefficients and reflectivity, respectively. [Joint work with Peyman P. Moghaddam was carried out as part of the SINBAD project, with financial support secured through ITF (the Industry Technology Facilitator) from the following organizations: BG Group, BP, ExxonMobil, and SHELL. Additional funding came from the NSERC Discovery Grants 22R81254.
Efficient visual tracking via low-complexity sparse representation
NASA Astrophysics Data System (ADS)
Lu, Weizhi; Zhang, Jinglin; Kpalma, Kidiyo; Ronsin, Joseph
2015-12-01
Thanks to its good performance on object recognition, sparse representation has recently been widely studied in the area of visual object tracking. Up to now, little attention has been paid to the complexity of sparse representation, while most works are focused on the performance improvement. By reducing the computation load related to sparse representation hundreds of times, this paper proposes by far the most computationally efficient tracking approach based on sparse representation. The proposal simply consists of two stages of sparse representation, one is for object detection and the other for object validation. Experimentally, it achieves better performance than some state-of-the-art methods in both accuracy and speed.
Galaxy redshift surveys with sparse sampling
Chiang, Chi-Ting; Wullstein, Philipp; Komatsu, Eiichiro; Jee, Inh; Jeong, Donghui; Blanc, Guillermo A.; Ciardullo, Robin; Gronwall, Caryl; Hagen, Alex; Schneider, Donald P.; Drory, Niv; Fabricius, Maximilian; Landriau, Martin; Finkelstein, Steven; Jogee, Shardha; Cooper, Erin Mentuch; Tuttle, Sarah; Gebhardt, Karl; Hill, Gary J.
2013-12-01
Survey observations of the three-dimensional locations of galaxies are a powerful approach to measure the distribution of matter in the universe, which can be used to learn about the nature of dark energy, physics of inflation, neutrino masses, etc. A competitive survey, however, requires a large volume (e.g., V{sub survey} ∼ 10Gpc{sup 3}) to be covered, and thus tends to be expensive. A ''sparse sampling'' method offers a more affordable solution to this problem: within a survey footprint covering a given survey volume, V{sub survey}, we observe only a fraction of the volume. The distribution of observed regions should be chosen such that their separation is smaller than the length scale corresponding to the wavenumber of interest. Then one can recover the power spectrum of galaxies with precision expected for a survey covering a volume of V{sub survey} (rather than the volume of the sum of observed regions) with the number density of galaxies given by the total number of observed galaxies divided by V{sub survey} (rather than the number density of galaxies within an observed region). We find that regularly-spaced sampling yields an unbiased power spectrum with no window function effect, and deviations from regularly-spaced sampling, which are unavoidable in realistic surveys, introduce calculable window function effects and increase the uncertainties of the recovered power spectrum. On the other hand, we show that the two-point correlation function (pair counting) is not affected by sparse sampling. While we discuss the sparse sampling method within the context of the forthcoming Hobby-Eberly Telescope Dark Energy Experiment, the method is general and can be applied to other galaxy surveys.
Effective dimension reduction for sparse functional data
YAO, F.; LEI, E.; WU, Y.
2015-01-01
Summary We propose a method of effective dimension reduction for functional data, emphasizing the sparse design where one observes only a few noisy and irregular measurements for some or all of the subjects. The proposed method borrows strength across the entire sample and provides a way to characterize the effective dimension reduction space, via functional cumulative slicing. Our theoretical study reveals a bias-variance trade-off associated with the regularizing truncation and decaying structures of the predictor process and the effective dimension reduction space. A simulation study and an application illustrate the superior finite-sample performance of the method. PMID:26566293
Sparse dynamics for partial differential equations
Schaeffer, Hayden; Caflisch, Russel; Hauck, Cory D.; Osher, Stanley
2013-01-01
We investigate the approximate dynamics of several differential equations when the solutions are restricted to a sparse subset of a given basis. The restriction is enforced at every time step by simply applying soft thresholding to the coefficients of the basis approximation. By reducing or compressing the information needed to represent the solution at every step, only the essential dynamics are represented. In many cases, there are natural bases derived from the differential equations, which promote sparsity. We find that our method successfully reduces the dynamics of convection equations, diffusion equations, weak shocks, and vorticity equations with high-frequency source terms. PMID:23533273
Guided wavefield reconstruction from sparse measurements
NASA Astrophysics Data System (ADS)
Mesnil, Olivier; Ruzzene, Massimo
2016-02-01
Guided wave measurements are at the basis of several Non-Destructive Evaluation (NDE) techniques. Although sparse measurements of guided wave obtained using piezoelectric sensors can efficiently detect and locate defects, extensive informa-tion on the shape and subsurface location of defects can be extracted from full-field measurements acquired by Laser Doppler Vibrometers (LDV). Wavefield acquisition from LDVs is generally a slow operation due to the fact that the wave propagation to record must be repeated for each point measurement and the initial conditions must be reached between each measurement. In this research, a Sparse Wavefield Reconstruction (SWR) process using Compressed Sensing is developed. The goal of this technique is to reduce the number of point measurements needed to apply NDE techniques by at least one order of magnitude by extrapolating the knowledge of a few randomly chosen measured pixels over an over-sampled grid. To achieve this, the Lamb wave propagation equation is used to formulate a basis of shape functions in which the wavefield has a sparse representation, in order to comply with the Compressed Sensing requirements and use l1-minimization solvers. The main assumption of this reconstruction process is that every material point of the studied area is a potential source. The Compressed Sensing matrix is defined as being the contribution that would have been received at a measurement location from each possible source, using the dispersion relations of the specimen computed using a Semi-Analytical Finite Element technique. The measurements are then processed through an l1-minimizer to find a minimum corresponding to the set of active sources and their corresponding excitation functions. This minimum represents the best combination of the parameters of the model matching the sparse measurements. Wavefields are then reconstructed using the propagation equation. The set of active sources found by minimization contains all the wave
Sparse Multivariate Regression With Covariance Estimation
Rothman, Adam J.; Levina, Elizaveta; Zhu, Ji
2014-01-01
We propose a procedure for constructing a sparse estimator of a multivariate regression coefficient matrix that accounts for correlation of the response variables. This method, which we call multivariate regression with covariance estimation (MRCE), involves penalized likelihood with simultaneous estimation of the regression coefficients and the covariance structure. An efficient optimization algorithm and a fast approximation are developed for computing MRCE. Using simulation studies, we show that the proposed method outperforms relevant competitors when the responses are highly correlated. We also apply the new method to a finance example on predicting asset returns. An R-package containing this dataset and code for computing MRCE and its approximation are available online. PMID:24963268
Parallel preconditioning techniques for sparse CG solvers
Basermann, A.; Reichel, B.; Schelthoff, C.
1996-12-31
Conjugate gradient (CG) methods to solve sparse systems of linear equations play an important role in numerical methods for solving discretized partial differential equations. The large size and the condition of many technical or physical applications in this area result in the need for efficient parallelization and preconditioning techniques of the CG method. In particular for very ill-conditioned matrices, sophisticated preconditioner are necessary to obtain both acceptable convergence and accuracy of CG. Here, we investigate variants of polynomial and incomplete Cholesky preconditioners that markedly reduce the iterations of the simply diagonally scaled CG and are shown to be well suited for massively parallel machines.
Distributed memory compiler design for sparse problems
NASA Technical Reports Server (NTRS)
Wu, Janet; Saltz, Joel; Berryman, Harry; Hiranandani, Seema
1991-01-01
A compiler and runtime support mechanism is described and demonstrated. The methods presented are capable of solving a wide range of sparse and unstructured problems in scientific computing. The compiler takes as input a FORTRAN 77 program enhanced with specifications for distributing data, and the compiler outputs a message passing program that runs on a distributed memory computer. The runtime support for this compiler is a library of primitives designed to efficiently support irregular patterns of distributed array accesses and irregular distributed array partitions. A variety of Intel iPSC/860 performance results obtained through the use of this compiler are presented.
A view of Kanerva's sparse distributed memory
NASA Technical Reports Server (NTRS)
Denning, P. J.
1986-01-01
Pentti Kanerva is working on a new class of computers, which are called pattern computers. Pattern computers may close the gap between capabilities of biological organisms to recognize and act on patterns (visual, auditory, tactile, or olfactory) and capabilities of modern computers. Combinations of numeric, symbolic, and pattern computers may one day be capable of sustaining robots. The overview of the requirements for a pattern computer, a summary of Kanerva's Sparse Distributed Memory (SDM), and examples of tasks this computer can be expected to perform well are given.
Learning sparse discriminative representations for land cover classification in the Arctic
NASA Astrophysics Data System (ADS)
Moody, Daniela I.; Brumby, Steven P.; Rowland, Joel C.; Gangodagamage, Chandana
2012-10-01
Neuroscience-inspired machine vision algorithms are of current interest in the areas of detection and monitoring of climate change impacts, and general Land Use/Land Cover classification using satellite image data. We describe an approach for automatic classification of land cover in multispectral satellite imagery of the Arctic using sparse representations over learned dictionaries. We demonstrate our method using DigitalGlobe Worldview-2 8-band visible/near infrared high spatial resolution imagery of the MacKenzie River basin. We use an on-line batch Hebbian learning rule to build spectral-textural dictionaries that are adapted to this multispectral data. We learn our dictionaries from millions of overlapping image patches and then use a pursuit search to generate sparse classification features. We explore unsupervised clustering in the sparse representation space to produce land-cover category labels. This approach combines spectral and spatial textural characteristics to detect geologic, vegetative, and hydrologic features. We compare our technique to standard remote sensing algorithms. Our results suggest that neuroscience-based models are a promising approach to practical pattern recognition problems in remote sensing, even for datasets using spectral bands not found in natural visual systems.
OSKI: A Library of Automatically Tuned Sparse Matrix Kernels
Vuduc, R; Demmel, J W; Yelick, K A
2005-07-19
The Optimized Sparse Kernel Interface (OSKI) is a collection of low-level primitives that provide automatically tuned computational kernels on sparse matrices, for use by solver libraries and applications. These kernels include sparse matrix-vector multiply and sparse triangular solve, among others. The primary aim of this interface is to hide the complex decision-making process needed to tune the performance of a kernel implementation for a particular user's sparse matrix and machine, while also exposing the steps and potentially non-trivial costs of tuning at run-time. This paper provides an overview of OSKI, which is based on our research on automatically tuned sparse kernels for modern cache-based superscalar machines.
Inferring sparse networks for noisy transient processes
NASA Astrophysics Data System (ADS)
Tran, Hoang M.; Bukkapatnam, Satish T. S.
2016-02-01
Inferring causal structures of real world complex networks from measured time series signals remains an open issue. The current approaches are inadequate to discern between direct versus indirect influences (i.e., the presence or absence of a directed arc connecting two nodes) in the presence of noise, sparse interactions, as well as nonlinear and transient dynamics of real world processes. We report a sparse regression (referred to as the -min) approach with theoretical bounds on the constraints on the allowable perturbation to recover the network structure that guarantees sparsity and robustness to noise. We also introduce averaging and perturbation procedures to further enhance prediction scores (i.e., reduce inference errors), and the numerical stability of -min approach. Extensive investigations have been conducted with multiple benchmark simulated genetic regulatory network and Michaelis-Menten dynamics, as well as real world data sets from DREAM5 challenge. These investigations suggest that our approach can significantly improve, oftentimes by 5 orders of magnitude over the methods reported previously for inferring the structure of dynamic networks, such as Bayesian network, network deconvolution, silencing and modular response analysis methods based on optimizing for sparsity, transients, noise and high dimensionality issues.
Inferring sparse networks for noisy transient processes.
Tran, Hoang M; Bukkapatnam, Satish T S
2016-01-01
Inferring causal structures of real world complex networks from measured time series signals remains an open issue. The current approaches are inadequate to discern between direct versus indirect influences (i.e., the presence or absence of a directed arc connecting two nodes) in the presence of noise, sparse interactions, as well as nonlinear and transient dynamics of real world processes. We report a sparse regression (referred to as the l1-min) approach with theoretical bounds on the constraints on the allowable perturbation to recover the network structure that guarantees sparsity and robustness to noise. We also introduce averaging and perturbation procedures to further enhance prediction scores (i.e., reduce inference errors), and the numerical stability of l1-min approach. Extensive investigations have been conducted with multiple benchmark simulated genetic regulatory network and Michaelis-Menten dynamics, as well as real world data sets from DREAM5 challenge. These investigations suggest that our approach can significantly improve, oftentimes by 5 orders of magnitude over the methods reported previously for inferring the structure of dynamic networks, such as Bayesian network, network deconvolution, silencing and modular response analysis methods based on optimizing for sparsity, transients, noise and high dimensionality issues. PMID:26916813
Towards robust topology of sparsely sampled data.
Correa, Carlos D; Lindstrom, Peter
2011-12-01
Sparse, irregular sampling is becoming a necessity for reconstructing large and high-dimensional signals. However, the analysis of this type of data remains a challenge. One issue is the robust selection of neighborhoods--a crucial part of analytic tools such as topological decomposition, clustering and gradient estimation. When extracting the topology of sparsely sampled data, common neighborhood strategies such as k-nearest neighbors may lead to inaccurate results, either due to missing neighborhood connections, which introduce false extrema, or due to spurious connections, which conceal true extrema. Other neighborhoods, such as the Delaunay triangulation, are costly to compute and store even in relatively low dimensions. In this paper, we address these issues. We present two new types of neighborhood graphs: a variation on and a generalization of empty region graphs, which considerably improve the robustness of neighborhood-based analysis tools, such as topological decomposition. Our findings suggest that these neighborhood graphs lead to more accurate topological representations of low- and high- dimensional data sets at relatively low cost, both in terms of storage and computation time. We describe the implications of our work in the analysis and visualization of scalar functions, and provide general strategies for computing and applying our neighborhood graphs towards robust data analysis. PMID:22034302
Neonatal Atlas Construction Using Sparse Representation
Shi, Feng; Wang, Li; Wu, Guorong; Li, Gang; Gilmore, John H.; Lin, Weili; Shen, Dinggang
2014-01-01
Atlas construction generally includes first an image registration step to normalize all images into a common space and then an atlas building step to fuse the information from all the aligned images. Although numerous atlas construction studies have been performed to improve the accuracy of the image registration step, unweighted or simply weighted average is often used in the atlas building step. In this article, we propose a novel patch-based sparse representation method for atlas construction after all images have been registered into the common space. By taking advantage of local sparse representation, more anatomical details can be recovered in the built atlas. To make the anatomical structures spatially smooth in the atlas, the anatomical feature constraints on group structure of representations and also the overlapping of neighboring patches are imposed to ensure the anatomical consistency between neighboring patches. The proposed method has been applied to 73 neonatal MR images with poor spatial resolution and low tissue contrast, for constructing a neonatal brain atlas with sharp anatomical details. Experimental results demonstrate that the proposed method can significantly enhance the quality of the constructed atlas by discovering more anatomical details especially in the highly convoluted cortical regions. The resulting atlas demonstrates superior performance of our atlas when applied to spatially normalizing three different neonatal datasets, compared with other start-of-the-art neonatal brain atlases. PMID:24638883
Topological sparse learning of dynamic form patterns.
Guthier, T; Willert, V; Eggert, J
2015-01-01
Motion is a crucial source of information for a variety of tasks in social interactions. The process of how humans recognize complex articulated movements such as gestures or face expressions remains largely unclear. There is an ongoing discussion if and how explicit low-level motion information, such as optical flow, is involved in the recognition process. Motivated by this discussion, we introduce a computational model that classifies the spatial configuration of gradient and optical flow patterns. The patterns are learned with an unsupervised learning algorithm based on translation-invariant nonnegative sparse coding called VNMF that extracts prototypical optical flow patterns shaped, for example, as moving heads or limb parts. A key element of the proposed system is a lateral inhibition term that suppresses activations of competing patterns in the learning process, leading to a low number of dominant and topological sparse activations. We analyze the classification performance of the gradient and optical flow patterns on three real-world human action recognition and one face expression recognition data set. The results indicate that the recognition of human actions can be achieved by gradient patterns alone, but adding optical flow patterns increases the classification performance. The combined patterns outperform other biological-inspired models and are competitive with current computer vision approaches. PMID:25248088
A density functional for sparse matter
NASA Astrophysics Data System (ADS)
Langreth, D. C.; Lundqvist, B. I.; Chakarova-Käck, S. D.; Cooper, V. R.; Dion, M.; Hyldgaard, P.; Kelkkanen, A.; Kleis, J.; Kong, Lingzhu; Li, Shen; Moses, P. G.; Murray, E.; Puzder, A.; Rydberg, H.; Schröder, E.; Thonhauser, T.
2009-02-01
Sparse matter is abundant and has both strong local bonds and weak nonbonding forces, in particular nonlocal van der Waals (vdW) forces between atoms separated by empty space. It encompasses a broad spectrum of systems, like soft matter, adsorption systems and biostructures. Density-functional theory (DFT), long since proven successful for dense matter, seems now to have come to a point, where useful extensions to sparse matter are available. In particular, a functional form, vdW-DF (Dion et al 2004 Phys. Rev. Lett. 92 246401; Thonhauser et al 2007 Phys. Rev. B 76 125112), has been proposed for the nonlocal correlations between electrons and applied to various relevant molecules and materials, including to those layered systems like graphite, boron nitride and molybdenum sulfide, to dimers of benzene, polycyclic aromatic hydrocarbons (PAHs), doped benzene, cytosine and DNA base pairs, to nonbonding forces in molecules, to adsorbed molecules, like benzene, naphthalene, phenol and adenine on graphite, alumina and metals, to polymer and carbon nanotube (CNT) crystals, and hydrogen storage in graphite and metal-organic frameworks (MOFs), and to the structure of DNA and of DNA with intercalators. Comparison with results from wavefunction calculations for the smaller systems and with experimental data for the extended ones show the vdW-DF path to be promising. This could have great ramifications.
Inferring sparse networks for noisy transient processes
Tran, Hoang M.; Bukkapatnam, Satish T.S.
2016-01-01
Inferring causal structures of real world complex networks from measured time series signals remains an open issue. The current approaches are inadequate to discern between direct versus indirect influences (i.e., the presence or absence of a directed arc connecting two nodes) in the presence of noise, sparse interactions, as well as nonlinear and transient dynamics of real world processes. We report a sparse regression (referred to as the -min) approach with theoretical bounds on the constraints on the allowable perturbation to recover the network structure that guarantees sparsity and robustness to noise. We also introduce averaging and perturbation procedures to further enhance prediction scores (i.e., reduce inference errors), and the numerical stability of -min approach. Extensive investigations have been conducted with multiple benchmark simulated genetic regulatory network and Michaelis-Menten dynamics, as well as real world data sets from DREAM5 challenge. These investigations suggest that our approach can significantly improve, oftentimes by 5 orders of magnitude over the methods reported previously for inferring the structure of dynamic networks, such as Bayesian network, network deconvolution, silencing and modular response analysis methods based on optimizing for sparsity, transients, noise and high dimensionality issues. PMID:26916813
Iterative Sparse Approximation of the Gravitational Potential
NASA Astrophysics Data System (ADS)
Telschow, R.
2012-04-01
In recent applications in the approximation of gravitational potential fields, several new challenges arise. We are concerned with a huge quantity of data (e.g. in case of the Earth) or strongly irregularly distributed data points (e.g. in case of the Juno mission to Jupiter), where both of these problems bring the established approximation methods to their limits. Our novel method, which is a matching pursuit, however, iteratively chooses a best basis out of a large redundant family of trial functions to reconstruct the signal. It is independent of the data points which makes it possible to take into account a much higher amount of data and, furthermore, handle irregularly distributed data, since the algorithm is able to combine arbitrary spherical basis functions, i.e., global as well as local trial functions. This additionaly results in a solution, which is sparse in the sense that it features more basis functions where the signal has a higher local detail density. Summarizing, we get a method which reconstructs large quantities of data with a preferably low number of basis functions, combining global as well as several localizing functions to a sparse basis and a solution which is locally adapted to the data density and also to the detail density of the signal.
Feature Selection and Pedestrian Detection Based on Sparse Representation
Yao, Shihong; Wang, Tao; Shen, Weiming; Pan, Shaoming; Chong, Yanwen; Ding, Fei
2015-01-01
Pedestrian detection have been currently devoted to the extraction of effective pedestrian features, which has become one of the obstacles in pedestrian detection application according to the variety of pedestrian features and their large dimension. Based on the theoretical analysis of six frequently-used features, SIFT, SURF, Haar, HOG, LBP and LSS, and their comparison with experimental results, this paper screens out the sparse feature subsets via sparse representation to investigate whether the sparse subsets have the same description abilities and the most stable features. When any two of the six features are fused, the fusion feature is sparsely represented to obtain its important components. Sparse subsets of the fusion features can be rapidly generated by avoiding calculation of the corresponding index of dimension numbers of these feature descriptors; thus, the calculation speed of the feature dimension reduction is improved and the pedestrian detection time is reduced. Experimental results show that sparse feature subsets are capable of keeping the important components of these six feature descriptors. The sparse features of HOG and LSS possess the same description ability and consume less time compared with their full features. The ratios of the sparse feature subsets of HOG and LSS to their full sets are the highest among the six, and thus these two features can be used to best describe the characteristics of the pedestrian and the sparse feature subsets of the combination of HOG-LSS show better distinguishing ability and parsimony. PMID:26295480
Feature Selection and Pedestrian Detection Based on Sparse Representation.
Yao, Shihong; Wang, Tao; Shen, Weiming; Pan, Shaoming; Chong, Yanwen; Ding, Fei
2015-01-01
Pedestrian detection have been currently devoted to the extraction of effective pedestrian features, which has become one of the obstacles in pedestrian detection application according to the variety of pedestrian features and their large dimension. Based on the theoretical analysis of six frequently-used features, SIFT, SURF, Haar, HOG, LBP and LSS, and their comparison with experimental results, this paper screens out the sparse feature subsets via sparse representation to investigate whether the sparse subsets have the same description abilities and the most stable features. When any two of the six features are fused, the fusion feature is sparsely represented to obtain its important components. Sparse subsets of the fusion features can be rapidly generated by avoiding calculation of the corresponding index of dimension numbers of these feature descriptors; thus, the calculation speed of the feature dimension reduction is improved and the pedestrian detection time is reduced. Experimental results show that sparse feature subsets are capable of keeping the important components of these six feature descriptors. The sparse features of HOG and LSS possess the same description ability and consume less time compared with their full features. The ratios of the sparse feature subsets of HOG and LSS to their full sets are the highest among the six, and thus these two features can be used to best describe the characteristics of the pedestrian and the sparse feature subsets of the combination of HOG-LSS show better distinguishing ability and parsimony. PMID:26295480
Genetic apertures: an improved sparse aperture design framework.
Salvaggio, Philip S; Schott, John R; McKeown, Donald M
2016-04-20
The majority of optical sparse aperture imaging research in the remote sensing field has been confined to a small set of aperture layouts. While these layouts possess some desirable properties for imaging, they may not be ideal for all applications. This work introduces an optimization framework for sparse aperture layouts based on genetic algorithms as well as a small set of fitness functions for incoherent sparse aperture image quality. The optimization results demonstrate the merits of existing designs and the opportunity for creating new sparse aperture layouts. PMID:27140086
Dose-shaping using targeted sparse optimization
Sayre, George A.; Ruan, Dan
2013-07-15
Purpose: Dose volume histograms (DVHs) are common tools in radiation therapy treatment planning to characterize plan quality. As statistical metrics, DVHs provide a compact summary of the underlying plan at the cost of losing spatial information: the same or similar dose-volume histograms can arise from substantially different spatial dose maps. This is exactly the reason why physicians and physicists scrutinize dose maps even after they satisfy all DVH endpoints numerically. However, up to this point, little has been done to control spatial phenomena, such as the spatial distribution of hot spots, which has significant clinical implications. To this end, the authors propose a novel objective function that enables a more direct tradeoff between target coverage, organ-sparing, and planning target volume (PTV) homogeneity, and presents our findings from four prostate cases, a pancreas case, and a head-and-neck case to illustrate the advantages and general applicability of our method.Methods: In designing the energy minimization objective (E{sub tot}{sup sparse}), the authors utilized the following robust cost functions: (1) an asymmetric linear well function to allow differential penalties for underdose, relaxation of prescription dose, and overdose in the PTV; (2) a two-piece linear function to heavily penalize high dose and mildly penalize low and intermediate dose in organs-at risk (OARs); and (3) a total variation energy, i.e., the L{sub 1} norm applied to the first-order approximation of the dose gradient in the PTV. By minimizing a weighted sum of these robust costs, general conformity to dose prescription and dose-gradient prescription is achieved while encouraging prescription violations to follow a Laplace distribution. In contrast, conventional quadratic objectives are associated with a Gaussian distribution of violations, which is less forgiving to large violations of prescription than the Laplace distribution. As a result, the proposed objective E{sub tot
Sparse distributed memory and related models
NASA Technical Reports Server (NTRS)
Kanerva, Pentti
1992-01-01
Described here is sparse distributed memory (SDM) as a neural-net associative memory. It is characterized by two weight matrices and by a large internal dimension - the number of hidden units is much larger than the number of input or output units. The first matrix, A, is fixed and possibly random, and the second matrix, C, is modifiable. The SDM is compared and contrasted to (1) computer memory, (2) correlation-matrix memory, (3) feet-forward artificial neural network, (4) cortex of the cerebellum, (5) Marr and Albus models of the cerebellum, and (6) Albus' cerebellar model arithmetic computer (CMAC). Several variations of the basic SDM design are discussed: the selected-coordinate and hyperplane designs of Jaeckel, the pseudorandom associative neural memory of Hassoun, and SDM with real-valued input variables by Prager and Fallside. SDM research conducted mainly at the Research Institute for Advanced Computer Science (RIACS) in 1986-1991 is highlighted.
Evolutionary induction of sparse neural trees
Zhang; Ohm; Muhlenbein
1997-01-01
This paper is concerned with the automatic induction of parsimonious neural networks. In contrast to other program induction situations, network induction entails parametric learning as well as structural adaptation. We present a novel representation scheme called neural trees that allows efficient learning of both network architectures and parameters by genetic search. A hybrid evolutionary method is developed for neural tree induction that combines genetic programming and the breeder genetic algorithm under the unified framework of the minimum description length principle. The method is successfully applied to the induction of higher order neural trees while still keeping the resulting structures sparse to ensure good generalization performance. Empirical results are provided on two chaotic time series prediction problems of practical interest. PMID:10021759
ESTIMATION OF FUNCTIONALS OF SPARSE COVARIANCE MATRICES
Fan, Jianqing; Rigollet, Philippe; Wang, Weichen
2016-01-01
High-dimensional statistical tests often ignore correlations to gain simplicity and stability leading to null distributions that depend on functionals of correlation matrices such as their Frobenius norm and other ℓr norms. Motivated by the computation of critical values of such tests, we investigate the difficulty of estimation the functionals of sparse correlation matrices. Specifically, we show that simple plug-in procedures based on thresholded estimators of correlation matrices are sparsity-adaptive and minimax optimal over a large class of correlation matrices. Akin to previous results on functional estimation, the minimax rates exhibit an elbow phenomenon. Our results are further illustrated in simulated data as well as an empirical study of data arising in financial econometrics. PMID:26806986
Eigensolver for a Sparse, Large Hermitian Matrix
NASA Technical Reports Server (NTRS)
Tisdale, E. Robert; Oyafuso, Fabiano; Klimeck, Gerhard; Brown, R. Chris
2003-01-01
A parallel-processing computer program finds a few eigenvalues in a sparse Hermitian matrix that contains as many as 100 million diagonal elements. This program finds the eigenvalues faster, using less memory, than do other, comparable eigensolver programs. This program implements a Lanczos algorithm in the American National Standards Institute/ International Organization for Standardization (ANSI/ISO) C computing language, using the Message Passing Interface (MPI) standard to complement an eigensolver in PARPACK. [PARPACK (Parallel Arnoldi Package) is an extension, to parallel-processing computer architectures, of ARPACK (Arnoldi Package), which is a collection of Fortran 77 subroutines that solve large-scale eigenvalue problems.] The eigensolver runs on Beowulf clusters of computers at the Jet Propulsion Laboratory (JPL).
Multiplication method for sparse interferometric fringes.
Liu, Cong; Zhang, Xingyi; Zhou, Youhe
2016-04-01
Fringe analysis in the interferometry has been of long-standing interest to the academic community. However, the process of sparse fringe is always a headache in the measurement, especially when the specimen is very small. Through theoretical derivation and experimental measurements, our work demonstrates a new method for fringe multiplication. Theoretically, arbitrary integral-multiple fringe multiplication can be acquired by using the interferogram phase as the parameter. We simulate digital images accordingly and find that not only the skeleton lines of the multiplied fringe are very convenient to extract, but also the main frequency of which can be easily separated from the DC component. Meanwhile, the experimental results have a good agreement with the theoretic ones in a validation using the classical photoelasticity. PMID:27137055
Predicting structure in nonsymmetric sparse matrix factorizations
Gilbert, J.R.; Ng, E.
1991-12-31
Many computations on sparse matrices have a phase that predicts the nonzero structure of the output, followed by a phase that actually performs the numerical computation. We study structure prediction for computations that involve nonsymmetric row and column permutations and nonsymmetric or non-square matrices. Our tools are bipartite graphs, matchings, and alternating paths. Our main new result concerns LU factorization with partial pivoting. We show that if a square matrix A has the strong Hall property (i.e., is fully indecomposable) then an upper bound due to George and Ng on the nonzero structure of L + U is as tight as possible. To show this, we prove a crucial result about alternating paths in strong Hall graphs. The alternating-paths theorem seems to be of independent interest: it can also be used to prove related results about structure prediction for QR factorization that are due to Coleman, Edenbrandt, Gilbert, Hare, Johnson, Olesky, Pothen, and van den Driessche.
Predicting structure in nonsymmetric sparse matrix factorizations
Gilbert, J.R. ); Ng, E. )
1991-01-01
Many computations on sparse matrices have a phase that predicts the nonzero structure of the output, followed by a phase that actually performs the numerical computation. We study structure prediction for computations that involve nonsymmetric row and column permutations and nonsymmetric or non-square matrices. Our tools are bipartite graphs, matchings, and alternating paths. Our main new result concerns LU factorization with partial pivoting. We show that if a square matrix A has the strong Hall property (i.e., is fully indecomposable) then an upper bound due to George and Ng on the nonzero structure of L + U is as tight as possible. To show this, we prove a crucial result about alternating paths in strong Hall graphs. The alternating-paths theorem seems to be of independent interest: it can also be used to prove related results about structure prediction for QR factorization that are due to Coleman, Edenbrandt, Gilbert, Hare, Johnson, Olesky, Pothen, and van den Driessche.
Predicting structure in nonsymmetric sparse matrix factorizations
Gilbert, J.R. ); Ng, E.G. )
1992-10-01
Many computations on sparse matrices have a phase that predicts the nonzero structure of the output, followed by a phase that actually performs the numerical computation. We study structure prediction for computations that involve nonsymmetric row and column permutations and nonsymmetric or non-square matrices. Our tools are bipartite graphs, matchings, and alternating paths. Our main new result concerns LU factorization with partial pivoting. We show that if a square matrix A has the strong Hall property (i.e., is fully indecomposable) then an upper bound due to George and Ng on the nonzero structure of L + U is as tight as possible. To show this, we prove a crucial result about alternating paths in strong Hall graphs. The alternating-paths theorem seems to be of independent interest: it can also be used to prove related results about structure prediction for QR factorization that are due to Coleman, Edenbrandt, Gilbert, Hare, Johnson, Olesky, Pothen, and van den Driessche.
Fast generation of sparse random kernel graphs
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in timemore » at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.« less
Fast generation of sparse random kernel graphs
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in time at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.
Fast Generation of Sparse Random Kernel Graphs
2015-01-01
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in time at most 𝒪(n(logn)2). As a practical example we show how to generate samples of power-law degree distribution graphs with tunable assortativity. PMID:26356296
Functional fixedness in a technologically sparse culture.
German, Tim P; Barrett, H Clark
2005-01-01
Problem solving can be inefficient when the solution requires subjects to generate an atypical function for an object and the object's typical function has been primed. Subjects become "fixed" on the design function of the object, and problem solving suffers relative to control conditions in which the object's function is not demonstrated. In the current study, such functional fixedness was demonstrated in a sample of adolescents (mean age of 16 years) among the Shuar of Ecuadorian Amazonia, whose technologically sparse culture provides limited access to large numbers of artifacts with highly specialized functions. This result suggests that design function may universally be the core property of artifact concepts in human semantic memory. PMID:15660843
Integer sparse distributed memory: analysis and results.
Snaider, Javier; Franklin, Stan; Strain, Steve; George, E Olusegun
2013-10-01
Sparse distributed memory is an auto-associative memory system that stores high dimensional Boolean vectors. Here we present an extension of the original SDM, the Integer SDM that uses modular arithmetic integer vectors rather than binary vectors. This extension preserves many of the desirable properties of the original SDM: auto-associativity, content addressability, distributed storage, and robustness over noisy inputs. In addition, it improves the representation capabilities of the memory and is more robust over normalization. It can also be extended to support forgetting and reliable sequence storage. We performed several simulations that test the noise robustness property and capacity of the memory. Theoretical analyses of the memory's fidelity and capacity are also presented. PMID:23747569
Encoding Cortical Dynamics in Sparse Features
Khan, Sheraz; Lefèvre, Julien; Baillet, Sylvain; Michmizos, Konstantinos P.; Ganesan, Santosh; Kitzbichler, Manfred G.; Zetino, Manuel; Hämäläinen, Matti S.; Papadelis, Christos; Kenet, Tal
2014-01-01
Distributed cortical solutions of magnetoencephalography (MEG) and electroencephalography (EEG) exhibit complex spatial and temporal dynamics. The extraction of patterns of interest and dynamic features from these cortical signals has so far relied on the expertise of investigators. There is a definite need in both clinical and neuroscience research for a method that will extract critical features from high-dimensional neuroimaging data in an automatic fashion. We have previously demonstrated the use of optical flow techniques for evaluating the kinematic properties of motion field projected on non-flat manifolds like in a cortical surface. We have further extended this framework to automatically detect features in the optical flow vector field by using the modified and extended 2-Riemannian Helmholtz–Hodge decomposition (HHD). Here, we applied these mathematical models on simulation and MEG data recorded from a healthy individual during a somatosensory experiment and an epilepsy pediatric patient during sleep. We tested whether our technique can automatically extract salient dynamical features of cortical activity. Simulation results indicated that we can precisely reproduce the simulated cortical dynamics with HHD; encode them in sparse features and represent the propagation of brain activity between distinct cortical areas. Using HHD, we decoded the somatosensory N20 component into two HHD features and represented the dynamics of brain activity as a traveling source between two primary somatosensory regions. In the epilepsy patient, we displayed the propagation of the epileptic activity around the margins of a brain lesion. Our findings indicate that HHD measures computed from cortical dynamics can: (i) quantitatively access the cortical dynamics in both healthy and disease brain in terms of sparse features and dynamic brain activity propagation between distinct cortical areas, and (ii) facilitate a reproducible, automated analysis of experimental and clinical
Bayesian learning of sparse multiscale image representations.
Hughes, James Michael; Rockmore, Daniel N; Wang, Yang
2013-12-01
Multiscale representations of images have become a standard tool in image analysis. Such representations offer a number of advantages over fixed-scale methods, including the potential for improved performance in denoising, compression, and the ability to represent distinct but complementary information that exists at various scales. A variety of multiresolution transforms exist, including both orthogonal decompositions such as wavelets as well as nonorthogonal, overcomplete representations. Recently, techniques for finding adaptive, sparse representations have yielded state-of-the-art results when applied to traditional image processing problems. Attempts at developing multiscale versions of these so-called dictionary learning models have yielded modest but encouraging results. However, none of these techniques has sought to combine a rigorous statistical formulation of the multiscale dictionary learning problem and the ability to share atoms across scales. We present a model for multiscale dictionary learning that overcomes some of the drawbacks of previous approaches by first decomposing an input into a pyramid of distinct frequency bands using a recursive filtering scheme, after which we perform dictionary learning and sparse coding on the individual levels of the resulting pyramid. The associated image model allows us to use a single set of adapted dictionary atoms that is shared--and learned--across all scales in the model. The underlying statistical model of our proposed method is fully Bayesian and allows for efficient inference of parameters, including the level of additive noise for denoising applications. We apply the proposed model to several common image processing problems including non-Gaussian and nonstationary denoising of real-world color images. PMID:24002002
Encoding cortical dynamics in sparse features.
Khan, Sheraz; Lefèvre, Julien; Baillet, Sylvain; Michmizos, Konstantinos P; Ganesan, Santosh; Kitzbichler, Manfred G; Zetino, Manuel; Hämäläinen, Matti S; Papadelis, Christos; Kenet, Tal
2014-01-01
Distributed cortical solutions of magnetoencephalography (MEG) and electroencephalography (EEG) exhibit complex spatial and temporal dynamics. The extraction of patterns of interest and dynamic features from these cortical signals has so far relied on the expertise of investigators. There is a definite need in both clinical and neuroscience research for a method that will extract critical features from high-dimensional neuroimaging data in an automatic fashion. We have previously demonstrated the use of optical flow techniques for evaluating the kinematic properties of motion field projected on non-flat manifolds like in a cortical surface. We have further extended this framework to automatically detect features in the optical flow vector field by using the modified and extended 2-Riemannian Helmholtz-Hodge decomposition (HHD). Here, we applied these mathematical models on simulation and MEG data recorded from a healthy individual during a somatosensory experiment and an epilepsy pediatric patient during sleep. We tested whether our technique can automatically extract salient dynamical features of cortical activity. Simulation results indicated that we can precisely reproduce the simulated cortical dynamics with HHD; encode them in sparse features and represent the propagation of brain activity between distinct cortical areas. Using HHD, we decoded the somatosensory N20 component into two HHD features and represented the dynamics of brain activity as a traveling source between two primary somatosensory regions. In the epilepsy patient, we displayed the propagation of the epileptic activity around the margins of a brain lesion. Our findings indicate that HHD measures computed from cortical dynamics can: (i) quantitatively access the cortical dynamics in both healthy and disease brain in terms of sparse features and dynamic brain activity propagation between distinct cortical areas, and (ii) facilitate a reproducible, automated analysis of experimental and clinical
Miniature Laboratory for Detecting Sparse Biomolecules
NASA Technical Reports Server (NTRS)
Lin, Ying; Yu, Nan
2005-01-01
A miniature laboratory system has been proposed for use in the field to detect sparsely distributed biomolecules. By emphasizing concentration and sorting of specimens prior to detection, the underlying system concept would make it possible to attain high detection sensitivities without the need to develop ever more sensitive biosensors. The original purpose of the proposal is to aid the search for signs of life on a remote planet by enabling the detection of specimens as sparse as a few molecules or microbes in a large amount of soil, dust, rocks, water/ice, or other raw sample material. Some version of the system could prove useful on Earth for remote sensing of biological contamination, including agents of biological warfare. Processing in this system would begin with dissolution of the raw sample material in a sample-separation vessel. The solution in the vessel would contain floating microscopic magnetic beads coated with substances that could engage in chemical reactions with various target functional groups that are parts of target molecules. The chemical reactions would cause the targeted molecules to be captured on the surfaces of the beads. By use of a controlled magnetic field, the beads would be concentrated in a specified location in the vessel. Once the beads were thus concentrated, the rest of the solution would be discarded. This procedure would obviate the filtration steps and thereby also eliminate the filter-clogging difficulties of typical prior sample-concentration schemes. For ferrous dust/soil samples, the dissolution would be done first in a separate vessel before the solution is transferred to the microbead-containing vessel.
Zhang, Di; He, Jiazhong; Zhao, Yun; Du, Minghui
2015-03-01
In magnetic resonance (MR) imaging, image spatial resolution is determined by various instrumental limitations and physical considerations. This paper presents a new algorithm for producing a high-resolution version of a low-resolution MR image. The proposed method consists of two consecutive steps: (1) reconstructs a high-resolution MR image from a given low-resolution observation via solving a joint sparse representation and nonlocal similarity L1-norm minimization problem; and (2) applies a sparse derivative prior based post-processing to suppress blurring effects. Extensive experiments on simulated brain MR images and two real clinical MR image datasets validate that the proposed method achieves much better results than many state-of-the-art algorithms in terms of both quantitative measures and visual perception. PMID:25638262
Group-based sparse representation for image restoration.
Zhang, Jian; Zhao, Debin; Gao, Wen
2014-08-01
Traditional patch-based sparse representation modeling of natural images usually suffer from two problems. First, it has to solve a large-scale optimization problem with high computational complexity in dictionary learning. Second, each patch is considered independently in dictionary learning and sparse coding, which ignores the relationship among patches, resulting in inaccurate sparse coding coefficients. In this paper, instead of using patch as the basic unit of sparse representation, we exploit the concept of group as the basic unit of sparse representation, which is composed of nonlocal patches with similar structures, and establish a novel sparse representation modeling of natural images, called group-based sparse representation (GSR). The proposed GSR is able to sparsely represent natural images in the domain of group, which enforces the intrinsic local sparsity and nonlocal self-similarity of images simultaneously in a unified framework. In addition, an effective self-adaptive dictionary learning method for each group with low complexity is designed, rather than dictionary learning from natural images. To make GSR tractable and robust, a split Bregman-based technique is developed to solve the proposed GSR-driven ℓ0 minimization problem for image restoration efficiently. Extensive experiments on image inpainting, image deblurring and image compressive sensing recovery manifest that the proposed GSR modeling outperforms many current state-of-the-art schemes in both peak signal-to-noise ratio and visual perception. PMID:24835225
Cervigram image segmentation based on reconstructive sparse representations
NASA Astrophysics Data System (ADS)
Zhang, Shaoting; Huang, Junzhou; Wang, Wei; Huang, Xiaolei; Metaxas, Dimitris
2010-03-01
We proposed an approach based on reconstructive sparse representations to segment tissues in optical images of the uterine cervix. Because of large variations in image appearance caused by the changing of the illumination and specular reflection, the color and texture features in optical images often overlap with each other and are not linearly separable. By leveraging sparse representations the data can be transformed to higher dimensions with sparse constraints and become more separated. K-SVD algorithm is employed to find sparse representations and corresponding dictionaries. The data can be reconstructed from its sparse representations and positive and/or negative dictionaries. Classification can be achieved based on comparing the reconstructive errors. In the experiments we applied our method to automatically segment the biomarker AcetoWhite (AW) regions in an archive of 60,000 images of the uterine cervix. Compared with other general methods, our approach showed lower space and time complexity and higher sensitivity.
Sparse Coding on Symmetric Positive Definite Manifolds Using Bregman Divergences.
Harandi, Mehrtash T; Hartley, Richard; Lovell, Brian; Sanderson, Conrad
2016-06-01
This paper introduces sparse coding and dictionary learning for symmetric positive definite (SPD) matrices, which are often used in machine learning, computer vision, and related areas. Unlike traditional sparse coding schemes that work in vector spaces, in this paper, we discuss how SPD matrices can be described by sparse combination of dictionary atoms, where the atoms are also SPD matrices. We propose to seek sparse coding by embedding the space of SPD matrices into the Hilbert spaces through two types of the Bregman matrix divergences. This not only leads to an efficient way of performing sparse coding but also an online and iterative scheme for dictionary learning. We apply the proposed methods to several computer vision tasks where images are represented by region covariance matrices. Our proposed algorithms outperform state-of-the-art methods on a wide range of classification tasks, including face recognition, action recognition, material classification, and texture categorization. PMID:25643414
ERIC Educational Resources Information Center
Sadler, Peter G.
The Institute for the Study of Sparsely Populated Areas is a multidisciplinary research unit which acts to coordinate, further, and initiate studies of the economic and social conditions of sparsely populated areas. Short summaries of the eight studies completed in the session of 1977-78 indicate work in such areas as the study of political life…
Classification of vegetation types in military region
NASA Astrophysics Data System (ADS)
Gonçalves, Miguel; Silva, Jose Silvestre; Bioucas-Dias, Jose
2015-10-01
In decision-making process regarding planning and execution of military operations, the terrain is a determining factor. Aerial photographs are a source of vital information for the success of an operation in hostile region, namely when the cartographic information behind enemy lines is scarce or non-existent. The objective of present work is the development of a tool capable of processing aerial photos. The methodology implemented starts with feature extraction, followed by the application of an automatic selector of features. The next step, using the k-fold cross validation technique, estimates the input parameters for the following classifiers: Sparse Multinomial Logist Regression (SMLR), K Nearest Neighbor (KNN), Linear Classifier using Principal Component Expansion on the Joint Data (PCLDC) and Multi-Class Support Vector Machine (MSVM). These classifiers were used in two different studies with distinct objectives: discrimination of vegetation's density and identification of vegetation's main components. It was found that the best classifier on the first approach is the Sparse Logistic Multinomial Regression (SMLR). On the second approach, the implemented methodology applied to high resolution images showed that the better performance was achieved by KNN classifier and PCLDC. Comparing the two approaches there is a multiscale issue, in which for different resolutions, the best solution to the problem requires different classifiers and the extraction of different features.
Ward, Andy L.
2007-11-26
Fluor Hanford (FH) is designing and assessing the performance of engineered barriers for final closure of 200-UW-1 waste sites. Engineered barriers must minimize the intrusion and water, plants and animals into the underlying waste to provide protection for human health and the environment. The Pacific Northwest National Laboratory (PNNL) developed Subsurface Transport Over Multiple Phases (STOMP) simulator is being used to optimize the performance of candidate barriers. Simulating barrier performance involves computation of mass and energy transfer within a soil-atmosphere-vegetation continuum and requires a variety of input parameters, some of which are more readily available than others. Required input includes parameter values for the geotechnical, physical, hydraulic, and thermal properties of the materials comprising the barrier and the structural fill on which it will be constructed as well as parameters to allow simulation of plant effects. This report provides a data package of the required parameters as well as the technical basis, rationale and methodology used to obtain the parameter values.
Anisotropic interpolation of sparse generalized image samples.
Bourquard, Aurélien; Unser, Michael
2013-02-01
Practical image-acquisition systems are often modeled as a continuous-domain prefilter followed by an ideal sampler, where generalized samples are obtained after convolution with the impulse response of the device. In this paper, our goal is to interpolate images from a given subset of such samples. We express our solution in the continuous domain, considering consistent resampling as a data-fidelity constraint. To make the problem well posed and ensure edge-preserving solutions, we develop an efficient anisotropic regularization approach that is based on an improved version of the edge-enhancing anisotropic diffusion equation. Following variational principles, our reconstruction algorithm minimizes successive quadratic cost functionals. To ensure fast convergence, we solve the corresponding sequence of linear problems by using multigrid iterations that are specifically tailored to their sparse structure. We conduct illustrative experiments and discuss the potential of our approach both in terms of algorithmic design and reconstruction quality. In particular, we present results that use as little as 2% of the image samples. PMID:22968212
Approximation and compression with sparse orthonormal transforms.
Sezer, Osman Gokhan; Guleryuz, Onur G; Altunbasak, Yucel
2015-08-01
We propose a new transform design method that targets the generation of compression-optimized transforms for next-generation multimedia applications. The fundamental idea behind transform compression is to exploit regularity within signals such that redundancy is minimized subject to a fidelity cost. Multimedia signals, in particular images and video, are well known to contain a diverse set of localized structures, leading to many different types of regularity and to nonstationary signal statistics. The proposed method designs sparse orthonormal transforms (SOTs) that automatically exploit regularity over different signal structures and provides an adaptation method that determines the best representation over localized regions. Unlike earlier work that is motivated by linear approximation constructs and model-based designs that are limited to specific types of signal regularity, our work uses general nonlinear approximation ideas and a data-driven setup to significantly broaden its reach. We show that our SOT designs provide a safe and principled extension of the Karhunen-Loeve transform (KLT) by reducing to the KLT on Gaussian processes and by automatically exploiting non-Gaussian statistics to significantly improve over the KLT on more general processes. We provide an algebraic optimization framework that generates optimized designs for any desired transform structure (multiresolution, block, lapped, and so on) with significantly better n -term approximation performance. For each structure, we propose a new prototype codec and test over a database of images. Simulation results show consistent increase in compression and approximation performance compared with conventional methods. PMID:25823033
Transformer fault diagnosis using continuous sparse autoencoder.
Wang, Lukun; Zhao, Xiaoying; Pei, Jiangnan; Tang, Gongyou
2016-01-01
This paper proposes a novel continuous sparse autoencoder (CSAE) which can be used in unsupervised feature learning. The CSAE adds Gaussian stochastic unit into activation function to extract features of nonlinear data. In this paper, CSAE is applied to solve the problem of transformer fault recognition. Firstly, based on dissolved gas analysis method, IEC three ratios are calculated by the concentrations of dissolved gases. Then IEC three ratios data is normalized to reduce data singularity and improve training speed. Secondly, deep belief network is established by two layers of CSAE and one layer of back propagation (BP) network. Thirdly, CSAE is adopted to unsupervised training and getting features. Then BP network is used for supervised training and getting transformer fault. Finally, the experimental data from IEC TC 10 dataset aims to illustrate the effectiveness of the presented approach. Comparative experiments clearly show that CSAE can extract features from the original data, and achieve a superior correct differentiation rate on transformer fault diagnosis. PMID:27119052
Sparse distributed memory: Principles and operation
NASA Technical Reports Server (NTRS)
Flynn, M. J.; Kanerva, P.; Bhadkamkar, N.
1989-01-01
Sparse distributed memory is a generalized random access memory (RAM) for long (1000 bit) binary words. Such words can be written into and read from the memory, and they can also be used to address the memory. The main attribute of the memory is sensitivity to similarity, meaning that a word can be read back not only by giving the original write address but also by giving one close to it as measured by the Hamming distance between addresses. Large memories of this kind are expected to have wide use in speech recognition and scene analysis, in signal detection and verification, and in adaptive control of automated equipment, in general, in dealing with real world information in real time. The memory can be realized as a simple, massively parallel computer. Digital technology has reached a point where building large memories is becoming practical. Major design issues were resolved which were faced in building the memories. The design is described of a prototype memory with 256 bit addresses and from 8 to 128 K locations for 256 bit words. A key aspect of the design is extensive use of dynamic RAM and other standard components.
Sparse distributed memory prototype: Principles of operation
NASA Technical Reports Server (NTRS)
Flynn, Michael J.; Kanerva, Pentti; Ahanin, Bahram; Bhadkamkar, Neal; Flaherty, Paul; Hickey, Philip
1988-01-01
Sparse distributed memory is a generalized random access memory (RAM) for long binary words. Such words can be written into and read from the memory, and they can be used to address the memory. The main attribute of the memory is sensitivity to similarity, meaning that a word can be read back not only by giving the original right address but also by giving one close to it as measured by the Hamming distance between addresses. Large memories of this kind are expected to have wide use in speech and scene analysis, in signal detection and verification, and in adaptive control of automated equipment. The memory can be realized as a simple, massively parallel computer. Digital technology has reached a point where building large memories is becoming practical. The research is aimed at resolving major design issues that have to be faced in building the memories. The design of a prototype memory with 256-bit addresses and from 8K to 128K locations for 256-bit words is described. A key aspect of the design is extensive use of dynamic RAM and other standard components.
Partially sparse imaging of stationary indoor scenes
NASA Astrophysics Data System (ADS)
Ahmad, Fauzia; Amin, Moeness G.; Dogaru, Traian
2014-12-01
In this paper, we exploit the notion of partial sparsity for scene reconstruction associated with through-the-wall radar imaging of stationary targets under reduced data volume. Partial sparsity implies that the scene being imaged consists of a sparse part and a dense part, with the support of the latter assumed to be known. For the problem at hand, sparsity is represented by a few stationary indoor targets, whereas the high scene density is defined by exterior and interior walls. Prior knowledge of wall positions and extent may be available either through building blueprints or from prior surveillance operations. The contributions of the exterior and interior walls are removed from the data through the use of projection matrices, which are determined from wall- and corner-specific dictionaries. The projected data, with enhanced sparsity, is then processed using l 1 norm reconstruction techniques. Numerical electromagnetic data is used to demonstrate the effectiveness of the proposed approach for imaging stationary indoor scenes using a reduced set of measurements.
Sparse coding for layered neural networks
NASA Astrophysics Data System (ADS)
Katayama, Katsuki; Sakata, Yasuo; Horiguchi, Tsuyoshi
2002-07-01
We investigate storage capacity of two types of fully connected layered neural networks with sparse coding when binary patterns are embedded into the networks by a Hebbian learning rule. One of them is a layered network, in which a transfer function of even layers is different from that of odd layers. The other is a layered network with intra-layer connections, in which the transfer function of inter-layer is different from that of intra-layer, and inter-layered neurons and intra-layered neurons are updated alternately. We derive recursion relations for order parameters by means of the signal-to-noise ratio method, and then apply the self-control threshold method proposed by Dominguez and Bollé to both layered networks with monotonic transfer functions. We find that a critical value αC of storage capacity is about 0.11|a ln a| -1 ( a≪1) for both layered networks, where a is a neuronal activity. It turns out that the basin of attraction is larger for both layered networks when the self-control threshold method is applied.
Sparse alignment for robust tensor learning.
Lai, Zhihui; Wong, Wai Keung; Xu, Yong; Zhao, Cairong; Sun, Mingming
2014-10-01
Multilinear/tensor extensions of manifold learning based algorithms have been widely used in computer vision and pattern recognition. This paper first provides a systematic analysis of the multilinear extensions for the most popular methods by using alignment techniques, thereby obtaining a general tensor alignment framework. From this framework, it is easy to show that the manifold learning based tensor learning methods are intrinsically different from the alignment techniques. Based on the alignment framework, a robust tensor learning method called sparse tensor alignment (STA) is then proposed for unsupervised tensor feature extraction. Different from the existing tensor learning methods, L1- and L2-norms are introduced to enhance the robustness in the alignment step of the STA. The advantage of the proposed technique is that the difficulty in selecting the size of the local neighborhood can be avoided in the manifold learning based tensor feature extraction algorithms. Although STA is an unsupervised learning method, the sparsity encodes the discriminative information in the alignment step and provides the robustness of STA. Extensive experiments on the well-known image databases as well as action and hand gesture databases by encoding object images as tensors demonstrate that the proposed STA algorithm gives the most competitive performance when compared with the tensor-based unsupervised learning methods. PMID:25291733
Optimal parallel solution of sparse triangular systems
NASA Technical Reports Server (NTRS)
Alvarado, Fernando L.; Schreiber, Robert
1990-01-01
A method for the parallel solution of triangular sets of equations is described that is appropriate when there are many right-handed sides. By preprocessing, the method can reduce the number of parallel steps required to solve Lx = b compared to parallel forward or backsolve. Applications are to iterative solvers with triangular preconditioners, to structural analysis, or to power systems applications, where there may be many right-handed sides (not all available a priori). The inverse of L is represented as a product of sparse triangular factors. The problem is to find a factored representation of this inverse of L with the smallest number of factors (or partitions), subject to the requirement that no new nonzero elements be created in the formation of these inverse factors. A method from an earlier reference is shown to solve this problem. This method is improved upon by constructing a permutation of the rows and columns of L that preserves triangularity and allow for the best possible such partition. A number of practical examples and algorithmic details are presented. The parallelism attainable is illustrated by means of elimination trees and clique trees.
Index statistical properties of sparse random graphs
NASA Astrophysics Data System (ADS)
Metz, F. L.; Stariolo, Daniel A.
2015-10-01
Using the replica method, we develop an analytical approach to compute the characteristic function for the probability PN(K ,λ ) that a large N ×N adjacency matrix of sparse random graphs has K eigenvalues below a threshold λ . The method allows to determine, in principle, all moments of PN(K ,λ ) , from which the typical sample-to-sample fluctuations can be fully characterized. For random graph models with localized eigenvectors, we show that the index variance scales linearly with N ≫1 for |λ |>0 , with a model-dependent prefactor that can be exactly calculated. Explicit results are discussed for Erdös-Rényi and regular random graphs, both exhibiting a prefactor with a nonmonotonic behavior as a function of λ . These results contrast with rotationally invariant random matrices, where the index variance scales only as lnN , with an universal prefactor that is independent of λ . Numerical diagonalization results confirm the exactness of our approach and, in addition, strongly support the Gaussian nature of the index fluctuations.
Partitioning sparse matrices with eigenvectors of graphs
NASA Technical Reports Server (NTRS)
Pothen, Alex; Simon, Horst D.; Liou, Kang-Pu
1990-01-01
The problem of computing a small vertex separator in a graph arises in the context of computing a good ordering for the parallel factorization of sparse, symmetric matrices. An algebraic approach for computing vertex separators is considered in this paper. It is shown that lower bounds on separator sizes can be obtained in terms of the eigenvalues of the Laplacian matrix associated with a graph. The Laplacian eigenvectors of grid graphs can be computed from Kronecker products involving the eigenvectors of path graphs, and these eigenvectors can be used to compute good separators in grid graphs. A heuristic algorithm is designed to compute a vertex separator in a general graph by first computing an edge separator in the graph from an eigenvector of the Laplacian matrix, and then using a maximum matching in a subgraph to compute the vertex separator. Results on the quality of the separators computed by the spectral algorithm are presented, and these are compared with separators obtained from other algorithms for computing separators. Finally, the time required to compute the Laplacian eigenvector is reported, and the accuracy with which the eigenvector must be computed to obtain good separators is considered. The spectral algorithm has the advantage that it can be implemented on a medium-size multiprocessor in a straightforward manner.
Optimized design and analysis of sparse-sampling FMRI experiments.
Perrachione, Tyler K; Ghosh, Satrajit S
2013-01-01
Sparse-sampling is an important methodological advance in functional magnetic resonance imaging (fMRI), in which silent delays are introduced between MR volume acquisitions, allowing for the presentation of auditory stimuli without contamination by acoustic scanner noise and for overt vocal responses without motion-induced artifacts in the functional time series. As such, the sparse-sampling technique has become a mainstay of principled fMRI research into the cognitive and systems neuroscience of speech, language, hearing, and music. Despite being in use for over a decade, there has been little systematic investigation of the acquisition parameters, experimental design considerations, and statistical analysis approaches that bear on the results and interpretation of sparse-sampling fMRI experiments. In this report, we examined how design and analysis choices related to the duration of repetition time (TR) delay (an acquisition parameter), stimulation rate (an experimental design parameter), and model basis function (an analysis parameter) act independently and interactively to affect the neural activation profiles observed in fMRI. First, we conducted a series of computational simulations to explore the parameter space of sparse design and analysis with respect to these variables; second, we validated the results of these simulations in a series of sparse-sampling fMRI experiments. Overall, these experiments suggest the employment of three methodological approaches that can, in many situations, substantially improve the detection of neurophysiological response in sparse fMRI: (1) Sparse analyses should utilize a physiologically informed model that incorporates hemodynamic response convolution to reduce model error. (2) The design of sparse fMRI experiments should maintain a high rate of stimulus presentation to maximize effect size. (3) TR delays of short to intermediate length can be used between acquisitions of sparse-sampled functional image volumes to increase
Electromagnetic Formation Flight (EMFF) for Sparse Aperture Arrays
NASA Technical Reports Server (NTRS)
Kwon, Daniel W.; Miller, David W.; Sedwick, Raymond J.
2004-01-01
Traditional methods of actuating spacecraft in sparse aperture arrays use propellant as a reaction mass. For formation flying systems, propellant becomes a critical consumable which can be quickly exhausted while maintaining relative orientation. Additional problems posed by propellant include optical contamination, plume impingement, thermal emission, and vibration excitation. For these missions where control of relative degrees of freedom is important, we consider using a system of electromagnets, in concert with reaction wheels, to replace the consumables. Electromagnetic Formation Flight sparse apertures, powered by solar energy, are designed differently from traditional propulsion systems, which are based on V. This paper investigates the design of sparse apertures both inside and outside the Earth's gravity field.
SPARSKIT: A basic tool kit for sparse matrix computations
NASA Technical Reports Server (NTRS)
Saad, Youcef
1990-01-01
Presented here are the main features of a tool package for manipulating and working with sparse matrices. One of the goals of the package is to provide basic tools to facilitate the exchange of software and data between researchers in sparse matrix computations. The starting point is the Harwell/Boeing collection of matrices for which the authors provide a number of tools. Among other things, the package provides programs for converting data structures, printing simple statistics on a matrix, plotting a matrix profile, and performing linear algebra operations with sparse matrices.
Unified inference for sparse and dense longitudinal models.
Kim, Seonjin; Zhao, Zhibiao
2013-03-01
In longitudinal data analysis, statistical inference for sparse data and dense data could be substantially different. For kernel smoothing estimate of the mean function, the convergence rates and limiting variance functions are different under the two scenarios. The latter phenomenon poses challenges for statistical inference as a subjective choice between the sparse and dense cases may lead to wrong conclusions. We develop self-normalization based methods that can adapt to the sparse and dense cases in a unified framework. Simulations show that the proposed methods outperform some existing methods. PMID:24966413
Method and apparatus for distinguishing actual sparse events from sparse event false alarms
Spalding, Richard E.; Grotbeck, Carter L.
2000-01-01
Remote sensing method and apparatus wherein sparse optical events are distinguished from false events. "Ghost" images of actual optical phenomena are generated using an optical beam splitter and optics configured to direct split beams to a single sensor or segmented sensor. True optical signals are distinguished from false signals or noise based on whether the ghost image is presence or absent. The invention obviates the need for dual sensor systems to effect a false target detection capability, thus significantly reducing system complexity and cost.
Arizona Vegetation Resource Inventory (AVRI) accuracy assessment
Szajgin, John; Pettinger, L.R.; Linden, D.S.; Ohlen, D.O.
1982-01-01
A quantitative accuracy assessment was performed for the vegetation classification map produced as part of the Arizona Vegetation Resource Inventory (AVRI) project. This project was a cooperative effort between the Bureau of Land Management (BLM) and the Earth Resources Observation Systems (EROS) Data Center. The objective of the accuracy assessment was to estimate (with a precision of ?10 percent at the 90 percent confidence level) the comission error in each of the eight level II hierarchical vegetation cover types. A stratified two-phase (double) cluster sample was used. Phase I consisted of 160 photointerpreted plots representing clusters of Landsat pixels, and phase II consisted of ground data collection at 80 of the phase I cluster sites. Ground data were used to refine the phase I error estimates by means of a linear regression model. The classified image was stratified by assigning each 15-pixel cluster to the stratum corresponding to the dominant cover type within each cluster. This method is known as stratified plurality sampling. Overall error was estimated to be 36 percent with a standard error of 2 percent. Estimated error for individual vegetation classes ranged from a low of 10 percent ?6 percent for evergreen woodland to 81 percent ?7 percent for cropland and pasture. Total cost of the accuracy assessment was $106,950 for the one-million-hectare study area. The combination of the stratified plurality sampling (SPS) method of sample allocation with double sampling provided the desired estimates within the required precision levels. The overall accuracy results confirmed that highly accurate digital classification of vegetation is difficult to perform in semiarid environments, due largely to the sparse vegetation cover. Nevertheless, these techniques show promise for providing more accurate information than is presently available for many BLM-administered lands.
Sparse source configurations for asteroid tomography
NASA Astrophysics Data System (ADS)
Pursiainen, S.; Kaasalainen, M.
2014-04-01
The objective of our recent research has been to develop non-invasive imaging techniques for future planetary research and mining activities involving a challenging in situ environment and tight payload limits [1]. This presentation will deal in particular with an approach in which the internal relative permittivity ∈r or the refractive index n = √ ∈r of an asteroid is to be recovered based on radio signal transmitted by a sparse set [2] of fixed or movable landers. To address important aspects of mission planning, we have analyzed different signal source configurations to find the minimal number of source positions needed for robust localization of anomalies, such as internal voids. Characteristic to this inverse problem are the large relative changes in signal speed caused by the high permittivity of typical asteroid minerals (e.g. basalt), leading to strong refractions and reflections of the signal. Finding an appropriate problemspecific signaling arrangement is an important premission goal for successful in situ measurements. This presentation will include inversion results obtained with laboratory-recorded travel time data y of the form in which n δ denotes a perturbation of a refractive index n = n δ + nbg; gi estimates the total noise due to different error sources; (ybg)i = ∫Ci nbg ds is an entry of noiseless background data ybg; and Ci is a signal path. Also simulated time-evolution data will be covered with respect to potential u satisfying the wave equation ∈rδ2/δt2+ ōδu/δt-∆u = f, where ō is a (latent) conductivity distribution and f is a source term. Special interest will be paid to inversion robustness regarding changes of the prior model and source positioning. Among other things, our analysis suggests that strongly refractive anomalies can be detected with three or four sources independently of their positioning.
Detecting novel genes with sparse arrays
Haiminen, Niina; Smit, Bart; Rautio, Jari; Vitikainen, Marika; Wiebe, Marilyn; Martinez, Diego; Chee, Christine; Kunkel, Joe; Sanchez, Charles; Nelson, Mary Anne; Pakula, Tiina; Saloheimo, Markku; Penttilä, Merja; Kivioja, Teemu
2014-01-01
Species-specific genes play an important role in defining the phenotype of an organism. However, current gene prediction methods can only efficiently find genes that share features such as sequence similarity or general sequence characteristics with previously known genes. Novel sequencing methods and tiling arrays can be used to find genes without prior information and they have demonstrated that novel genes can still be found from extensively studied model organisms. Unfortunately, these methods are expensive and thus are not easily applicable, e.g., to finding genes that are expressed only in very specific conditions. We demonstrate a method for finding novel genes with sparse arrays, applying it on the 33.9 Mb genome of the filamentous fungus Trichoderma reesei. Our computational method does not require normalisations between arrays and it takes into account the multiple-testing problem typical for analysis of microarray data. In contrast to tiling arrays, that use overlapping probes, only one 25mer microarray oligonucleotide probe was used for every 100 b. Thus, only relatively little space on a microarray slide was required to cover the intergenic regions of a genome. The analysis was done as a by-product of a conventional microarray experiment with no additional costs. We found at least 23 good candidates for novel transcripts that could code for proteins and all of which were expressed at high levels. Candidate genes were found to neighbour ire1 and cre1 and many other regulatory genes. Our simple, low-cost method can easily be applied to finding novel species-specific genes without prior knowledge of their sequence properties. PMID:20691772
Finding One Community in a Sparse Graph
NASA Astrophysics Data System (ADS)
Montanari, Andrea
2015-10-01
We consider a random sparse graph with bounded average degree, in which a subset of vertices has higher connectivity than the background. In particular, the average degree inside this subset of vertices is larger than outside (but still bounded). Given a realization of such graph, we aim at identifying the hidden subset of vertices. This can be regarded as a model for the problem of finding a tightly knitted community in a social network, or a cluster in a relational dataset. In this paper we present two sets of contributions: ( i) We use the cavity method from spin glass theory to derive an exact phase diagram for the reconstruction problem. In particular, as the difference in edge probability increases, the problem undergoes two phase transitions, a static phase transition and a dynamic one. ( ii) We establish rigorous bounds on the dynamic phase transition and prove that, above a certain threshold, a local algorithm (belief propagation) correctly identify most of the hidden set. Below the same threshold no local algorithm can achieve this goal. However, in this regime the subset can be identified by exhaustive search. For small hidden sets and large average degree, the phase transition for local algorithms takes an intriguingly simple form. Local algorithms succeed with high probability for deg _in - deg _out > √{deg _out/e} and fail for deg _in - deg _out < √{deg _out/e} (with deg _in, deg _out the average degrees inside and outside the community). We argue that spectral algorithms are also ineffective in the latter regime. It is an open problem whether any polynomial time algorithms might succeed for deg _in - deg _out < √{deg _out/e}.
Image inpainting based on sparse representations with a perceptual metric
NASA Astrophysics Data System (ADS)
Ogawa, Takahiro; Haseyama, Miki
2013-12-01
This paper presents an image inpainting method based on sparse representations optimized with respect to a perceptual metric. In the proposed method, the structural similarity (SSIM) index is utilized as a criterion to optimize the representation performance of image data. Specifically, the proposed method enables the formulation of two important procedures in the sparse representation problem, 'estimation of sparse representation coefficients' and 'update of the dictionary', based on the SSIM index. Then, using the generated dictionary, approximation of target patches including missing areas via the SSIM-based sparse representation becomes feasible. Consequently, image inpainting for which procedures are totally derived from the SSIM index is realized. Experimental results show that the proposed method enables successful inpainting of missing areas.
Multiple kernel learning for sparse representation-based classification.
Shrivastava, Ashish; Patel, Vishal M; Chellappa, Rama
2014-07-01
In this paper, we propose a multiple kernel learning (MKL) algorithm that is based on the sparse representation-based classification (SRC) method. Taking advantage of the nonlinear kernel SRC in efficiently representing the nonlinearities in the high-dimensional feature space, we propose an MKL method based on the kernel alignment criteria. Our method uses a two step training method to learn the kernel weights and sparse codes. At each iteration, the sparse codes are updated first while fixing the kernel mixing coefficients, and then the kernel mixing coefficients are updated while fixing the sparse codes. These two steps are repeated until a stopping criteria is met. The effectiveness of the proposed method is demonstrated using several publicly available image classification databases and it is shown that this method can perform significantly better than many competitive image classification algorithms. PMID:24835226
Ensemble polarimetric SAR image classification based on contextual sparse representation
NASA Astrophysics Data System (ADS)
Zhang, Lamei; Wang, Xiao; Zou, Bin; Qiao, Zhijun
2016-05-01
Polarimetric SAR image interpretation has become one of the most interesting topics, in which the construction of the reasonable and effective technique of image classification is of key importance. Sparse representation represents the data using the most succinct sparse atoms of the over-complete dictionary and the advantages of sparse representation also have been confirmed in the field of PolSAR classification. However, it is not perfect, like the ordinary classifier, at different aspects. So ensemble learning is introduced to improve the issue, which makes a plurality of different learners training and obtained the integrated results by combining the individual learner to get more accurate and ideal learning results. Therefore, this paper presents a polarimetric SAR image classification method based on the ensemble learning of sparse representation to achieve the optimal classification.
Marginality and needs of dietary valine for broilers fed certain all-vegetable diets
Technology Transfer Automated Retrieval System (TEKTRAN)
Valine is likely the fourth limiting amino acid in most diets based of corn and soybean meal. However, its exact needs are not well known, and information regarding it is sparse. A series of studies were conducted to validate valine’s limitation in all-vegetable diets fed to broilers, and subseque...
Retrieval of Understory NDVI in Sparse Boreal Forests By MODIS Brdf Data
NASA Astrophysics Data System (ADS)
Yang, W.; Kobayashi, H.; Suzuki, R.; Nasahara, K. N.
2014-12-01
Global products of leaf area index (LAI) usually show large uncertainties in sparsely vegetated areas. The reason is that the understory contribution is not negligible in reflectance modeling for the case of low to intermediate canopy cover. Therefore many efforts have been carried out on inclusion of understory properties in the LAI estimation algorithms. Compared with conventional data bank method, estimation of forest understory property from satellite data is superior in the studies at global or continental scale during a long periods. However, the existing remote sensing method based on multi-angular observations is very complicated to implement. Alternatively, a simple method to retrieve understory NDVI (NDVIu) for sparse boreal forests was proposed in this study. The method is based on the property that the bi-directional variation of NDVIu is much smaller than that of the canopy-level NDVI. To retrieve NDVIu for a certain pixel, linear extrapolation was applied using the pixels within a 5×5 target-pixel-centered window. The NDVI values were reconstructed from the MODIS BRDF data corresponding to eight different solar-view angles. NDVIu was estimated as the average of the NDVI values corresponding to the position where the stand NDVI has the smallest angular variation. Validation by noise-free simulation dataset yielded high agreement between estimated and true NDVIu with R2 and RMSE of 0.99 and 0.03, respectively. By the MODIS BRDF data, we got the estimate of NDVIu close to the in situ measured value (0.61 vs. 0.66 for estimate and measurement, respectively), and also reasonable seasonal patterns of NDVIu in 2010-2013. The results imply a potential application of the retrieved NDVIu to improve the estimation of overstory LAI for sparse boreal forests.
Out-of-Core Solutions of Complex Sparse Linear Equations
NASA Technical Reports Server (NTRS)
Yip, E. L.
1982-01-01
ETCLIB is library of subroutines for obtaining out-of-core solutions of complex sparse linear equations. Routines apply to dense and sparse matrices too large to be stored in core. Useful for solving any set of linear equations, but particularly useful in cases where coefficient matrix has no special properties that guarantee convergence with any of interative processes. The only assumption made is that coefficient matrix is not singular.
Testing of Error-Correcting Sparse Permutation Channel Codes
NASA Technical Reports Server (NTRS)
Shcheglov, Kirill, V.; Orlov, Sergei S.
2008-01-01
A computer program performs Monte Carlo direct numerical simulations for testing sparse permutation channel codes, which offer strong error-correction capabilities at high code rates and are considered especially suitable for storage of digital data in holographic and volume memories. A word in a code of this type is characterized by, among other things, a sparseness parameter (M) and a fixed number (K) of 1 or "on" bits in a channel block length of N.
Kernelized Elastic Net Regularization: Generalization Bounds, and Sparse Recovery.
Feng, Yunlong; Lv, Shao-Gao; Hang, Hanyuan; Suykens, Johan A K
2016-03-01
Kernelized elastic net regularization (KENReg) is a kernelization of the well-known elastic net regularization (Zou & Hastie, 2005 ). The kernel in KENReg is not required to be a Mercer kernel since it learns from a kernelized dictionary in the coefficient space. Feng, Yang, Zhao, Lv, and Suykens ( 2014 ) showed that KENReg has some nice properties including stability, sparseness, and generalization. In this letter, we continue our study on KENReg by conducting a refined learning theory analysis. This letter makes the following three main contributions. First, we present refined error analysis on the generalization performance of KENReg. The main difficulty of analyzing the generalization error of KENReg lies in characterizing the population version of its empirical target function. We overcome this by introducing a weighted Banach space associated with the elastic net regularization. We are then able to conduct elaborated learning theory analysis and obtain fast convergence rates under proper complexity and regularity assumptions. Second, we study the sparse recovery problem in KENReg with fixed design and show that the kernelization may improve the sparse recovery ability compared to the classical elastic net regularization. Finally, we discuss the interplay among different properties of KENReg that include sparseness, stability, and generalization. We show that the stability of KENReg leads to generalization, and its sparseness confidence can be derived from generalization. Moreover, KENReg is stable and can be simultaneously sparse, which makes it attractive theoretically and practically. PMID:26735744
Visual Tracking Based on Extreme Learning Machine and Sparse Representation
Wang, Baoxian; Tang, Linbo; Yang, Jinglin; Zhao, Baojun; Wang, Shuigen
2015-01-01
The existing sparse representation-based visual trackers mostly suffer from both being time consuming and having poor robustness problems. To address these issues, a novel tracking method is presented via combining sparse representation and an emerging learning technique, namely extreme learning machine (ELM). Specifically, visual tracking can be divided into two consecutive processes. Firstly, ELM is utilized to find the optimal separate hyperplane between the target observations and background ones. Thus, the trained ELM classification function is able to remove most of the candidate samples related to background contents efficiently, thereby reducing the total computational cost of the following sparse representation. Secondly, to further combine ELM and sparse representation, the resultant confidence values (i.e., probabilities to be a target) of samples on the ELM classification function are used to construct a new manifold learning constraint term of the sparse representation framework, which tends to achieve robuster results. Moreover, the accelerated proximal gradient method is used for deriving the optimal solution (in matrix form) of the constrained sparse tracking model. Additionally, the matrix form solution allows the candidate samples to be calculated in parallel, thereby leading to a higher efficiency. Experiments demonstrate the effectiveness of the proposed tracker. PMID:26506359
Vector sparse representation of color image using quaternion matrix analysis.
Xu, Yi; Yu, Licheng; Xu, Hongteng; Zhang, Hao; Nguyen, Truong
2015-04-01
Traditional sparse image models treat color image pixel as a scalar, which represents color channels separately or concatenate color channels as a monochrome image. In this paper, we propose a vector sparse representation model for color images using quaternion matrix analysis. As a new tool for color image representation, its potential applications in several image-processing tasks are presented, including color image reconstruction, denoising, inpainting, and super-resolution. The proposed model represents the color image as a quaternion matrix, where a quaternion-based dictionary learning algorithm is presented using the K-quaternion singular value decomposition (QSVD) (generalized K-means clustering for QSVD) method. It conducts the sparse basis selection in quaternion space, which uniformly transforms the channel images to an orthogonal color space. In this new color space, it is significant that the inherent color structures can be completely preserved during vector reconstruction. Moreover, the proposed sparse model is more efficient comparing with the current sparse models for image restoration tasks due to lower redundancy between the atoms of different color channels. The experimental results demonstrate that the proposed sparse image model avoids the hue bias issue successfully and shows its potential as a general and powerful tool in color image analysis and processing domain. PMID:25643407
A sparse embedding and least variance encoding approach to hashing.
Zhu, Xiaofeng; Zhang, Lei; Huang, Zi
2014-09-01
Hashing is becoming increasingly important in large-scale image retrieval for fast approximate similarity search and efficient data storage. Many popular hashing methods aim to preserve the kNN graph of high dimensional data points in the low dimensional manifold space, which is, however, difficult to achieve when the number of samples is big. In this paper, we propose an effective and efficient hashing approach by sparsely embedding a sample in the training sample space and encoding the sparse embedding vector over a learned dictionary. To this end, we partition the sample space into clusters via a linear spectral clustering method, and then represent each sample as a sparse vector of normalized probabilities that it falls into its several closest clusters. This actually embeds each sample sparsely in the sample space. The sparse embedding vector is employed as the feature of each sample for hashing. We then propose a least variance encoding model, which learns a dictionary to encode the sparse embedding feature, and consequently binarize the coding coefficients as the hash codes. The dictionary and the binarization threshold are jointly optimized in our model. Experimental results on benchmark data sets demonstrated the effectiveness of the proposed approach in comparison with state-of-the-art methods. PMID:24968174
Visual tracking based on extreme learning machine and sparse representation.
Wang, Baoxian; Tang, Linbo; Yang, Jinglin; Zhao, Baojun; Wang, Shuigen
2015-01-01
The existing sparse representation-based visual trackers mostly suffer from both being time consuming and having poor robustness problems. To address these issues, a novel tracking method is presented via combining sparse representation and an emerging learning technique, namely extreme learning machine (ELM). Specifically, visual tracking can be divided into two consecutive processes. Firstly, ELM is utilized to find the optimal separate hyperplane between the target observations and background ones. Thus, the trained ELM classification function is able to remove most of the candidate samples related to background contents efficiently, thereby reducing the total computational cost of the following sparse representation. Secondly, to further combine ELM and sparse representation, the resultant confidence values (i.e., probabilities to be a target) of samples on the ELM classification function are used to construct a new manifold learning constraint term of the sparse representation framework, which tends to achieve robuster results. Moreover, the accelerated proximal gradient method is used for deriving the optimal solution (in matrix form) of the constrained sparse tracking model. Additionally, the matrix form solution allows the candidate samples to be calculated in parallel, thereby leading to a higher efficiency. Experiments demonstrate the effectiveness of the proposed tracker. PMID:26506359
Representation-Independent Iteration of Sparse Data Arrays
NASA Technical Reports Server (NTRS)
James, Mark
2007-01-01
An approach is defined that describes a method of iterating over massively large arrays containing sparse data using an approach that is implementation independent of how the contents of the sparse arrays are laid out in memory. What is unique and important here is the decoupling of the iteration over the sparse set of array elements from how they are internally represented in memory. This enables this approach to be backward compatible with existing schemes for representing sparse arrays as well as new approaches. What is novel here is a new approach for efficiently iterating over sparse arrays that is independent of the underlying memory layout representation of the array. A functional interface is defined for implementing sparse arrays in any modern programming language with a particular focus for the Chapel programming language. Examples are provided that show the translation of a loop that computes a matrix vector product into this representation for both the distributed and not-distributed cases. This work is directly applicable to NASA and its High Productivity Computing Systems (HPCS) program that JPL and our current program are engaged in. The goal of this program is to create powerful, scalable, and economically viable high-powered computer systems suitable for use in national security and industry by 2010. This is important to NASA for its computationally intensive requirements for analyzing and understanding the volumes of science data from our returned missions.
Water Control on Vegetation Growth Pattern in Eurasia from GRACE
NASA Astrophysics Data System (ADS)
A, G.; Velicogna, I.; Kimball, J. S.; Kim, Y.
2014-12-01
High latitude ecosystem productivity is constrained by cold temperature and moisture limitations to plant growth, while these environmental restrictions may be changing with global warming. Satellite data driven assessments indicate that over the past three decades, rapid warming in the northern high latitudes has resulted in earlier and longer potential growing seasons and widespread greening, due to general relaxation of cold temperature constraints to vegetation productivity. However, warming may have also increased water stress limitations to growth. In this study, we use GRACE (Gravity Recovery and Climate Experiment) derived total water storage (TWS), 2-m air temperature (T) from ERA-interim reanalysis, normalized difference vegetation index (NDVI) data from MODIS (Collection 5) and satellite data driven vegetation gross primary productivity (GPP) estimates as surrogates for vegetation growth, for the period August 2002-December 2013 to evaluate terrestrial water supply controls to vegetation growth changes over the three major river basins of northern Eurasia. We find that during the analyzed period, the apparent growth response follows regional vegetation, moisture and temperature gradients and is spatially complex. In the drier southwest characterized by grassland, vegetation growth is mainly controlled by TWS availability. In the central region, dominated by cold temperature and water limited boreal forest, T is the main control on vegetation growth. In the Lena basin, where vegetation includes both boreal forest and water limited grassland, both T and TWS impact vegetation growth. We suggest that GRACE TWS estimates provide reliable observational constraints on water availability to vegetation that supplement sparse soil moisture observations and satellite precipitation estimates with unknown bias.
... the non-starchy vegetables . Starchy vegetables like potatoes, corn and peas are included in the " Grains and ... or Chinese spinach Artichoke Artichoke hearts Asparagus Baby corn Bamboo shoots Beans (green, wax, Italian) Bean sprouts ...
A healthy diet includes adding vegetables and fruit every day. Vegetables like broccoli, green beans, leafy greens, zucchini, cauliflower, cabbage, carrots, and tomatoes are low in calories and high in fiber, ...
Ecogeomorphology of Sand Dunes Shaped by Vegetation
NASA Astrophysics Data System (ADS)
Tsoar, H.
2014-12-01
Two dune types associated with vegetation are known: Parabolic and Vegetated Linear Dunes (VLDs), the latters are the dominant dune type in the world deserts. Parabolic dunes are formed in humid, sub-humid and semi-arid environments (rather than arid) where vegetation is nearby. VLDs are known today in semiarid and arid lands where the average yearly rainfall is ≥100 mm, enough to support sparse cover of vegetation. These two dune types are formed by unidirectional winds although they demonstrate a different form and have a distinct dynamics. Conceptual and mathematical models of dunes mobility and stability, based on three control parameters: wind power (DP), average annual precipitation (p), and the human impact parameter (μ) show that where human impact is negligible the effect of wind power (DP) on vegetative cover is substantial. The average yearly rainfall of 60-80 mm is the threshold of annual average rainfall for vegetation growth on dune sand. The model is shown to follow a hysteresis path, which explains the bistability of active and stabilized dunes under the same climatic conditions with respect to wind power. We have discerned formation of parabolic dunes from barchans and transverse dunes in the coastal plain of Israel where a decrease in human activity during the second half of the 20th century caused establishment of vegetation on the crest of the dunes, a process that changed the dynamics of these barchans and transverse dunes and led to a change in the shape of the windward slope from convex to concave. These dunes gradually became parabolic. It seems that VLDs in Australia or the Kalahari have always been vegetated to some degree, though the shrubs were sparser in colder periods when the aeolian erosion was sizeable. Those ancient conditions are characterized by higher wind power and lower rainfall that can reduce, but not completely destroy, the vegetation cover, leading to the formation of lee (shadow) dunes behind each shrub. Formation of
Technology Transfer Automated Retrieval System (TEKTRAN)
The vegetative shoot apical meristem is a highly organized yet dynamic structure. It is responsible for maintaining the proliferation of a population of undifferentiated stem cells through-out vegetative development, as well as for initiating lateral organs in stereotypical patterns. The vegetative...
NASA Technical Reports Server (NTRS)
Schrumpf, B. J.
1973-01-01
Unique characteristics of ERTS imagery can be used to inventory natural vegetation. While satellite images can seldom be interpreted and identified directly in terms of vegetation types, such types can be inferred by interpretation of physical terrain features and through an understanding of the ecology of the vegetation.
NASA Astrophysics Data System (ADS)
Gong, T. T.; Lei, H. M.; Yang, D. W.; Jiao, Y.; Yang, H. B.
2014-12-01
Vegetation change is a primary factor that affect evapotranspiration (ET), which is an important process in the hydrological cycle. In this study, an attempt is made to analyze the effects of vegetation change on ET using continuous observation data from eddy-covariance (EC) measurements over three periods (1 July 2011 to 30 June 2014) of a study site in a sparse shrubland study site located in the Loess Plateau of China, which is a fragile ecosystem experiencing serious soil desiccation. In our study, vegetation change includes phenological change and land use change. Phenological process of vegetation is validated to have a remarkable positive effect on ET in a rate of 1.83 ± 0.01 along with vegetation greening. Land use change at our study site was due to the native vegetation being cut-off by human activities, converting sparse shrubland to bare soil. With land use condition changing during the three years, annual total ET was observed to increase 103 mm, suggesting that soil evaporation consumes more water than canopy transpiration. In summary, the effects of vegetation change on ET suggest that both vegetation greening and increased area of exposed soil would aggravate the soil desiccation at our site in the north Loess Plateau.
Geostatistical estimation of signal-to-noise ratios for spectral vegetation indices
NASA Astrophysics Data System (ADS)
Ji, Lei; Zhang, Li; Rover, Jennifer; Wylie, Bruce K.; Chen, Xuexia
2014-10-01
In the past 40 years, many spectral vegetation indices have been developed to quantify vegetation biophysical parameters. An ideal vegetation index should contain the maximum level of signal related to specific biophysical characteristics and the minimum level of noise such as background soil influences and atmospheric effects. However, accurate quantification of signal and noise in a vegetation index remains a challenge, because it requires a large number of field measurements or laboratory experiments. In this study, we applied a geostatistical method to estimate signal-to-noise ratio (S/N) for spectral vegetation indices. Based on the sample semivariogram of vegetation index images, we used the standardized noise to quantify the noise component of vegetation indices. In a case study in the grasslands and shrublands of the western United States, we demonstrated the geostatistical method for evaluating S/N for a series of soil-adjusted vegetation indices derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor. The soil-adjusted vegetation indices were found to have higher S/N values than the traditional normalized difference vegetation index (NDVI) and simple ratio (SR) in the sparsely vegetated areas. This study shows that the proposed geostatistical analysis can constitute an efficient technique for estimating signal and noise components in vegetation indices.
Geostatistical estimation of signal-to-noise ratios for spectral vegetation indices
Ji, Lei; Zhang, Li; Rover, Jennifer R.; Wylie, Bruce K.; Chen, Xuexia
2014-01-01
In the past 40 years, many spectral vegetation indices have been developed to quantify vegetation biophysical parameters. An ideal vegetation index should contain the maximum level of signal related to specific biophysical characteristics and the minimum level of noise such as background soil influences and atmospheric effects. However, accurate quantification of signal and noise in a vegetation index remains a challenge, because it requires a large number of field measurements or laboratory experiments. In this study, we applied a geostatistical method to estimate signal-to-noise ratio (S/N) for spectral vegetation indices. Based on the sample semivariogram of vegetation index images, we used the standardized noise to quantify the noise component of vegetation indices. In a case study in the grasslands and shrublands of the western United States, we demonstrated the geostatistical method for evaluating S/N for a series of soil-adjusted vegetation indices derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor. The soil-adjusted vegetation indices were found to have higher S/N values than the traditional normalized difference vegetation index (NDVI) and simple ratio (SR) in the sparsely vegetated areas. This study shows that the proposed geostatistical analysis can constitute an efficient technique for estimating signal and noise components in vegetation indices.
Sparse modeling of spatial environmental variables associated with asthma.
Chang, Timothy S; Gangnon, Ronald E; David Page, C; Buckingham, William R; Tandias, Aman; Cowan, Kelly J; Tomasallo, Carrie D; Arndt, Brian G; Hanrahan, Lawrence P; Guilbert, Theresa W
2015-02-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin's Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5-50years over a three-year period. Each patient's home address was geocoded to one of 3456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin's geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. PMID:25533437
Sparse modeling of spatial environmental variables associated with asthma
Chang, Timothy S.; Gangnon, Ronald E.; Page, C. David; Buckingham, William R.; Tandias, Aman; Cowan, Kelly J.; Tomasallo, Carrie D.; Arndt, Brian G.; Hanrahan, Lawrence P.; Guilbert, Theresa W.
2014-01-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin’s Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5–50 years over a three-year period. Each patient’s home address was geocoded to one of 3,456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin’s geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. PMID:25533437
Pryde, E.H.
1982-01-01
Suggested standards for vegetable oils and ester fuels, as well as ASTM specifications for No. 2 diesel oil are given. The following physical properties were discussed: cetane number, cloud point, distillation temperatures, flash point, pour point, turbidity, viscosity, free fatty acids, iodine value, phosphorus, and wax. It was apparent that vegetable oils and their esters cannot meet ASTM specifications D975 for No. 2 diesel oil for use in the diesel engine. Vegetable oil modification or engine design modification may make it possible eventually for vegetable oils to become suitable alternative fuels. Vegetable oils must be recognized as experimental fuels until modifications have been tested thoroughly and generally accepted. 1 table. (DP)
Sparse approximation problem: how rapid simulated annealing succeeds and fails
NASA Astrophysics Data System (ADS)
Obuchi, Tomoyuki; Kabashima, Yoshiyuki
2016-03-01
Information processing techniques based on sparseness have been actively studied in several disciplines. Among them, a mathematical framework to approximately express a given dataset by a combination of a small number of basis vectors of an overcomplete basis is termed the sparse approximation. In this paper, we apply simulated annealing, a metaheuristic algorithm for general optimization problems, to sparse approximation in the situation where the given data have a planted sparse representation and noise is present. The result in the noiseless case shows that our simulated annealing works well in a reasonable parameter region: the planted solution is found fairly rapidly. This is true even in the case where a common relaxation of the sparse approximation problem, the G-relaxation, is ineffective. On the other hand, when the dimensionality of the data is close to the number of non-zero components, another metastable state emerges, and our algorithm fails to find the planted solution. This phenomenon is associated with a first-order phase transition. In the case of very strong noise, it is no longer meaningful to search for the planted solution. In this situation, our algorithm determines a solution with close-to-minimum distortion fairly quickly.
Sparsely sampling the sky: Regular vs. random sampling
NASA Astrophysics Data System (ADS)
Paykari, P.; Pires, S.; Starck, J.-L.; Jaffe, A. H.
2015-09-01
Aims: The next generation of galaxy surveys, aiming to observe millions of galaxies, are expensive both in time and money. This raises questions regarding the optimal investment of this time and money for future surveys. In a previous work, we have shown that a sparse sampling strategy could be a powerful substitute for the - usually favoured - contiguous observation of the sky. In our previous paper, regular sparse sampling was investigated, where the sparse observed patches were regularly distributed on the sky. The regularity of the mask introduces a periodic pattern in the window function, which induces periodic correlations at specific scales. Methods: In this paper, we use a Bayesian experimental design to investigate a "random" sparse sampling approach, where the observed patches are randomly distributed over the total sparsely sampled area. Results: We find that in this setting, the induced correlation is evenly distributed amongst all scales as there is no preferred scale in the window function. Conclusions: This is desirable when we are interested in any specific scale in the galaxy power spectrum, such as the matter-radiation equality scale. As the figure of merit shows, however, there is no preference between regular and random sampling to constrain the overall galaxy power spectrum and the cosmological parameters.
X-ray computed tomography using curvelet sparse regularization
Wieczorek, Matthias Vogel, Jakob; Lasser, Tobias; Frikel, Jürgen; Demaret, Laurent; Eggl, Elena; Pfeiffer, Franz; Kopp, Felix; Noël, Peter B.
2015-04-15
Purpose: Reconstruction of x-ray computed tomography (CT) data remains a mathematically challenging problem in medical imaging. Complementing the standard analytical reconstruction methods, sparse regularization is growing in importance, as it allows inclusion of prior knowledge. The paper presents a method for sparse regularization based on the curvelet frame for the application to iterative reconstruction in x-ray computed tomography. Methods: In this work, the authors present an iterative reconstruction approach based on the alternating direction method of multipliers using curvelet sparse regularization. Results: Evaluation of the method is performed on a specifically crafted numerical phantom dataset to highlight the method’s strengths. Additional evaluation is performed on two real datasets from commercial scanners with different noise characteristics, a clinical bone sample acquired in a micro-CT and a human abdomen scanned in a diagnostic CT. The results clearly illustrate that curvelet sparse regularization has characteristic strengths. In particular, it improves the restoration and resolution of highly directional, high contrast features with smooth contrast variations. The authors also compare this approach to the popular technique of total variation and to traditional filtered backprojection. Conclusions: The authors conclude that curvelet sparse regularization is able to improve reconstruction quality by reducing noise while preserving highly directional features.
Image fusion via nonlocal sparse K-SVD dictionary learning.
Li, Ying; Li, Fangyi; Bai, Bendu; Shen, Qiang
2016-03-01
Image fusion aims to merge two or more images captured via various sensors of the same scene to construct a more informative image by integrating their details. Generally, such integration is achieved through the manipulation of the representations of the images concerned. Sparse representation plays an important role in the effective description of images, offering a great potential in a variety of image processing tasks, including image fusion. Supported by sparse representation, in this paper, an approach for image fusion by the use of a novel dictionary learning scheme is proposed. The nonlocal self-similarity property of the images is exploited, not only at the stage of learning the underlying description dictionary but during the process of image fusion. In particular, the property of nonlocal self-similarity is combined with the traditional sparse dictionary. This results in an improved learned dictionary, hereafter referred to as the nonlocal sparse K-SVD dictionary (where K-SVD stands for the K times singular value decomposition that is commonly used in the literature), and abbreviated to NL_SK_SVD. The performance of the NL_SK_SVD dictionary is applied for image fusion using simultaneous orthogonal matching pursuit. The proposed approach is evaluated with different types of images, and compared with a number of alternative image fusion techniques. The resultant superior fused images using the present approach demonstrates the efficacy of the NL_SK_SVD dictionary in sparse image representation. PMID:26974648
Online Hierarchical Sparse Representation of Multifeature for Robust Object Tracking
Qu, Shiru
2016-01-01
Object tracking based on sparse representation has given promising tracking results in recent years. However, the trackers under the framework of sparse representation always overemphasize the sparse representation and ignore the correlation of visual information. In addition, the sparse coding methods only encode the local region independently and ignore the spatial neighborhood information of the image. In this paper, we propose a robust tracking algorithm. Firstly, multiple complementary features are used to describe the object appearance; the appearance model of the tracked target is modeled by instantaneous and stable appearance features simultaneously. A two-stage sparse-coded method which takes the spatial neighborhood information of the image patch and the computation burden into consideration is used to compute the reconstructed object appearance. Then, the reliability of each tracker is measured by the tracking likelihood function of transient and reconstructed appearance models. Finally, the most reliable tracker is obtained by a well established particle filter framework; the training set and the template library are incrementally updated based on the current tracking results. Experiment results on different challenging video sequences show that the proposed algorithm performs well with superior tracking accuracy and robustness.
[Investigation of Multi-Angle Polarization Properties of Vegetation Based on RSP].
Jiao, Jian-nan; Zhao, Hai-meng; Yang, Bin; Yan, Lei
2016-02-01
Polarization detection provides us with novel information to reflect the target attribute. Compared with traditional remote sensing methods, multi-angle polarization has relatively stable correlation and regularity. RSP(research scanning polarimeter)is an airborne prototype for the APS(aerosol polarimetery sensor) developed by the USA, which can provide with us the polarization detection information of 9 channels. We can get optical properties and physical characteristics of vegetation by analyzing stable multi-angle and multi-band polarization detection information from preprocessing scanning polarization data of flight test. In this paper, after making registration based on flight attitude information, a comparative analysis is made between characteristics of reflectance and polarization reflectance with visible light and near infrared band of the view zenith angles between--30 degree and 65 degree, based on dense area and sparse area(close to bare field) of vegetation. The results show that both dense area and sparse area demonstrate regular characteristics of polarization degree. The area close to hot spot area has highest reflectance energy. In contrast,. it has relatively least energy of polarization degree, which can prevent strong reflectance energy from influencing the stability of detector. Because the degree of polarization in dense area of vegetation is higher than that in sparse area at visible light band while that in concentration area of vegetation is lower than sparse area at near infrared light band, it shows that the visible light band information of dense area of vegetation that the sensor received is dominated by single scattering while the near infrared light band information of dense area of vegetation is dominated by multiple scattering. PMID:27209749
Uncovering Transcriptional Regulatory Networks by Sparse Bayesian Factor Model
NASA Astrophysics Data System (ADS)
Meng, Jia; Zhang, Jianqiu(Michelle); Qi, Yuan(Alan); Chen, Yidong; Huang, Yufei
2010-12-01
The problem of uncovering transcriptional regulation by transcription factors (TFs) based on microarray data is considered. A novel Bayesian sparse correlated rectified factor model (BSCRFM) is proposed that models the unknown TF protein level activity, the correlated regulations between TFs, and the sparse nature of TF-regulated genes. The model admits prior knowledge from existing database regarding TF-regulated target genes based on a sparse prior and through a developed Gibbs sampling algorithm, a context-specific transcriptional regulatory network specific to the experimental condition of the microarray data can be obtained. The proposed model and the Gibbs sampling algorithm were evaluated on the simulated systems, and results demonstrated the validity and effectiveness of the proposed approach. The proposed model was then applied to the breast cancer microarray data of patients with Estrogen Receptor positive ([InlineEquation not available: see fulltext.]) status and Estrogen Receptor negative ([InlineEquation not available: see fulltext.]) status, respectively.
PSPIKE: A Parallel Hybrid Sparse Linear System Solver
NASA Astrophysics Data System (ADS)
Manguoglu, Murat; Sameh, Ahmed H.; Schenk, Olaf
The availability of large-scale computing platforms comprised of tens of thousands of multicore processors motivates the need for the next generation of highly scalable sparse linear system solvers. These solvers must optimize parallel performance, processor (serial) performance, as well as memory requirements, while being robust across broad classes of applications and systems. In this paper, we present a new parallel solver that combines the desirable characteristics of direct methods (robustness) and effective iterative solvers (low computational cost), while alleviating their drawbacks (memory requirements, lack of robustness). Our proposed hybrid solver is based on the general sparse solver PARDISO, and the “Spike” family of hybrid solvers. The resulting algorithm, called PSPIKE, is as robust as direct solvers, more reliable than classical preconditioned Krylov subspace methods, and much more scalable than direct sparse solvers. We support our performance and parallel scalability claims using detailed experimental studies and comparison with direct solvers, as well as classical preconditioned Krylov methods.
Automatic landslide and mudflow detection method via multichannel sparse representation
NASA Astrophysics Data System (ADS)
Chao, Chen; Zhou, Jianjun; Hao, Zhuo; Sun, Bo; He, Jun; Ge, Fengxiang
2015-10-01
Landslide and mudflow detection is an important application of aerial images and high resolution remote sensing images, which is crucial for national security and disaster relief. Since the high resolution images are often large in size, it's necessary to develop an efficient algorithm for landslide and mudflow detection. Based on the theory of sparse representation and, we propose a novel automatic landslide and mudflow detection method in this paper, which combines multi-channel sparse representation and eight neighbor judgment methods. The whole process of the detection is totally automatic. We make the experiment on a high resolution image of ZhouQu district of Gansu province in China on August, 2010 and get a promising result which proved the effective of using sparse representation on landslide and mudflow detection.
A Multiobjective Sparse Feature Learning Model for Deep Neural Networks.
Gong, Maoguo; Liu, Jia; Li, Hao; Cai, Qing; Su, Linzhi
2015-12-01
Hierarchical deep neural networks are currently popular learning models for imitating the hierarchical architecture of human brain. Single-layer feature extractors are the bricks to build deep networks. Sparse feature learning models are popular models that can learn useful representations. But most of those models need a user-defined constant to control the sparsity of representations. In this paper, we propose a multiobjective sparse feature learning model based on the autoencoder. The parameters of the model are learnt by optimizing two objectives, reconstruction error and the sparsity of hidden units simultaneously to find a reasonable compromise between them automatically. We design a multiobjective induced learning procedure for this model based on a multiobjective evolutionary algorithm. In the experiments, we demonstrate that the learning procedure is effective, and the proposed multiobjective model can learn useful sparse features. PMID:26340790
Sparse representation based face recognition using weighted regions
NASA Astrophysics Data System (ADS)
Bilgazyev, Emil; Yeniaras, E.; Uyanik, I.; Unan, Mahmut; Leiss, E. L.
2013-12-01
Face recognition is a challenging research topic, especially when the training (gallery) and recognition (probe) images are acquired using different cameras under varying conditions. Even a small noise or occlusion in the images can compromise the accuracy of recognition. Lately, sparse encoding based classification algorithms gave promising results for such uncontrollable scenarios. In this paper, we introduce a novel methodology by modeling the sparse encoding with weighted patches to increase the robustness of face recognition even further. In the training phase, we define a mask (i.e., weight matrix) using a sparse representation selecting the facial regions, and in the recognition phase, we perform comparison on selected facial regions. The algorithm was evaluated both quantitatively and qualitatively using two comprehensive surveillance facial image databases, i.e., SCfaceandMFPV, with the results clearly superior to common state-of-the-art methodologies in different scenarios.
Human interaction recognition through two-phase sparse coding
NASA Astrophysics Data System (ADS)
Zhang, B.; Conci, N.; De Natale, Francesco G. B.
2014-03-01
In this paper, we propose a novel method to recognize two-person interactions through a two-phase sparse coding approach. In the first phase, we adopt the non-negative sparse coding on the spatio-temporal interest points (STIPs) extracted from videos, and then construct the feature vector for each video by sum-pooling and l2-normalization. At the second stage, we apply the label-consistent KSVD (LC-KSVD) algorithm on the video feature vectors to train a new dictionary. The algorithm has been validated on the TV human interaction dataset, and the experimental results show that the classification performance is considerably improved compared with the standard bag-of-words approach and the single layer non-negative sparse coding.
Analyzing Sparse Dictionaries for Online Learning With Kernels
NASA Astrophysics Data System (ADS)
Honeine, Paul
2015-12-01
Many signal processing and machine learning methods share essentially the same linear-in-the-parameter model, with as many parameters as available samples as in kernel-based machines. Sparse approximation is essential in many disciplines, with new challenges emerging in online learning with kernels. To this end, several sparsity measures have been proposed in the literature to quantify sparse dictionaries and constructing relevant ones, the most prolific ones being the distance, the approximation, the coherence and the Babel measures. In this paper, we analyze sparse dictionaries based on these measures. By conducting an eigenvalue analysis, we show that these sparsity measures share many properties, including the linear independence condition and inducing a well-posed optimization problem. Furthermore, we prove that there exists a quasi-isometry between the parameter (i.e., dual) space and the dictionary's induced feature space.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
Vitello, P; Fried, L E; Pudliner, B; McAbee, T
2003-07-14
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values, called the ''cache'' is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamic calculations, CHEETAH uses MPI communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.
P-SPARSLIB: A parallel sparse iterative solution package
Saad, Y.
1994-12-31
Iterative methods are gaining popularity in engineering and sciences at a time where the computational environment is changing rapidly. P-SPARSLIB is a project to build a software library for sparse matrix computations on parallel computers. The emphasis is on iterative methods and the use of distributed sparse matrices, an extension of the domain decomposition approach to general sparse matrices. One of the goals of this project is to develop a software package geared towards specific applications. For example, the author will test the performance and usefulness of P-SPARSLIB modules on linear systems arising from CFD applications. Equally important is the goal of portability. In the long run, the author wishes to ensure that this package is portable on a variety of platforms, including SIMD environments and shared memory environments.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
NASA Astrophysics Data System (ADS)
Vitello, Peter; Fried, Laurence E.; Pudliner, Brian; McAbee, Tom
2004-07-01
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamic calculations, CHEETAH uses parallel communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH coupled to an ALE hydrocode. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.
Sparse Sensing of Aerodynamic Loads on Insect Wings
NASA Astrophysics Data System (ADS)
Manohar, Krithika; Brunton, Steven; Kutz, J. Nathan
2015-11-01
We investigate how insects use sparse sensors on their wings to detect aerodynamic loading and wing deformation using a coupled fluid-structure model given periodically flapping input motion. Recent observations suggest that insects collect sensor information about their wing deformation to inform control actions for maneuvering and rejecting gust disturbances. Given a small number of point measurements of the chordwise aerodynamic loads from the sparse sensors, we reconstruct the entire chordwise loading using sparsesensing - a signal processing technique that reconstructs a signal from a small number of measurements using l1 norm minimization of sparse modal coefficients in some basis. We compare reconstructions from sensors randomly sampled from probability distributions biased toward different regions along the wing chord. In this manner, we determine the preferred regions along the chord for sensor placement and for estimating chordwise loads to inform control decisions in flight.
Sparse-based multispectral image encryption via ptychography
NASA Astrophysics Data System (ADS)
Rawat, Nitin; Shi, Yishi; Kim, Byoungho; Lee, Byung-Geun
2015-12-01
Recently, we proposed a model of securing a ptychography-based monochromatic image encryption system via the classical Photon-counting imaging (PCI) technique. In this study, we examine a single-channel multispectral sparse-based photon-counting ptychography imaging (SMPI)-based cryptosystem. A ptychography-based cryptosystem creates a complex object wave field, which can be reconstructed by a series of diffraction intensity patterns through an aperture movement. The PCI sensor records only a few complex Bayer patterned samples that have been utilized in the decryption process. Sparse sensing and nonlinear properties of the classical PCI system, together with the scanning probes, enlarge the key space, and such a combination therefore enhances the system's security. We demonstrate that the sparse samples have adequate information for image decryption, as well as information authentication by means of optical correlation.
Joint sparse representation for robust multimodal biometrics recognition.
Shekhar, Sumit; Patel, Vishal M; Nasrabadi, Nasser M; Chellappa, Rama
2014-01-01
Traditional biometric recognition systems rely on a single biometric signature for authentication. While the advantage of using multiple sources of information for establishing the identity has been widely recognized, computational models for multimodal biometrics recognition have only recently received attention. We propose a multimodal sparse representation method, which represents the test data by a sparse linear combination of training data, while constraining the observations from different modalities of the test subject to share their sparse representations. Thus, we simultaneously take into account correlations as well as coupling information among biometric modalities. A multimodal quality measure is also proposed to weigh each modality as it gets fused. Furthermore, we also kernelize the algorithm to handle nonlinearity in data. The optimization problem is solved using an efficient alternative direction method. Various experiments show that the proposed method compares favorably with competing fusion-based methods. PMID:24231870
SAR target classification based on multiscale sparse representation
NASA Astrophysics Data System (ADS)
Ruan, Huaiyu; Zhang, Rong; Li, Jingge; Zhan, Yibing
2016-03-01
We propose a novel multiscale sparse representation approach for SAR target classification. It firstly extracts the dense SIFT descriptors on multiple scales, then trains a global multiscale dictionary by sparse coding algorithm. After obtaining the sparse representation, the method applies spatial pyramid matching (SPM) and max pooling to summarize the features for each image. The proposed method can provide more information and descriptive ability than single-scale ones. Moreover, it costs less extra computation than existing multiscale methods which compute a dictionary for each scale. The MSTAR database and ship database collected from TerraSAR-X images are used in classification setup. Results show that the best overall classification rate of the proposed approach can achieve 98.83% on the MSTAR database and 92.67% on the TerraSAR-X ship database.
A note on rank reduction in sparse multivariate regression
Chen, Kun; Chan, Kung-Sik
2016-01-01
A reduced-rank regression with sparse singular value decomposition (RSSVD) approach was proposed by Chen et al. for conducting variable selection in a reduced-rank model. To jointly model the multivariate response, the method efficiently constructs a prespecified number of latent variables as some sparse linear combinations of the predictors. Here, we generalize the method to also perform rank reduction, and enable its usage in reduced-rank vector autoregressive (VAR) modeling to perform automatic rank determination and order selection. We show that in the context of stationary time-series data, the generalized approach correctly identifies both the model rank and the sparse dependence structure between the multivariate response and the predictors, with probability one asymptotically. We demonstrate the efficacy of the proposed method by simulations and analyzing a macro-economical multivariate time series using a reduced-rank VAR model. PMID:26997938
The sparseness of neuronal responses in ferret primary visual cortex.
Tolhurst, David J; Smyth, Darragh; Thompson, Ian D
2009-02-25
Various arguments suggest that neuronal coding of natural sensory stimuli should be sparse (i.e., individual neurons should respond rarely but should respond reliably). We examined sparseness of visual cortical neurons in anesthetized ferret to flashed natural scenes. Response behavior differed widely between neurons. The median firing rate of 4.1 impulses per second was slightly higher than predicted from consideration of metabolic load. Thirteen percent of neurons (12 of 89) responded to <5% of the images, but one-half responded to >25% of images. Multivariate analysis of the range of sparseness values showed that 67% of the variance was accounted for by differing response patterns to moving gratings. Repeat presentation of images showed that response variance for natural images exaggerated sparseness measures; variance was scaled with mean response, but with a lower Fano factor than for the responses to moving gratings. This response variability and the "soft" sparse responses (Rehn and Sommer, 2007) raise the question of what constitutes a reliable neuronal response and imply parallel signaling by multiple neurons. We investigated whether the temporal structure of responses might be reliable enough to give additional information about natural scenes. Poststimulus time histogram shape was similar for "strong" and "weak" stimuli, with no systematic change in first-spike latency with stimulus strength. The variance of first-spike latency for repeat presentations of the same image was greater than the latency variance between images. In general, responses to flashed natural scenes do not seem compatible with a sparse encoding in which neurons fire rarely but reliably. PMID:19244512
Learning control for robotic manipulators with sparse data
NASA Technical Reports Server (NTRS)
Morita, Atsushi; Dubowsky, Steven; Hootsmans, Norbert A. M.
1987-01-01
Learning control algorithms have been proposed for error compensation in repetitive robotic manipulator tasks. It is shown that the performance of such control algorithms can be seriously degraded when the feedback data they use is relatively sparse in time, such as might be provided by vision systems. It is also shown that learning control algorithms can be modified to compensate for the effects of sparse data and thereby yield performance which approaches that of systems without limitations on the sensory information available for control.
Content addressable systolic array for sparse matrix computation
Wing, O.
1983-01-01
A systolic array is proposed which is specifically designed to solve a system of sparse linear equations. The array consists of a number of processing elements connected in a ring. Each processing element has its own content addressable memory where the nonzero elements of the sparse matrix are stored. Matrix elements to which elementary operations are applied are extracted from the memory by content addressing. The system of equations is solved in a systolic fashion and the solution is obtained in nz+5n-2 steps where nz is the number of nonzero elements along and below the diagonal and n is the number of equations. 13 references.
Reconstruction Techniques for Sparse Multistatic Linear Array Microwave Imaging
Sheen, David M.; Hall, Thomas E.
2014-06-09
Sequentially-switched linear arrays are an enabling technology for a number of near-field microwave imaging applications. Electronically sequencing along the array axis followed by mechanical scanning along an orthogonal axis allows dense sampling of a two-dimensional aperture in near real-time. In this paper, a sparse multi-static array technique will be described along with associated Fourier-Transform-based and back-projection-based image reconstruction algorithms. Simulated and measured imaging results are presented that show the effectiveness of the sparse array technique along with the merits and weaknesses of each image reconstruction approach.
Sparse Matrix for ECG Identification with Two-Lead Features
Tseng, Kuo-Kun; Luo, Jiao; Wang, Wenmin; Haiting, Dong
2015-01-01
Electrocardiograph (ECG) human identification has the potential to improve biometric security. However, improvements in ECG identification and feature extraction are required. Previous work has focused on single lead ECG signals. Our work proposes a new algorithm for human identification by mapping two-lead ECG signals onto a two-dimensional matrix then employing a sparse matrix method to process the matrix. And that is the first application of sparse matrix techniques for ECG identification. Moreover, the results of our experiments demonstrate the benefits of our approach over existing methods. PMID:25961074
D Super-Resolution Approach for Sparse Laser Scanner Data
NASA Astrophysics Data System (ADS)
Hosseinyalamdary, S.; Yilmaz, A.
2015-08-01
Laser scanner point cloud has been emerging in Photogrammetry and computer vision to achieve high level tasks such as object tracking, object recognition and scene understanding. However, low cost laser scanners are noisy, sparse and prone to systematic errors. This paper proposes a novel 3D super resolution approach to reconstruct surface of the objects in the scene. This method works on sparse, unorganized point clouds and has superior performance over other surface recovery approaches. Since the proposed approach uses anisotropic diffusion equation, it does not deteriorate the object boundaries and it preserves topology of the object.
Three-dimensional sparse-aperture moving-target imaging
NASA Astrophysics Data System (ADS)
Ferrara, Matthew; Jackson, Julie; Stuff, Mark
2008-04-01
If a target's motion can be determined, the problem of reconstructing a 3D target image becomes a sparse-aperture imaging problem. That is, the data lies on a random trajectory in k-space, which constitutes a sparse data collection that yields very low-resolution images if backprojection or other standard imaging techniques are used. This paper investigates two moving-target imaging algorithms: the first is a greedy algorithm based on the CLEAN technique, and the second is a version of Basis Pursuit Denoising. The two imaging algorithms are compared for a realistic moving-target motion history applied to a Xpatch-generated backhoe data set.
A LONE code for the sparse control of quantum systems
NASA Astrophysics Data System (ADS)
Ciaramella, G.; Borzì, A.
2016-03-01
In many applications with quantum spin systems, control functions with a sparse and pulse-shaped structure are often required. These controls can be obtained by solving quantum optimal control problems with L1-penalized cost functionals. In this paper, the MATLAB package LONE is presented aimed to solving L1-penalized optimal control problems governed by unitary-operator quantum spin models. This package implements a new strategy that includes a globalized semi-smooth Krylov-Newton scheme and a continuation procedure. Results of numerical experiments demonstrate the ability of the LONE code in computing accurate sparse optimal control solutions.
BIRD: A general interface for sparse distributed memory simulators
NASA Technical Reports Server (NTRS)
Rogers, David
1990-01-01
Kanerva's sparse distributed memory (SDM) has now been implemented for at least six different computers, including SUN3 workstations, the Apple Macintosh, and the Connection Machine. A common interface for input of commands would both aid testing of programs on a broad range of computer architectures and assist users in transferring results from research environments to applications. A common interface also allows secondary programs to generate command sequences for a sparse distributed memory, which may then be executed on the appropriate hardware. The BIRD program is an attempt to create such an interface. Simplifying access to different simulators should assist developers in finding appropriate uses for SDM.