STOMP Sparse Vegetation Evapotranspiration Model for the Water-Air-Energy Operational Mode
Ward, Anderson L.; White, Mark D.; Freeman, Eugene J.; Zhang, Z. F.
2005-09-15
The Water-Air-Energy (WAE) Operational Mode of the Subsurface Transport Over Multiple Phases (STOMP) numerical simulator solves the coupled conservation equations for water mass, air mass, and thermal energy in multiple dimensions. This addendum describes the theory, input file formatting, and application of a soil-vegetation-atmosphere transfer (SVAT) scheme for STOMP that is based on a sparse vegetation evapotranspiration model. The SVAT scheme is implemented as a boundary condition on the upper surface of the computational domain and has capabilities for simulating evaporation from bare surfaces as well as evapotranspiration from sparsely vegetated surfaces populated with single or multiple plant species in response to meteorological forcings. With this extension, the model calculates water mass, air mass and thermal energy across a boundary surface in addition to root-water transport between the subsurface and atmosphere. This mode represents the barrier extension of the WAE mode and is designated as STOMP-WAE-B. Input for STOMP-WAE-B is specified via three input cards and include: atmospheric conditions through the Atmospheric Conditions Card; time-invariant plant species data through the Plant Properties Card; and time varying plant species data through the Boundary Conditions Card. Two optional cards, the Observed Data and UCODE Control Cards allow use of STOMP-WAE with UCODE in an inverse mode to estimate model parameters. STOMP-WAE was validated by solving a number of test problems from the literature that included experimental observations as well as analytical or numerical solutions. Several of the UNSAT-H verification problems are included along with a benchmark simulation derived from a recently published intercode comparison for barrier design tools. Results show that STOMP is able to meet, and in most cases, exceed performance of other commonly used simulation codes without having to resort to may of their simplifying assumptions. Use of the fully
ERIC Educational Resources Information Center
Martinez, Luz Adriana
2007-01-01
The Specialized Training of Military Parents, better known by many as STOMP, is a federally funded Parent Training and Information (PTI) center established to assist military families who have children with specialized education or health needs. STOMP exists to empower military parents, individuals with disabilities, and service providers with…
Partitioning evapotranspiration in sparsely vegetated rangeland using a portable chamber
Stannard, D.I.; Weltz, M.A.
2006-01-01
A portable chamber was used to separate evapotranspiration (ET) from a sparse, mixed-species shrub canopy in southeastern Arizona, United States, into vegetation and soil components. Chamber measurements were made of ET from the five dominant species, and from bare soil, on 3 days during the monsoon season when the soil surface was dry. The chamber measurements were assembled into landscape ET using a simple geometric model of the vegetated land surface. Chamber estimates of landscape ET were well correlated with, but about 26% greater than, simultaneous eddy-correlation measurements. Excessive air speed inside the chamber appears to be the primary cause of the overestimate. Overall, transpiration accounted for 84% of landscape ET, and bare soil evaporation for 16%. Desert zinnia, a small (???0.1 m high) but abundant species, was the greatest water user, both per unit area of shrub and of landscape. Partitioning of ET into components varied as a function of air temperature and shallow soil moisture. Transpiration from shorter species was more highly correlated with air temperature whereas transpiration from taller species was more highly correlated with shallow soil moisture. Application of these results to a full drying cycle between rainfalls at a similar site suggests that during the monsoon, ET at such sites may be about equally partitioned between transpiration and bare soil evaporation.
Wu, Jun-Jun; Gao, Zhi-Hai; Li, Zeng-Yuan; Wang, Hong-Yan; Pang, Yong; Sun, Bin; Li, Chang-Long; Li, Xu-Zhi; Zhang, Jiu-Xing
2014-03-01
In order to estimate the sparse vegetation information accurately in desertification region, taking southeast of Sunite Right Banner, Inner Mongolia, as the test site and Tiangong-1 hyperspectral image as the main data, sparse vegetation coverage and biomass were retrieved based on normalized difference vegetation index (NDVI) and soil adjusted vegetation index (SAVI), combined with the field investigation data. Then the advantages and disadvantages between them were compared. Firstly, the correlation between vegetation indexes and vegetation coverage under different bands combination was analyzed, as well as the biomass. Secondly, the best bands combination was determined when the maximum correlation coefficient turned up between vegetation indexes (VI) and vegetation parameters. It showed that the maximum correlation coefficient between vegetation parameters and NDVI could reach as high as 0.7, while that of SAVI could nearly reach 0.8. The center wavelength of red band in the best bands combination for NDVI was 630nm, and that of the near infrared (NIR) band was 910 nm. Whereas, when the center wavelength was 620 and 920 nm respectively, they were the best combination for SAVI. Finally, the linear regression models were established to retrieve vegetation coverage and biomass based on Tiangong-1 VIs. R2 of all models was more than 0.5, while that of the model based on SAVI was higher than that based on NDVI, especially, the R2 of vegetation coverage retrieve model based on SAVI was as high as 0.59. By intersection validation, the standard errors RMSE based on SAVI models were lower than that of the model based on NDVI. The results showed that the abundant spectral information of Tiangong-1 hyperspectral image can reflect the actual vegetaion condition effectively, and SAVI can estimate the sparse vegetation information more accurately than NDVI in desertification region.
STOMP Subsurface Transport Over Multiple Phases: STOMP-CO2 and STOMP-CO2e Guide: Version 1.0
White, Mark D.; Bacon, Diana H.; McGrail, B. Peter; Watson, David J.; White, Signe K.; Zhang, Z. F.
2012-04-03
This STOMP (Subsurface Transport Over Multiple Phases) guide document describes the theory, use, and application of the STOMP-CO2 and STOMP-CO2e operational modes. These operational modes of the STOMP simulator are configured to solve problems involving the sequestration of CO2 in geologic saline reservoirs. STOMP-CO2 is the isothermal version and STOMP-CO2e is the nonisothermal version. These core operational modes solve the governing conservation equations for component flow and transport through geologic media; where, the STOMP-CO2 components are water, CO2 and salt and the STOMP-CO2e operational mode also includes an energy conservation equation. Geochemistry can be included in the problem solution via the ECKEChem (Equilibrium-Conservation-Kinetic-Equation Chemistry) module, and geomechanics via the EPRMech (Elastic-Plastic-Rock Mechanics) module. This addendum is designed to provide the new user with a full guide for the core capabilities of the STOMP-CO2 and -CO2e simulators, and to provide the experienced user with a quick reference on implementing features. Several benchmark problems are provided in this addendum, which serve as starting points for developing inputs for more complex problems and as demonstrations of the simulator’s capabilities.
Energy budgets and resistances to energy transport in sparsely vegetated rangeland
Nichols, W.D.
1992-01-01
Partitioning available energy between plants and bare soil in sparsely vegetated rangelands will allow hydrologists and others to gain a greater understanding of water use by native vegetation, especially phreatophytes. Standard methods of conducting energy budget studies result in measurements of latent and sensible heat fluxes above the plant canopy which therefore include the energy fluxes from both the canopy and the soil. One-dimensional theoretical numerical models have been proposed recently for the partitioning of energy in sparse crops. Bowen ratio and other micrometeorological data collected over phreatophytes growing in areas of shallow ground water in central Nevada were used to evaluate the feasibility of using these models, which are based on surface and within-canopy aerodynamic resistances, to determine heat and water vapor transport in sparsely vegetated rangelands. The models appear to provide reasonably good estimates of sensible heat flux from the soil and latent heat flux from the canopy. Estimates of latent heat flux from the soil were less satisfactory. Sensible heat flux from the canopy was not well predicted by the present resistance formulations. Also, estimates of total above-canopy fluxes were not satisfactory when using a single value for above-canopy bulk aerodynamic resistance. ?? 1992.
Biomechanical assessment of the injury risk of stomping.
Schirmer, Frederike; Muggenthaler, Holger; Hubig, Michael; Schenkl, Sebastian; Koch, Markus; Blickhan, Reinhard; Mall, Gita
2016-05-01
Forensic case work as well as literature shows that severe head injuries, e.g., with basilar fractures and cerebral hemorrhages due to stomps can be seen; however, there is no data basis concerning contact forces and potential influencing factors. The objective of this work was to generate a data basis of contact forces in stomping by performing experimental measurements and subsequent statistical analyses. Fifty-five volunteers participated in the present study. Each participant performed several stomps onto force plates with sturdy/soft footwear as well as with/without an elastic layer imitating the scalp. Ground reaction forces induced by jumps were also measured for sturdy and soft footwear. The results show statistically significant dependencies between maximum ground reaction forces and body weight and body height. A statistically significant influence of footwear on stomping force could only be found in tests with an elastic layer and in the jumping setup. Mean maximum stomping forces for the female volunteers were between 4694 and 5970 N; male volunteers were able to produce mean peak stomping forces between 8494 and 9016 N. Jumping forces were approximately twice the stomping forces for both male and female test persons. Regardless of footwear and gender, it can be claimed that a forceful stomp or jump to someone's head supported on the ground can cause facial and skull fractures. Thus, forceful stomps or jumps to someone's head can cause potential fatal injuries independent of footwear, gender, or fitness level.
Turbulent heat transfer from a sparsely vegetated surface - Two-component representation
NASA Technical Reports Server (NTRS)
Otterman, J.; Novak, M. D.; Starr, D. O'C.
1993-01-01
The conventional calculation of heat fluxes from a vegetated surface involving the coefficient of turbulent heat transfer which increases logarithmically with surface roughness, is inappropriate such highly structured surfaces as desert scrub or open forest. An approach is developed here for computing sensible heat flux from sparsely vegetated surfaces, where the absorption of insolation and the transfer of absorbed heat to the atmosphere are calculated separately for the plants and for the soil. This approach is applied to a desert-scrub surface in the northern Sinai, for which the turbulent transfer coefficient of sensible heat flux from the plants is much larger than that from the soil below, as shown by an analysis of plant, soil, and air temperatures. The plant density is expressed as the sum of products (plant-height) x (plant-diameter) of plants per unit horizontal surface area. The solar heat absorbed by the plants is assumed to be transferred immediately to the airflow. The effective turbulent transfer coefficient k(g-eff) for sensible heat from the desert-scrub/soil surface computed under this assumption increases sharply with increasing solar zenith angle, as the plants absorb a greater fraction of the incoming irradiation. The surface absorptivity (the coalbedo) also increases sharply with increasing solar zenith angle, and thus the sensible heat flux from such complex surfaces is a much broader function of time of day than when computed under constant k(g-eff) and constant albedo assumptions.
STOMP Subsurface Transport Over Multiple Phases: User`s guide
White, M.D.; Oostrom, M.
1997-10-01
The U.S. Department of Energy, through the Office of Technology Development, has requested the demonstration of remediation technologies for the cleanup of volatile organic compounds and associated radionuclides within the soil and groundwater at arid sites. This demonstration program, called the VOC-Arid Soils Integrated Demonstration Program (Arid-ID), has been initially directed at a volume of unsaturated and saturated soil contaminated with carbon tetrachloride, on the Hanford Site near Richland, Washington. A principal subtask of the Arid-ID program involves the development of an integrated engineering simulator for evaluating the effectiveness and efficiency of various remediation technologies. The engineering simulator`s intended users include scientists and engineers who are investigating soil physics phenomena associated with remediation technologies. Principal design goals for the engineer simulator include broad applicability, verified algorithms, quality assurance controls, and validated simulations against laboratory and field-scale experiments. An important goal for the simulator development subtask involves the ability to scale laboratory and field-scale experiments to full-scale remediation technologies, and to transfer acquired technology to other arid sites. The STOMP (Subsurface Transport Over Multiple Phases) simulator has been developed by the Pacific Northwest National Laboratory for modeling remediation technologies. Information on the use, application, and theoretical basis of the STOMP simulator theory and discussions on the governing equations, constitutive relations, and numerical solution algorithms for the STOMP simulator.
STOMP Subsurface Transport Over Multiple Phases: Application guide
Nichols, W.E.; Aimo, N.J.; Oostrom, M.; White, M.D.
1997-09-01
The U.S. Department of Energy (DOE), through the Office of Technology Development, has requested the demonstration of remediation technologies for the cleanup of volatile organic compounds and associated radionuclides within the soil and ground water at arid sites. This demonstration program, called the VOC-Arid Soils Integrated Demonstrated Program (Arid-ID), has been initially directed at a volume of unsaturated and saturated soil contaminated with carbon tetrachloride on the Hanford Site near Richland, Washington. A principal subtask of the Arid-ID program involves the development of an integrated engineering simulator for evaluating the effectiveness and efficiency of various remediation technologies. The engineering simulator`s intended users include scientists and engineers who are investigating soil physics phenomena associated with remediation technologies. Principal design goals for the engineering simulator include broad applicability, verified algorithms, quality assurance controls, and validated simulations against laboratory and field-scale experiments. An important goal for the simulator development subtask involves the ability to scale laboratory and field-scale experiments to full-scale remediation technologies, and to transfer acquired technology to other arid sites. The STOMP (Subsurface Transport Over Multiple Phases) simulator has been developed by the Pacific Northwest Laboratory for modeling remediation technologies. Information on the use, application, and theoretical basis of the STOMP simulator are documented in three companion guide guides. This document, the Application Guide, provides a suite of example applications of the STOMP simulator.
NASA Astrophysics Data System (ADS)
Kumar, U.; Milesi, C.; Nemani, R. R.; Raja, S. Kumar; Ganguly, S.; Wang, W.
2015-06-01
In this paper, we explore the possibility of sparse regression, a new direction in unmixing, for vegetation and urban area classification. SUnSAL (Sparse unmixing via variable splitting and augmented Lagrangian) in both unconstrained and constrained forms (with the abundance non-negativity and abundance sum-to-one constraints) were used with a set of global endmembers (substrate, vegetation and dark objects) to unmix a set of computer simulated noise-free and noisy data (with Gaussian noise of different signal-to-noise ratio) in order to judge the robustness of the algorithm. The error in the fractional estimate was examined for varying noise power (variance): 2, 4, 8, 16, 32, 64, 128 and 256. In the second set of experiments, a spectrally diverse collection of 11 scenes of Level 1 terrain corrected, cloud free Landsat-5 TM data representing an agricultural setup in Fresno, California, USA were used. The corresponding ground data for validation were collected on the same days of satellite overpass. Finally in the third set of experiments, a clear sky Landsat-5 TM data for an area near the Golden Gate Bridge, San Francisco (an urbanized landscape), California, USA were used to assess the algorithm. The fractional estimates of the 30 m Landsat-5 TM data were compared with the fractional estimates of a high-resolution World View-2 data (2 m spatial resolution) obtained using a fully constrained least squares algorithm. The results were evaluated using descriptive statistics, correlation coefficient, RMSE, probability of success and bivariate distribution function, which showed that constrained model was better than unconstrained form.
NASA Astrophysics Data System (ADS)
Trautz, A.; Illangasekare, T. H.; Tilton, N.
2015-12-01
Soil moisture is a fundamental state variable that provides the water necessary for plant growth and evapotranspiration. Soil moisture has been extensively studied in the context of bare surface soils and root zones. Less attention has focused on the effects of sparse vegetation distributions, such as those typical of agricultural cropland and other natural surface environments, on soil moisture dynamics. The current study explores root zone, bulk soil, and near-surface atmosphere interactions in terms of soil moisture under different distributions of sparse vegetation using multi-scale laboratory experimentation and numerical simulation. This research is driven by the need to advance our fundamental understanding of soil moisture dynamics in the context of improving water conservation and next generation heat and mass transfer numerical models. Experimentation is performed in a two-dimensional 7.3 m long intermediate scale soil tank interfaced with a climate-controlled wind tunnel, both of which are outfitted with current sensor technologies for measuring atmospheric and soil variables. The soil tank is packed so that a sparsely vegetated soil is surrounded by bulk bare soil; the two regions are separated by porous membranes to isolate the root zone from the bulk soil. Results show that in the absence of vegetation, evaporation rates vary along the soil tank in response to longitudinal changes in humidity; soil dries fastest upstream where evaporation rates are highest. In the presence of vegetation, soil moisture in the bulk soil closest to a vegetated region decreases more rapidly than the bulk soil farther away. Evapotranspiration rates in this region are also higher than the bulk soil region. This study is the first step towards the development of more generalized models that account for non-uniformly distributed vegetation and land surfaces exhibiting micro-topology.
STOMP Subsurface Transport Over Multiple Phases, Version 4.0, User’s Guide
White, Mark D.; Oostrom, Martinus
2006-06-09
This guide describes the general use, input file formatting, compilation and execution of the STOMP (Subsurface Transport Over Multiple Phases) simulator, a scientific tool for analyzing single and multiple phase subsurface flow and transport. A description of the simulator’s governing equations, constitutive functions and numerical solution algorithms are provided in a companion theory guide. In writing these guides for the STOMP simulator, the authors have assumed that the reader comprehends concepts and theories associated with multiple-phase hydrology, heat transfer, thermodynamics, radioactive chain decay, and relative permeability-saturation-capillary pressure constitutive relations. The authors further assume that the reader is familiar with the computing environment on which they plan to compile and execute the STOMP simulator. Source codes for the sequential versions of the simulator are available in pure FORTRAN 77 or mixed FORTRAN 77/90 forms. The pure FORTRAN 77 source code form requires a parameters file to define the memory requirements for the array elements. The mixed FORTRAN 77/90 form of the source code uses dynamic memory allocation to define memory requirements, based on a FORTRAN 90 preprocessor STEP, that reads the input files. The simulator utilizes a variable source code configuration, which allows the execution memory and speed to be tailored to the problem specifics, and essentially requires that the source code be assembled and compiled through a software maintenance utility. The memory requirements for executing the simulator are dependent on the complexity of physical system to be modeled and the size and dimensionality of the computational domain. Likewise execution speed depends on the problem complexity, size and dimensionality of the computational domain, and computer performance. Selected operational modes of the STOMP simulator are available for scalable execution on multiple processor (i.e., parallel) computers. These versions
STOMP: A Software Architecture for the Design and Simulation UAV-Based Sensor Networks
Jones, E D; Roberts, R S; Hsia, T C S
2002-10-28
This paper presents the Simulation, Tactical Operations and Mission Planning (STOMP) software architecture and framework for simulating, controlling and communicating with unmanned air vehicles (UAVs) servicing large distributed sensor networks. STOMP provides hardware-in-the-loop capability enabling real UAVs and sensors to feedback state information, route data and receive command and control requests while interacting with other real or virtual objects thereby enhancing support for simulation of dynamic and complex events.
STOMP-ECKEChem: An Engineering Perspective on Reactive Transport in Geologic Media
White, Mark D.; Fang, Yilin
2012-04-04
ECKEChem (Equilibrium, Conservation, Kinetic Equation Chemistry) is a reactive transport module for the STOMP suite of multifluid subsurface flow and transport simulators that was developed from an engineering perspective. STOMP comprises a suite of operational modes that are distinguished by the solved coupled conservation equations with capabilities for a variety of subsurface applications (e.g., environmental remediation and stewardship, geologic sequestration of greenhouse gases, gas hydrate production, and oil shale production). The ECKEChem module was designed to provide integrated reactive transport capabilities across the suite of STOMP simulator operational modes. The initial application for the ECKEChem module was in the simulation of the mineralization reactions that occurred with the injection of supercritical carbon dioxide into deep Columbia River basalt formations, where it was implemented in the STOMP-CO2 simulator. The STOMP-ECKEChem solution approach to modeling reactive transport in multifluid geologic media is founded on an engineering perspective: (1) sequential non-iterative coupling between the flow and reactive transport is sufficient, (2) reactive transport can be modeled by operator splitting with local geochemistry and global transport, (3) geochemistry can be expressed as a system of coupled nonlinear equilibrium, conservation and kinetic equations, (4) a limited number of kinetic equation forms are used in geochemical practice. This chapter describes the conceptual approach to converting a geochemical reaction network into a series of equilibrium, conservation and kinetic equations, the implementation of ECKEChem in STOMP, the numerical solution approach, and a demonstration of the simulator on a complex application involving desorption of uranium from contaminated field-textured sediments.
Lu, Zhiming; Meyer, D.J.
2002-01-01
We used interferometric methods on a pair of repeat-pass ERS-1 synthetic aperture radar (SAR) images to study soil moisture changes over sparsely vegetated targets. The intensity of the SAR image acquired at one time was higher than that of an image acquired at an earlier time. We used a correlation image computed from the SAR image pair to study the cause of the observed changes in SAR intensity. Because a reduction of correlation over areas with intensity changes was not observed, we interpreted the intensity changes as not being caused by changes in roughness/structure, but by a change in soil moisture owing to rainfall. An increase in soil moisture ranging from 5% to 20% is the most likely explanation for the increase of intensity. These analyses imply that both intensity and phase information should be used in SAR change detection applications.
NASA Astrophysics Data System (ADS)
Powell, R. L.; Goulden, M.; Peterson, S.; Roberts, D. A.; Still, C. J.
2015-12-01
Temperature is a primary environmental control on biological systems and processes at a range of spatial and temporal scales, from controlling biochemical processes such as photosynthesis to influencing continental-scale species distribution. The Landsat satellite series provides a long record (since the mid-1980s) of relatively high spatial resolution thermal infrared (TIR) imagery, from which we derive land surface temperature (LST) grids. Here, we investigate fine spatial resolution factors that influence Landsat-derived LST over a spectrally and spatially heterogeneous landscape. We focus on paired sites (inside/outside a 1994 fire scar) within a pinyon-juniper scrubland in Southern California. The sites have nearly identical micro-meteorology and vegetation species composition, but distinctly different vegetation abundance and structure. The tower at the unburned site includes a number of in-situ imaging tools to quantify vegetation properties, including a thermal camera on a pan-tilt mount, allowing hourly characterization of landscape component temperatures (e.g., sunlit canopy, bare soil, leaf litter). We use these in-situ measurements to assess the impact of fine-scale landscape heterogeneity on estimates of LST, including sensitivity to (i) the relative abundance of component materials, (ii) directional effects due to solar and viewing geometry, (iii) duration of sunlit exposure for each compositional type, and (iv) air temperature. To scale these properties to Landsat spatial resolution (~100-m), we characterize the sub-pixel composition of landscape components (in addition to shade) by applying spectral mixture analysis (SMA) to the Landsat Operational Land Imager (OLI) spectral bands and test the sensitivity of the relationships established with the in-situ data at this coarser scale. The effects of vegetation abundance and cover height versus other controls on satellite-derived estimates of LST will be assessed by comparing estimates at the burned vs
Moustafa, Gihan G.; Shaaban, F. E.; Hadeed, A. H. Abo; Elhady, Walaa M.
2016-01-01
Aim: The current study was directed to investigate the immunotoxic and oxidative stress effects of Roundup and Stomp herbicides and their combination on Nile catfish (Clarias gariepinus). Materials and Methods: The experiment was carried out on 120 fish that randomly divided into four equal groups with three replicates: The first group kept as control, the second group exposed to 1/2 96 h lethal concentration 50 (LC50) of Roundup, the third group exposed to 1/2 96 h LC50 of Stomp, and the fourth one exposed to a combination of Roundup and Stomp at previously-mentioned doses. The experiment was terminated after 15 days; blood samples were obtained at 1st, 8th, and 15th days of treatment where the sera were separated for estimation of antioxidant enzymes. Meanwhile, at 15th day of exposure part of blood was collected from all groups with an anticoagulant for evaluation of phagocytic activity, then the fish were sacrificed, and specimens from the liver of all groups were obtained for histopathological examination. Results: Our results indicated that both herbicides either individually or in combination elucidated significant decrease in phagocytic activity that was highly marked in group exposed to both herbicides. Furthermore, our data elicited an obvious elevation in the levels of superoxide dismutase (SOD), catalase (CAT), and glutathione peroxidase (GPx). Meanwhile, the data depicted reduction in levels of reduced glutathione (GSH) and glutathione-S-transferase (GST). Histopathological investigation of liver proved the aforementioned results. Conclusion: It could be concluded that either Roundup or Stomp alone cause significant deleterious effects on aquatic vertebrates. However, the use of their combination enhanced their toxic effects. Toxicity can end up in humans through the food chain. PMID:27397989
NASA Astrophysics Data System (ADS)
Peng, Z.; Zhou, J.; Li, M.; Liu, S.; Song, L.; Zhu, L.
2015-12-01
Validations of the remotely sensed land surface temperature (LST) are necessary due to the strong sensitivity of temperature as an input parameter for many models. LST products with large errors may lead to significant uncertainties in subsequent applications. Most of the current validation methods rely on direct LST comparisons between the ground measurements at a single site and its corresponding pixel. However, due to the scale-mismatch, it is inappropriate to use the single-site data to evaluate the whole pixel value, especially over sparsely vegetated pixels or other heterogeneous pixels. Based on the geometric optical model, a ground-based realistic three-dimensional scene model (RTDS) is developed by taking into account the sun-ground target-sensor geometry, to calculate the surface temperature for a mixed pixel. In this model, the ground-based realistic three-dimensional scene is reconstructed, and the surface of the scene is divided into little surface elements. The sun light is treated as parallel rays in order to calculate the Boolean value of each surface element observed by the remote sensor to judge shadow or illumination, and thus we get the fraction of each component of the scene using face-factor. The RTDS modeled LST has the same extent as the field-of-view of the remote sensor. Therefore, it is comparable with the estimates of remotely sensed LST. The measurements over a sparsely vegetated surface in the downstream area of Heihe River basin are used to test the RTDS model. The results show that the calculated LST with the RTDS model agree well with the ground measurements. When the size of element is less than 0.1 m2, the modeled surface temperatures have negligible differences. The influences of different distributions of ground scene on the results are analyzed and the results indicate that the distribution of ground scene has a great impact on the calculated surface temperature. The proposed RTDS model provides a bridge to solve the problem of
Using the STOMP (Seismic TOMography Program) Program for tomography with strong ray bending
Beatty, J.A.; Berryman, J.G.
1987-08-31
Accurate tomographic reconstructions of sound wave speed and attenuation are more difficult to obtain than are the corresponding reconstructions for x-rays or high frequency electromagnetic probes. The source of the difficulty is the common occurrence of large contrasts in acoustic or seismic wave speeds, leading to refraction and ray-bending effects. A new algorithm based on Fermat's principle has been developed to treat these problems. A description of the code STOMP (for Seismic TOMography Program) implementing the new algorithm is presented here together with a brief users manual for applications to borehole-to-borehole tomography. 3 refs., 3 figs.
A bloodstain pattern interpretation in a homicide case involving an apparent "stomping".
Ristenbatt, R R; Shaler, R C
1995-01-01
A New York City homicide case was investigated at the request of the District Attorney's office. The deceased had been violently beaten about the face, neck, and chest area. The bludgeoning left the victim unrecognizable, and produced blood-spatter patterns at the scene that attained heights over nine feet. The suspect claimed that he reacted in self-defense to the victim's attack with a table leg at which point he "knocked him down" and possibly "kicked him a few times." Our investigation was intended to determine whether the bloodspatter patterns observed at the crime scene were consistent with the statements made by the defendant. Conclusions were drawn from an analysis of the crime scene, autopsy photos and report, physical evidence submitted to the laboratory, and reconstruction experimentation performed at the Office of Chief Medical Examiner (OCME). The spatter patterns observed at the scene were found to be consistent with those that would be produced from a "stomping" incident. PMID:7876798
A bloodstain pattern interpretation in a homicide case involving an apparent "stomping".
Ristenbatt, R R; Shaler, R C
1995-01-01
A New York City homicide case was investigated at the request of the District Attorney's office. The deceased had been violently beaten about the face, neck, and chest area. The bludgeoning left the victim unrecognizable, and produced blood-spatter patterns at the scene that attained heights over nine feet. The suspect claimed that he reacted in self-defense to the victim's attack with a table leg at which point he "knocked him down" and possibly "kicked him a few times." Our investigation was intended to determine whether the bloodspatter patterns observed at the crime scene were consistent with the statements made by the defendant. Conclusions were drawn from an analysis of the crime scene, autopsy photos and report, physical evidence submitted to the laboratory, and reconstruction experimentation performed at the Office of Chief Medical Examiner (OCME). The spatter patterns observed at the scene were found to be consistent with those that would be produced from a "stomping" incident.
Numerical Simulations of Urea Hydrolysis and Calcite Precipitation in Porous Media Using STOMP
Luanjing Guo; Hai Huang; Bill X. Hu
2010-11-01
Subsurface radionuclide and trace metal contaminants throughout the U.S. Department of Energy (DOE) complex pose one of DOE’s greatest challenges for long-term stewardship. One promising in situ immobilization approach of these contaminants is engineered mineral (co)precipitation of calcite driven by urea hydrolysis that is catalyzed by enzyme urease. The tight nonlinear coupling among flow, transport, reaction and reaction-induced property changes of media of this approach was studied by reactive transport simulations with systematically increasing level of complexities of reaction network and physical/chemical heterogeneities using a numerical simulator named STOMP. Sensitivity studies on the reaction rates of both urea hydrolysis and calcite precipitation are performed via controlling urease enzyme concentration and precipitation rate constant according to the rate models employed. We have found that the rate of ureolysis is a dominating factor in the amount of precipitated mineral; however, the spatial distribution of the precipitates depends on both rates of ureolysis and calcite precipitation. A maximum 5% reduction in the porosity was observed within the simulation time period of 6 pore volumes in our 1-dimensional (1D) column simulations. When a low permeability inclusion is considered in the 2D simulations, the altered flow fields redistribute mineral forming constituents, leading to a distorted precipitation reaction front. The simulations also indicate that mineral precipitation occurs along the boundary of the low permeability zone, which implies that contaminants in the low permeability zone could be encapsulated and isolated from the flow paths.
Sparse Regression as a Sparse Eigenvalue Problem
NASA Technical Reports Server (NTRS)
Moghaddam, Baback; Gruber, Amit; Weiss, Yair; Avidan, Shai
2008-01-01
We extend the l0-norm "subspectral" algorithms for sparse-LDA [5] and sparse-PCA [6] to general quadratic costs such as MSE in linear (kernel) regression. The resulting "Sparse Least Squares" (SLS) problem is also NP-hard, by way of its equivalence to a rank-1 sparse eigenvalue problem (e.g., binary sparse-LDA [7]). Specifically, for a general quadratic cost we use a highly-efficient technique for direct eigenvalue computation using partitioned matrix inverses which leads to dramatic x103 speed-ups over standard eigenvalue decomposition. This increased efficiency mitigates the O(n4) scaling behaviour that up to now has limited the previous algorithms' utility for high-dimensional learning problems. Moreover, the new computation prioritizes the role of the less-myopic backward elimination stage which becomes more efficient than forward selection. Similarly, branch-and-bound search for Exact Sparse Least Squares (ESLS) also benefits from partitioned matrix inverse techniques. Our Greedy Sparse Least Squares (GSLS) generalizes Natarajan's algorithm [9] also known as Order-Recursive Matching Pursuit (ORMP). Specifically, the forward half of GSLS is exactly equivalent to ORMP but more efficient. By including the backward pass, which only doubles the computation, we can achieve lower MSE than ORMP. Experimental comparisons to the state-of-the-art LARS algorithm [3] show forward-GSLS is faster, more accurate and more flexible in terms of choice of regularization
A Modified Sparse Representation Method for Facial Expression Recognition.
Wang, Wei; Xu, LiHong
2016-01-01
In this paper, we carry on research on a facial expression recognition method, which is based on modified sparse representation recognition (MSRR) method. On the first stage, we use Haar-like+LPP to extract feature and reduce dimension. On the second stage, we adopt LC-K-SVD (Label Consistent K-SVD) method to train the dictionary, instead of adopting directly the dictionary from samples, and add block dictionary training into the training process. On the third stage, stOMP (stagewise orthogonal matching pursuit) method is used to speed up the convergence of OMP (orthogonal matching pursuit). Besides, a dynamic regularization factor is added to iteration process to suppress noises and enhance accuracy. We verify the proposed method from the aspect of training samples, dimension, feature extraction and dimension reduction methods and noises in self-built database and Japan's JAFFE and CMU's CK database. Further, we compare this sparse method with classic SVM and RVM and analyze the recognition effect and time efficiency. The result of simulation experiment has shown that the coefficient of MSRR method contains classifying information, which is capable of improving the computing speed and achieving a satisfying recognition result. PMID:26880878
Han, Chun-Mei; Ran, Juan; Zhang, Hui; Li, Fa-Sheng; Li, Yan; Gu, Qing-Bao
2012-10-01
Frequently accidental spill of hazard materials into soil environment posed significant threats to human health and natural environment in China. In this paper, simulated nonaqueous phase liquid (NAPLs) toluene sudden spill in fluvo-aquic soil was performed in a two-dimensional tank to investigate the migration and distribution characteristics of toluene as well as its simulation through STOMP model. Visual observation showed that the horizontal expansion of toluene with concentration > 1 g x kg(-1) and > 20 g x kg(-1) were approximate 2.3 and 3 times the length of the corresponding vertical transportation, respectively. The result revealed that toluene exerted a preferential tendency to lateral spread compared to the vertical migration trends, which may contribute to the impeding effect as a result of the low permeable capacity of soil (permeability coefficient was 0.12 cm x h(-1)). The behavior and fate of toluene in heterogeneous soil layers (combined fluvo-aquic-sand and sand-fluvo-aquic soil layers, respectively) after spill were simulated using STOMP model and the results indicated that the significant difference in relative permeability of interface layer, due to the much higher value of permeability coefficient of sand (29.70 cm x h(-1)) than that of fluvo-aquic soil, played an important role on the redistribution of toluene after its spill into the heterogeneous soil layers. For practical purposes, the results of this study may be beneficial to identify the distribution property of NAPLs in time after its release into the soil environment.
Ray, J.; Lee, J.; Yadav, V.; Lefantzi, S.; Michalak, A. M.; van Bloemen Waanders, B.
2014-08-20
We present a sparse reconstruction scheme that can also be used to ensure non-negativity when fitting wavelet-based random field models to limited observations in non-rectangular geometries. The method is relevant when multiresolution fields are estimated using linear inverse problems. Examples include the estimation of emission fields for many anthropogenic pollutants using atmospheric inversion or hydraulic conductivity in aquifers from flow measurements. The scheme is based on three new developments. Firstly, we extend an existing sparse reconstruction method, Stagewise Orthogonal Matching Pursuit (StOMP), to incorporate prior information on the target field. Secondly, we develop an iterative method that uses StOMP tomore » impose non-negativity on the estimated field. Finally, we devise a method, based on compressive sensing, to limit the estimated field within an irregularly shaped domain. We demonstrate the method on the estimation of fossil-fuel CO2 (ffCO2) emissions in the lower 48 states of the US. The application uses a recently developed multiresolution random field model and synthetic observations of ffCO2 concentrations from a limited set of measurement sites. We find that our method for limiting the estimated field within an irregularly shaped region is about a factor of 10 faster than conventional approaches. It also reduces the overall computational cost by a factor of two. Further, the sparse reconstruction scheme imposes non-negativity without introducing strong nonlinearities, such as those introduced by employing log-transformed fields, and thus reaps the benefits of simplicity and computational speed that are characteristic of linear inverse problems.« less
Sparse pseudospectral approximation method
NASA Astrophysics Data System (ADS)
Constantine, Paul G.; Eldred, Michael S.; Phipps, Eric T.
2012-07-01
Multivariate global polynomial approximations - such as polynomial chaos or stochastic collocation methods - are now in widespread use for sensitivity analysis and uncertainty quantification. The pseudospectral variety of these methods uses a numerical integration rule to approximate the Fourier-type coefficients of a truncated expansion in orthogonal polynomials. For problems in more than two or three dimensions, a sparse grid numerical integration rule offers accuracy with a smaller node set compared to tensor product approximation. However, when using a sparse rule to approximately integrate these coefficients, one often finds unacceptable errors in the coefficients associated with higher degree polynomials. By reexamining Smolyak's algorithm and exploiting the connections between interpolation and projection in tensor product spaces, we construct a sparse pseudospectral approximation method that accurately reproduces the coefficients of basis functions that naturally correspond to the sparse grid integration rule. The compelling numerical results show that this is the proper way to use sparse grid integration rules for pseudospectral approximation.
sparse-msrf:A package for sparse modeling and estimation of fossil-fuel CO2 emission fields
2014-10-06
The software is used to fit models of emission fields (e.g., fossil-fuel CO2 emissions) to sparse measurements of gaseous concentrations. Its primary aim is to provide an implementation and a demonstration for the algorithms and models developed in J. Ray, V. Yadav, A. M. Michalak, B. van Bloemen Waanders and S. A. McKenna, "A multiresolution spatial parameterization for the estimation of fossil-fuel carbon dioxide emissions via atmospheric inversions", accepted, Geoscientific Model Development, 2014. The software can be used to estimate emissions of non-reactive gases such as fossil-fuel CO2, methane etc. The software uses a proxy of the emission field being estimated (e.g., for fossil-fuel CO2, a population density map is a good proxy) to construct a wavelet model for the emission field. It then uses a shrinkage regression algorithm called Stagewise Orthogonal Matching Pursuit (StOMP) to fit the wavelet model to concentration measurements, using an atmospheric transport model to relate emission and concentration fields. Algorithmic novelties described in the paper above (1) ensure that the estimated emission fields are non-negative, (2) allow the use of guesses for emission fields to accelerate the estimation processes and (3) ensure that under/overestimates in the guesses do not skew the estimation.
sparse-msrf:A package for sparse modeling and estimation of fossil-fuel CO2 emission fields
2014-10-06
The software is used to fit models of emission fields (e.g., fossil-fuel CO2 emissions) to sparse measurements of gaseous concentrations. Its primary aim is to provide an implementation and a demonstration for the algorithms and models developed in J. Ray, V. Yadav, A. M. Michalak, B. van Bloemen Waanders and S. A. McKenna, "A multiresolution spatial parameterization for the estimation of fossil-fuel carbon dioxide emissions via atmospheric inversions", accepted, Geoscientific Model Development, 2014. The softwaremore » can be used to estimate emissions of non-reactive gases such as fossil-fuel CO2, methane etc. The software uses a proxy of the emission field being estimated (e.g., for fossil-fuel CO2, a population density map is a good proxy) to construct a wavelet model for the emission field. It then uses a shrinkage regression algorithm called Stagewise Orthogonal Matching Pursuit (StOMP) to fit the wavelet model to concentration measurements, using an atmospheric transport model to relate emission and concentration fields. Algorithmic novelties described in the paper above (1) ensure that the estimated emission fields are non-negative, (2) allow the use of guesses for emission fields to accelerate the estimation processes and (3) ensure that under/overestimates in the guesses do not skew the estimation.« less
Sparse distributed memory overview
NASA Technical Reports Server (NTRS)
Raugh, Mike
1990-01-01
The Sparse Distributed Memory (SDM) project is investigating the theory and applications of massively parallel computing architecture, called sparse distributed memory, that will support the storage and retrieval of sensory and motor patterns characteristic of autonomous systems. The immediate objectives of the project are centered in studies of the memory itself and in the use of the memory to solve problems in speech, vision, and robotics. Investigation of methods for encoding sensory data is an important part of the research. Examples of NASA missions that may benefit from this work are Space Station, planetary rovers, and solar exploration. Sparse distributed memory offers promising technology for systems that must learn through experience and be capable of adapting to new circumstances, and for operating any large complex system requiring automatic monitoring and control. Sparse distributed memory is a massively parallel architecture motivated by efforts to understand how the human brain works. Sparse distributed memory is an associative memory, able to retrieve information from cues that only partially match patterns stored in the memory. It is able to store long temporal sequences derived from the behavior of a complex system, such as progressive records of the system's sensory data and correlated records of the system's motor controls.
Ray, J.; Lee, J.; Yadav, V.; Lefantzi, S.; Michalak, A. M.; van Bloemen Waanders, B.
2015-04-29
Atmospheric inversions are frequently used to estimate fluxes of atmospheric greenhouse gases (e.g., biospheric CO2 flux fields) at Earth's surface. These inversions typically assume that flux departures from a prior model are spatially smoothly varying, which are then modeled using a multi-variate Gaussian. When the field being estimated is spatially rough, multi-variate Gaussian models are difficult to construct and a wavelet-based field model may be more suitable. Unfortunately, such models are very high dimensional and are most conveniently used when the estimation method can simultaneously perform data-driven model simplification (removal of model parameters that cannot be reliably estimated) and fitting.more » Such sparse reconstruction methods are typically not used in atmospheric inversions. In this work, we devise a sparse reconstruction method, and illustrate it in an idealized atmospheric inversion problem for the estimation of fossil fuel CO2 (ffCO2) emissions in the lower 48 states of the USA. Our new method is based on stagewise orthogonal matching pursuit (StOMP), a method used to reconstruct compressively sensed images. Our adaptations bestow three properties to the sparse reconstruction procedure which are useful in atmospheric inversions. We have modified StOMP to incorporate prior information on the emission field being estimated and to enforce non-negativity on the estimated field. Finally, though based on wavelets, our method allows for the estimation of fields in non-rectangular geometries, e.g., emission fields inside geographical and political boundaries. Our idealized inversions use a recently developed multi-resolution (i.e., wavelet-based) random field model developed for ffCO2 emissions and synthetic observations of ffCO2 concentrations from a limited set of measurement sites. We find that our method for limiting the estimated field within an irregularly shaped region is about a factor of 10 faster than conventional approaches. It also
LOFAR sparse image reconstruction
NASA Astrophysics Data System (ADS)
Garsden, H.; Girard, J. N.; Starck, J. L.; Corbel, S.; Tasse, C.; Woiselle, A.; McKean, J. P.; van Amesfoort, A. S.; Anderson, J.; Avruch, I. M.; Beck, R.; Bentum, M. J.; Best, P.; Breitling, F.; Broderick, J.; Brüggen, M.; Butcher, H. R.; Ciardi, B.; de Gasperin, F.; de Geus, E.; de Vos, M.; Duscha, S.; Eislöffel, J.; Engels, D.; Falcke, H.; Fallows, R. A.; Fender, R.; Ferrari, C.; Frieswijk, W.; Garrett, M. A.; Grießmeier, J.; Gunst, A. W.; Hassall, T. E.; Heald, G.; Hoeft, M.; Hörandel, J.; van der Horst, A.; Juette, E.; Karastergiou, A.; Kondratiev, V. I.; Kramer, M.; Kuniyoshi, M.; Kuper, G.; Mann, G.; Markoff, S.; McFadden, R.; McKay-Bukowski, D.; Mulcahy, D. D.; Munk, H.; Norden, M. J.; Orru, E.; Paas, H.; Pandey-Pommier, M.; Pandey, V. N.; Pietka, G.; Pizzo, R.; Polatidis, A. G.; Renting, A.; Röttgering, H.; Rowlinson, A.; Schwarz, D.; Sluman, J.; Smirnov, O.; Stappers, B. W.; Steinmetz, M.; Stewart, A.; Swinbank, J.; Tagger, M.; Tang, Y.; Tasse, C.; Thoudam, S.; Toribio, C.; Vermeulen, R.; Vocks, C.; van Weeren, R. J.; Wijnholds, S. J.; Wise, M. W.; Wucknitz, O.; Yatawatta, S.; Zarka, P.; Zensus, A.
2015-03-01
Context. The LOw Frequency ARray (LOFAR) radio telescope is a giant digital phased array interferometer with multiple antennas distributed in Europe. It provides discrete sets of Fourier components of the sky brightness. Recovering the original brightness distribution with aperture synthesis forms an inverse problem that can be solved by various deconvolution and minimization methods. Aims: Recent papers have established a clear link between the discrete nature of radio interferometry measurement and the "compressed sensing" (CS) theory, which supports sparse reconstruction methods to form an image from the measured visibilities. Empowered by proximal theory, CS offers a sound framework for efficient global minimization and sparse data representation using fast algorithms. Combined with instrumental direction-dependent effects (DDE) in the scope of a real instrument, we developed and validated a new method based on this framework. Methods: We implemented a sparse reconstruction method in the standard LOFAR imaging tool and compared the photometric and resolution performance of this new imager with that of CLEAN-based methods (CLEAN and MS-CLEAN) with simulated and real LOFAR data. Results: We show that i) sparse reconstruction performs as well as CLEAN in recovering the flux of point sources; ii) performs much better on extended objects (the root mean square error is reduced by a factor of up to 10); and iii) provides a solution with an effective angular resolution 2-3 times better than the CLEAN images. Conclusions: Sparse recovery gives a correct photometry on high dynamic and wide-field images and improved realistic structures of extended sources (of simulated and real LOFAR datasets). This sparse reconstruction method is compatible with modern interferometric imagers that handle DDE corrections (A- and W-projections) required for current and future instruments such as LOFAR and SKA.
Structured Multifrontal Sparse Solver
2014-05-01
StruMF is an algebraic structured preconditioner for the interative solution of large sparse linear systems. The preconditioner corresponds to a multifrontal variant of sparse LU factorization in which some dense blocks of the factors are approximated with low-rank matrices. It is algebraic in that it only requires the linear system itself, and the approximation threshold that determines the accuracy of individual low-rank approximations. Favourable rank properties are obtained using a block partitioning which is amore » refinement of the partitioning induced by nested dissection ordering.« less
Sparse inpainting and isotropy
Feeney, Stephen M.; McEwen, Jason D.; Peiris, Hiranya V.; Marinucci, Domenico; Cammarota, Valentina; Wandelt, Benjamin D. E-mail: marinucc@axp.mat.uniroma2.it E-mail: h.peiris@ucl.ac.uk E-mail: cammarot@axp.mat.uniroma2.it
2014-01-01
Sparse inpainting techniques are gaining in popularity as a tool for cosmological data analysis, in particular for handling data which present masked regions and missing observations. We investigate here the relationship between sparse inpainting techniques using the spherical harmonic basis as a dictionary and the isotropy properties of cosmological maps, as for instance those arising from cosmic microwave background (CMB) experiments. In particular, we investigate the possibility that inpainted maps may exhibit anisotropies in the behaviour of higher-order angular polyspectra. We provide analytic computations and simulations of inpainted maps for a Gaussian isotropic model of CMB data, suggesting that the resulting angular trispectrum may exhibit small but non-negligible deviations from isotropy.
NASA Technical Reports Server (NTRS)
Kanerva, Pentti
1988-01-01
Theoretical models of the human brain and proposed neural-network computers are developed analytically. Chapters are devoted to the mathematical foundations, background material from computer science, the theory of idealized neurons, neurons as address decoders, and the search of memory for the best match. Consideration is given to sparse memory, distributed storage, the storage and retrieval of sequences, the construction of distributed memory, and the organization of an autonomous learning system.
Kanerva, P.
1988-01-01
Theoretical models of the human brain and proposed neural-network computers are developed analytically. Chapters are devoted to the mathematical foundations, background material from computer science, the theory of idealized neurons, neurons as address decoders, and the search of memory for the best match. Consideration is given to sparse memory, distributed storage, the storage and retrieval of sequences, the construction of distributed memory, and the organization of an autonomous learning system. 63 refs.
Sparse matrix test collections
Duff, I.
1996-12-31
This workshop will discuss plans for coordinating and developing sets of test matrices for the comparison and testing of sparse linear algebra software. We will talk of plans for the next release (Release 2) of the Harwell-Boeing Collection and recent work on improving the accessibility of this Collection and others through the World Wide Web. There will only be three talks of about 15 to 20 minutes followed by a discussion from the floor.
and Drayton Munster, Miroslav Stoyanov
2013-09-20
Sparse Grids are the family of methods of choice for multidimensional integration and interpolation in low to moderate number of dimensions. The method is to select extend a one dimensional set of abscissas, weights and basis functions by taking a subset of all possible tensor products. The module provides the ability to create global and local approximations based on polynomials and wavelets. The software has three components, a library, a wrapper for the library that provides a command line interface via text files ad a MATLAB interface via the command line tool.
2007-04-12
The Sparse Image Format (SIF) is a file format for storing spare raster images. It works by breaking an image down into tiles. Space is savid by only storing non-uniform tiles, i.e. tiles with at least two different pixel values. If a tile is completely uniform, its common pixel value is stored instead of the complete tile raster. The software is a library in the C language used for manipulating files in SIF format. Itmore » supports large files (> 2GB) and is designed to build in Windows and Linux environments.« less
2013-09-20
Sparse Grids are the family of methods of choice for multidimensional integration and interpolation in low to moderate number of dimensions. The method is to select extend a one dimensional set of abscissas, weights and basis functions by taking a subset of all possible tensor products. The module provides the ability to create global and local approximations based on polynomials and wavelets. The software has three components, a library, a wrapper for the library thatmore » provides a command line interface via text files ad a MATLAB interface via the command line tool.« less
Eads, Damian Ryan
2007-04-12
The Sparse Image Format (SIF) is a file format for storing spare raster images. It works by breaking an image down into tiles. Space is savid by only storing non-uniform tiles, i.e. tiles with at least two different pixel values. If a tile is completely uniform, its common pixel value is stored instead of the complete tile raster. The software is a library in the C language used for manipulating files in SIF format. It supports large files (> 2GB) and is designed to build in Windows and Linux environments.
White, Mark D.; McGrail, B. Peter
2005-12-01
flow and transport simulator, STOMP (Subsurface Transport Over Multiple Phases). Prior to these code development activities, the STOMP simulator included sequential and scalable implementations for numerically simulating the injection of supercritical CO2 into deep saline aquifers. Additionally, the sequential implementations included operational modes that considered nonisothermal conditions and kinetic dissolution of CO2 into the saline aqueous phase. This addendum documents the advancement of these numerical simulation capabilities to include reactive transport in the STOMP simulator through the inclusion of the recently PNNL developed batch geochemistry solution module ECKEChem (Equilibrium-Conservation-Kinetic Equation Chemistry). Potential geologic reservoirs for sequestering CO2 include deep saline aquifers, hydrate-bearing formations, depleted or partially depleted natural gas and petroleum reservoirs, and coal beds. The mechanisms for sequestering carbon dioxide in geologic reservoirs include physical trapping, dissolution in the reservoir fluids, hydraulic trapping (hysteretic entrapment of nonwetting fluids), and chemical reaction. This document and the associated code development and verification work are concerned with the chemistry of injecting CO2 into geologic reservoirs. As geologic sequestration of CO2 via chemical reaction, namely precipitation reactions, are most dominate in deep saline aquifers, the principal focus of this document is the numerical simulation of CO2 injection, migration, and geochemical reaction in deep saline aquifers. The ECKEChem batch chemistry module was developed in a fashion that would allow its implementation into all operational modes of the STOMP simulator, making it a more versatile chemistry component. Additionally, this approach allows for verification of the ECKEChem module against more classical reactive transport problems involving aqueous systems.
NASA Technical Reports Server (NTRS)
Denning, Peter J.
1989-01-01
Sparse distributed memory was proposed be Pentti Kanerva as a realizable architecture that could store large patterns and retrieve them based on partial matches with patterns representing current sensory inputs. This memory exhibits behaviors, both in theory and in experiment, that resemble those previously unapproached by machines - e.g., rapid recognition of faces or odors, discovery of new connections between seemingly unrelated ideas, continuation of a sequence of events when given a cue from the middle, knowing that one doesn't know, or getting stuck with an answer on the tip of one's tongue. These behaviors are now within reach of machines that can be incorporated into the computing systems of robots capable of seeing, talking, and manipulating. Kanerva's theory is a break with the Western rationalistic tradition, allowing a new interpretation of learning and cognition that respects biology and the mysteries of individual human beings.
Zhang, Lihe; Lu, Huchuan; Du, Dandan; Liu, Luning
2016-02-01
In this paper, we propose a novel tracking framework based on a sparse and discriminative hashing method. Different from the previous work, we treat object tracking as an approximate nearest neighbor searching process in a binary space. Using the hash functions, the target templates and the candidates can be projected into the Hamming space, facilitating the distance calculation and tracking efficiency. First, we integrate both the inter-class and intra-class information to train multiple hash functions for better classification, while most classifiers in previous tracking methods usually neglect the inter-class correlation, which may cause the inaccuracy. Then, we introduce sparsity into the hash coefficient vectors for dynamic feature selection, which is crucial to select the discriminative and stable features to adapt to visual variations during the tracking process. Extensive experiments on various challenging sequences show that the proposed algorithm performs favorably against the state-of-the-art methods.
Estimating sparse precision matrices
NASA Astrophysics Data System (ADS)
Padmanabhan, Nikhil; White, Martin; Zhou, Harrison H.; O'Connell, Ross
2016-08-01
We apply a method recently introduced to the statistical literature to directly estimate the precision matrix from an ensemble of samples drawn from a corresponding Gaussian distribution. Motivated by the observation that cosmological precision matrices are often approximately sparse, the method allows one to exploit this sparsity of the precision matrix to more quickly converge to an asymptotic 1/sqrt{N_sim} rate while simultaneously providing an error model for all of the terms. Such an estimate can be used as the starting point for further regularization efforts which can improve upon the 1/sqrt{N_sim} limit above, and incorporating such additional steps is straightforward within this framework. We demonstrate the technique with toy models and with an example motivated by large-scale structure two-point analysis, showing significant improvements in the rate of convergence. For the large-scale structure example, we find errors on the precision matrix which are factors of 5 smaller than for the sample precision matrix for thousands of simulations or, alternatively, convergence to the same error level with more than an order of magnitude fewer simulations.
NASA Astrophysics Data System (ADS)
Hernandez, Svea
2012-10-01
CTE measurements are made using the "internal sparse field test", along the parallelaxis. The "POS=" optional parameter, introduced during cycle 11, is used to provideoff-center MSM positioning of some slits. All exposures are internals.
NASA Astrophysics Data System (ADS)
Wolfe, Michael
2011-10-01
CTE measurements are made using the "internal sparse field test", along the parallelaxis. The "POS=" optional parameter, introduced during cycle 11, is used to provideoff-center MSM positionings of some slits. All exposures are internals.
NASA Astrophysics Data System (ADS)
Wolfe, Michael
2010-09-01
CTE measurements are made using the "internal sparse field test", along the parallelaxis. The "POS=" optional parameter, introduced during cycle 11, is used to provideoff-center MSM positionings of some slits. All exposures are internals.
NASA Astrophysics Data System (ADS)
Hernandez, Svea
2013-10-01
CTE measurements are made using the "internal sparse field test", along the parallelaxis. The "POS=" optional parameter, introduced during cycle 11, is used to provideoff-center MSM positioning of some slits. All exposures are internals.
NASA Astrophysics Data System (ADS)
Wolfe, Michael
2009-07-01
CTE measurements are made using the "internal sparse field test", along the parallelaxis. The "POS=" optional parameter, introduced during cycle 11, is used to provideoff-center MSM positionings of some slits. All exposures are internals.
Threaded Operations on Sparse Matrices
Sneed, Brett
2015-09-01
We investigate the use of sparse matrices and OpenMP multi-threading on linear algebra operations involving them. Several sparse matrix data structures are presented. Implementation of the multi- threading primarily occurs in the level one and two BLAS functions used within the four algorithms investigated{the Power Method, Conjugate Gradient, Biconjugate Gradient, and Jacobi's Method. The bene ts of launching threads once per high level algorithm are explored.
Sparse Methods for Biomedical Data
Ye, Jieping; Liu, Jun
2013-01-01
Following recent technological revolutions, the investigation of massive biomedical data with growing scale, diversity, and complexity has taken a center stage in modern data analysis. Although complex, the underlying representations of many biomedical data are often sparse. For example, for a certain disease such as leukemia, even though humans have tens of thousands of genes, only a few genes are relevant to the disease; a gene network is sparse since a regulatory pathway involves only a small number of genes; many biomedical signals are sparse or compressible in the sense that they have concise representations when expressed in a proper basis. Therefore, finding sparse representations is fundamentally important for scientific discovery. Sparse methods based on the ℓ1 norm have attracted a great amount of research efforts in the past decade due to its sparsity-inducing property, convenient convexity, and strong theoretical guarantees. They have achieved great success in various applications such as biomarker selection, biological network construction, and magnetic resonance imaging. In this paper, we review state-of-the-art sparse methods and their applications to biomedical data. PMID:24076585
Wavelet Sparse Approximate Inverse Preconditioners
NASA Technical Reports Server (NTRS)
Chan, Tony F.; Tang, W.-P.; Wan, W. L.
1996-01-01
There is an increasing interest in using sparse approximate inverses as preconditioners for Krylov subspace iterative methods. Recent studies of Grote and Huckle and Chow and Saad also show that sparse approximate inverse preconditioner can be effective for a variety of matrices, e.g. Harwell-Boeing collections. Nonetheless a drawback is that it requires rapid decay of the inverse entries so that sparse approximate inverse is possible. However, for the class of matrices that, come from elliptic PDE problems, this assumption may not necessarily hold. Our main idea is to look for a basis, other than the standard one, such that a sparse representation of the inverse is feasible. A crucial observation is that the kind of matrices we are interested in typically have a piecewise smooth inverse. We exploit this fact, by applying wavelet techniques to construct a better sparse approximate inverse in the wavelet basis. We shall justify theoretically and numerically that our approach is effective for matrices with smooth inverse. We emphasize that in this paper we have only presented the idea of wavelet approximate inverses and demonstrated its potential but have not yet developed a highly refined and efficient algorithm.
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
1997-07-01
CTE measurements are made using the "sparse field test", along both the serial and parallel axes. This program needs special commanding to provide {a} off-center MSM positionings of some slits, and {b} the ability to read out with any amplifier {A, B, C, or D}. All exposures are internals.
Amesos2 Templated Direct Sparse Solver Package
2011-05-24
Amesos2 is a templated direct sparse solver package. Amesos2 provides interfaces to direct sparse solvers, rather than providing native solver capabilities. Amesos2 is a derivative work of the Trilinos package Amesos.
Finding communities in sparse networks
NASA Astrophysics Data System (ADS)
Singh, Abhinav; Humphries, Mark D.
2015-03-01
Spectral algorithms based on matrix representations of networks are often used to detect communities, but classic spectral methods based on the adjacency matrix and its variants fail in sparse networks. New spectral methods based on non-backtracking random walks have recently been introduced that successfully detect communities in many sparse networks. However, the spectrum of non-backtracking random walks ignores hanging trees in networks that can contain information about their community structure. We introduce the reluctant backtracking operators that explicitly account for hanging trees as they admit a small probability of returning to the immediately previous node, unlike the non-backtracking operators that forbid an immediate return. We show that the reluctant backtracking operators can detect communities in certain sparse networks where the non-backtracking operators cannot, while performing comparably on benchmark stochastic block model networks and real world networks. We also show that the spectrum of the reluctant backtracking operator approximately optimises the standard modularity function. Interestingly, for this family of non- and reluctant-backtracking operators the main determinant of performance on real-world networks is whether or not they are normalised to conserve probability at each node.
Highly parallel sparse Cholesky factorization
NASA Technical Reports Server (NTRS)
Gilbert, John R.; Schreiber, Robert
1990-01-01
Several fine grained parallel algorithms were developed and compared to compute the Cholesky factorization of a sparse matrix. The experimental implementations are on the Connection Machine, a distributed memory SIMD machine whose programming model conceptually supplies one processor per data element. In contrast to special purpose algorithms in which the matrix structure conforms to the connection structure of the machine, the focus is on matrices with arbitrary sparsity structure. The most promising algorithm is one whose inner loop performs several dense factorizations simultaneously on a 2-D grid of processors. Virtually any massively parallel dense factorization algorithm can be used as the key subroutine. The sparse code attains execution rates comparable to those of the dense subroutine. Although at present architectural limitations prevent the dense factorization from realizing its potential efficiency, it is concluded that a regular data parallel architecture can be used efficiently to solve arbitrarily structured sparse problems. A performance model is also presented and it is used to analyze the algorithms.
Sparse Matrices in MATLAB: Design and Implementation
NASA Technical Reports Server (NTRS)
Gilbert, John R.; Moler, Cleve; Schreiber, Robert
1992-01-01
The matrix computation language and environment MATLAB is extended to include sparse matrix storage and operations. The only change to the outward appearance of the MATLAB language is a pair of commands to create full or sparse matrices. Nearly all the operations of MATLAB now apply equally to full or sparse matrices, without any explicit action by the user. The sparse data structure represents a matrix in space proportional to the number of nonzero entries, and most of the operations compute sparse results in time proportional to the number of arithmetic operations on nonzeros.
Robust Sparse Blind Source Separation
NASA Astrophysics Data System (ADS)
Chenot, Cecile; Bobin, Jerome; Rapin, Jeremy
2015-11-01
Blind Source Separation is a widely used technique to analyze multichannel data. In many real-world applications, its results can be significantly hampered by the presence of unknown outliers. In this paper, a novel algorithm coined rGMCA (robust Generalized Morphological Component Analysis) is introduced to retrieve sparse sources in the presence of outliers. It explicitly estimates the sources, the mixing matrix, and the outliers. It also takes advantage of the estimation of the outliers to further implement a weighting scheme, which provides a highly robust separation procedure. Numerical experiments demonstrate the efficiency of rGMCA to estimate the mixing matrix in comparison with standard BSS techniques.
Image fusion using sparse overcomplete feature dictionaries
Brumby, Steven P.; Bettencourt, Luis; Kenyon, Garrett T.; Chartrand, Rick; Wohlberg, Brendt
2015-10-06
Approaches for deciding what individuals in a population of visual system "neurons" are looking for using sparse overcomplete feature dictionaries are provided. A sparse overcomplete feature dictionary may be learned for an image dataset and a local sparse representation of the image dataset may be built using the learned feature dictionary. A local maximum pooling operation may be applied on the local sparse representation to produce a translation-tolerant representation of the image dataset. An object may then be classified and/or clustered within the translation-tolerant representation of the image dataset using a supervised classification algorithm and/or an unsupervised clustering algorithm.
Sparse representation for the ISAR image reconstruction
NASA Astrophysics Data System (ADS)
Hu, Mengqi; Montalbo, John; Li, Shuxia; Sun, Ligang; Qiao, Zhijun G.
2016-05-01
In this paper, a sparse representation of the data for an inverse synthetic aperture radar (ISAR) system is provided in two dimensions. The proposed sparse representation motivates the use a of a Convex Optimization that recovers the image with far less samples, which is required by Nyquist-Shannon sampling theorem to increases the efficiency and decrease the cost of calculation in radar imaging.
Sparse Modeling for Astronomical Data Analysis
NASA Astrophysics Data System (ADS)
Ikeda, Shiro; Odaka, Hirokazu; Uemura, Makoto
2016-03-01
For astronomical data analysis, there have been proposed multiple methods based on sparse modeling. We have proposed a method for Compton camera imaging. The proposed approach is a sparse modeling method, but the derived algorithm is different from LASSO. We explain the problem and how we derived the method.
Approximate inverse preconditioners for general sparse matrices
Chow, E.; Saad, Y.
1994-12-31
Preconditioned Krylov subspace methods are often very efficient in solving sparse linear matrices that arise from the discretization of elliptic partial differential equations. However, for general sparse indifinite matrices, the usual ILU preconditioners fail, often because of the fact that the resulting factors L and U give rise to unstable forward and backward sweeps. In such cases, alternative preconditioners based on approximate inverses may be attractive. We are currently developing a number of such preconditioners based on iterating on each column to get the approximate inverse. For this approach to be efficient, the iteration must be done in sparse mode, i.e., we must use sparse-matrix by sparse-vector type operatoins. We will discuss a few options and compare their performance on standard problems from the Harwell-Boeing collection.
Maximum constrained sparse coding for image representation
NASA Astrophysics Data System (ADS)
Zhang, Jie; Zhao, Danpei; Jiang, Zhiguo
2015-12-01
Sparse coding exhibits good performance in many computer vision applications by finding bases which capture highlevel semantics of the data and learning sparse coefficients in terms of the bases. However, due to the fact that bases are non-orthogonal, sparse coding can hardly preserve the samples' similarity, which is important for discrimination. In this paper, a new image representing method called maximum constrained sparse coding (MCSC) is proposed. Sparse representation with more active coefficients means more similarity information, and the infinite norm is added to the solution for this purpose. We solve the optimizer by constraining the codes' maximum and releasing the residual to other dictionary atoms. Experimental results on image clustering show that our method can preserve the similarity of adjacent samples and maintain the sparsity of code simultaneously.
Large-scale sparse singular value computations
NASA Technical Reports Server (NTRS)
Berry, Michael W.
1992-01-01
Four numerical methods for computing the singular value decomposition (SVD) of large sparse matrices on a multiprocessor architecture are presented. Lanczos and subspace iteration-based methods for determining several of the largest singular triplets (singular values and corresponding left and right-singular vectors) for sparse matrices arising from two practical applications: information retrieval and seismic reflection tomography are emphasized. The target architectures for implementations are the CRAY-2S/4-128 and Alliant FX/80. The sparse SVD problem is well motivated by recent information-retrieval techniques in which dominant singular values and their corresponding singular vectors of large sparse term-document matrices are desired, and by nonlinear inverse problems from seismic tomography applications which require approximate pseudo-inverses of large sparse Jacobian matrices.
Efficient nearest neighbors via robust sparse hashing.
Cherian, Anoop; Sra, Suvrit; Morellas, Vassilios; Papanikolopoulos, Nikolaos
2014-08-01
This paper presents a new nearest neighbor (NN) retrieval framework: robust sparse hashing (RSH). Our approach is inspired by the success of dictionary learning for sparse coding. Our key idea is to sparse code the data using a learned dictionary, and then to generate hash codes out of these sparse codes for accurate and fast NN retrieval. But, direct application of sparse coding to NN retrieval poses a technical difficulty: when data are noisy or uncertain (which is the case with most real-world data sets), for a query point, an exact match of the hash code generated from the sparse code seldom happens, thereby breaking the NN retrieval. Borrowing ideas from robust optimization theory, we circumvent this difficulty via our novel robust dictionary learning and sparse coding framework called RSH, by learning dictionaries on the robustified counterparts of the perturbed data points. The algorithm is applied to NN retrieval on both simulated and real-world data. Our results demonstrate that RSH holds significant promise for efficient NN retrieval against the state of the art.
Sparse High Dimensional Models in Economics.
Fan, Jianqing; Lv, Jinchi; Qi, Lei
2011-09-01
This paper reviews the literature on sparse high dimensional models and discusses some applications in economics and finance. Recent developments of theory, methods, and implementations in penalized least squares and penalized likelihood methods are highlighted. These variable selection methods are proved to be effective in high dimensional sparse modeling. The limits of dimensionality that regularization methods can handle, the role of penalty functions, and their statistical properties are detailed. Some recent advances in ultra-high dimensional sparse modeling are also briefly discussed. PMID:22022635
Sparse High Dimensional Models in Economics
Fan, Jianqing; Lv, Jinchi; Qi, Lei
2010-01-01
This paper reviews the literature on sparse high dimensional models and discusses some applications in economics and finance. Recent developments of theory, methods, and implementations in penalized least squares and penalized likelihood methods are highlighted. These variable selection methods are proved to be effective in high dimensional sparse modeling. The limits of dimensionality that regularization methods can handle, the role of penalty functions, and their statistical properties are detailed. Some recent advances in ultra-high dimensional sparse modeling are also briefly discussed. PMID:22022635
Enhancing Scalability of Sparse Direct Methods
Li, Xiaoye S.; Demmel, James; Grigori, Laura; Gu, Ming; Xia,Jianlin; Jardin, Steve; Sovinec, Carl; Lee, Lie-Quan
2007-07-23
TOPS is providing high-performance, scalable sparse direct solvers, which have had significant impacts on the SciDAC applications, including fusion simulation (CEMM), accelerator modeling (COMPASS), as well as many other mission-critical applications in DOE and elsewhere. Our recent developments have been focusing on new techniques to overcome scalability bottleneck of direct methods, in both time and memory. These include parallelizing symbolic analysis phase and developing linear-complexity sparse factorization methods. The new techniques will make sparse direct methods more widely usable in large 3D simulations on highly-parallel petascale computers.
On finding supernodes for sparse matrix computations
Liu, J.W.H. . Dept. of Computer Science); Ng, E.; Peyton, B.W. )
1990-06-01
A simple characterization of fundamental supernodes is given in terms of the row subtrees of sparse Cholesky factors in the elimination tree. Using this characterization, we present an efficient algorithm that determines the set of such supernodes in time proportional to the number of nonzeros and equations in the original matrix. Experimental results are included to demonstrate the use of this algorithm in the context of sparse supernodal symbolic factorization. 18 refs., 3 figs., 3 tabs.
Deploying temporary networks for upscaling of sparse network stations
NASA Astrophysics Data System (ADS)
Coopersmith, Evan J.; Cosh, Michael H.; Bell, Jesse E.; Kelly, Victoria; Hall, Mark; Palecki, Michael A.; Temimi, Marouane
2016-10-01
Soil observations networks at the national scale play an integral role in hydrologic modeling, drought assessment, agricultural decision support, and our ability to understand climate change. Understanding soil moisture variability is necessary to apply these measurements to model calibration, business and consumer applications, or even human health issues. The installation of soil moisture sensors as sparse, national networks is necessitated by limited financial resources. However, this results in the incomplete sampling of the local heterogeneity of soil type, vegetation cover, topography, and the fine spatial distribution of precipitation events. To this end, temporary networks can be installed in the areas surrounding a permanent installation within a sparse network. The temporary networks deployed in this study provide a more representative average at the 3 km and 9 km scales, localized about the permanent gauge. The value of such temporary networks is demonstrated at test sites in Millbrook, New York and Crossville, Tennessee. The capacity of a single U.S. Climate Reference Network (USCRN) sensor set to approximate the average of a temporary network at the 3 km and 9 km scales using a simple linear scaling function is tested. The capacity of a temporary network to provide reliable estimates with diminishing numbers of sensors, the temporal stability of those networks, and ultimately, the relationship of the variability of those networks to soil moisture conditions at the permanent sensor are investigated. In this manner, this work demonstrates the single-season installation of a temporary network as a mechanism to characterize the soil moisture variability at a permanent gauge within a sparse network.
Finding Nonoverlapping Substructures of a Sparse Matrix
Pinar, Ali; Vassilevska, Virginia
2005-08-11
Many applications of scientific computing rely on computations on sparse matrices. The design of efficient implementations of sparse matrix kernels is crucial for the overall efficiency of these applications. Due to the high compute-to-memory ratio and irregular memory access patterns, the performance of sparse matrix kernels is often far away from the peak performance on a modern processor. Alternative data structures have been proposed, which split the original matrix A into A{sub d} and A{sub s}, so that A{sub d} contains all dense blocks of a specified size in the matrix, and A{sub s} contains the remaining entries. This enables the use of dense matrix kernels on the entries of A{sub d} producing better memory performance. In this work, we study the problem of finding a maximum number of nonoverlapping dense blocks in a sparse matrix, which is previously not studied in the sparse matrix community. We show that the maximum nonoverlapping dense blocks problem is NP-complete by using a reduction from the maximum independent set problem on cubic planar graphs. We also propose a 2/3-approximation algorithm that runs in linear time in the number of nonzeros in the matrix. This extended abstract focuses on our results for 2x2 dense blocks. However we show that our results can be generalized to arbitrary sized dense blocks, and many other oriented substructures, which can be exploited to improve the memory performance of sparse matrix operations.
Sparse extreme learning machine for classification.
Bai, Zuo; Huang, Guang-Bin; Wang, Danwei; Wang, Han; Westover, M Brandon
2014-10-01
Extreme learning machine (ELM) was initially proposed for single-hidden-layer feedforward neural networks (SLFNs). In the hidden layer (feature mapping), nodes are randomly generated independently of training data. Furthermore, a unified ELM was proposed, providing a single framework to simplify and unify different learning methods, such as SLFNs, least square support vector machines, proximal support vector machines, and so on. However, the solution of unified ELM is dense, and thus, usually plenty of storage space and testing time are required for large-scale applications. In this paper, a sparse ELM is proposed as an alternative solution for classification, reducing storage space and testing time. In addition, unified ELM obtains the solution by matrix inversion, whose computational complexity is between quadratic and cubic with respect to the training size. It still requires plenty of training time for large-scale problems, even though it is much faster than many other traditional methods. In this paper, an efficient training algorithm is specifically developed for sparse ELM. The quadratic programming problem involved in sparse ELM is divided into a series of smallest possible sub-problems, each of which are solved analytically. Compared with SVM, sparse ELM obtains better generalization performance with much faster training speed. Compared with unified ELM, sparse ELM achieves similar generalization performance for binary classification applications, and when dealing with large-scale binary classification problems, sparse ELM realizes even faster training speed than unified ELM. PMID:25222727
Visual tracking via robust multitask sparse prototypes
NASA Astrophysics Data System (ADS)
Zhang, Huanlong; Hu, Shiqiang; Yu, Junyang
2015-03-01
Sparse representation has been applied to an online subspace learning-based tracking problem. To handle partial occlusion effectively, some researchers introduce l1 regularization to principal component analysis (PCA) reconstruction. However, in these traditional tracking methods, the representation of each object observation is often viewed as an individual task so the inter-relationship between PCA basis vectors is ignored. We propose a new online visual tracking algorithm with multitask sparse prototypes, which combines multitask sparse learning with PCA-based subspace representation. We first extend a visual tracking algorithm with sparse prototypes in multitask learning framework to mine inter-relations between subtasks. Then, to avoid the problem that enforcing all subtasks to share the same structure may result in degraded tracking results, we impose group sparse constraints on the coefficients of PCA basis vectors and element-wise sparse constraints on the error coefficients, respectively. Finally, we show that the proposed optimization problem can be effectively solved using the accelerated proximal gradient method with the fast convergence. Experimental results compared with the state-of-the-art tracking methods demonstrate that the proposed algorithm achieves favorable performance when the object undergoes partial occlusion, motion blur, and illumination changes.
Learning discriminative dictionary for group sparse representation.
Sun, Yubao; Liu, Qingshan; Tang, Jinhui; Tao, Dacheng
2014-09-01
In recent years, sparse representation has been widely used in object recognition applications. How to learn the dictionary is a key issue to sparse representation. A popular method is to use l1 norm as the sparsity measurement of representation coefficients for dictionary learning. However, the l1 norm treats each atom in the dictionary independently, so the learned dictionary cannot well capture the multisubspaces structural information of the data. In addition, the learned subdictionary for each class usually shares some common atoms, which weakens the discriminative ability of the reconstruction error of each subdictionary. This paper presents a new dictionary learning model to improve sparse representation for image classification, which targets at learning a class-specific subdictionary for each class and a common subdictionary shared by all classes. The model is composed of a discriminative fidelity, a weighted group sparse constraint, and a subdictionary incoherence term. The discriminative fidelity encourages each class-specific subdictionary to sparsely represent the samples in the corresponding class. The weighted group sparse constraint term aims at capturing the structural information of the data. The subdictionary incoherence term is to make all subdictionaries independent as much as possible. Because the common subdictionary represents features shared by all classes, we only use the reconstruction error of each class-specific subdictionary for classification. Extensive experiments are conducted on several public image databases, and the experimental results demonstrate the power of the proposed method, compared with the state-of-the-arts.
Robust face recognition via sparse representation.
Wright, John; Yang, Allen Y; Ganesh, Arvind; Sastry, S Shankar; Ma, Yi
2009-02-01
We consider the problem of automatically recognizing human faces from frontal views with varying expression and illumination, as well as occlusion and disguise. We cast the recognition problem as one of classifying among multiple linear regression models and argue that new theory from sparse signal representation offers the key to addressing this problem. Based on a sparse representation computed by l{1}-minimization, we propose a general classification algorithm for (image-based) object recognition. This new framework provides new insights into two crucial issues in face recognition: feature extraction and robustness to occlusion. For feature extraction, we show that if sparsity in the recognition problem is properly harnessed, the choice of features is no longer critical. What is critical, however, is whether the number of features is sufficiently large and whether the sparse representation is correctly computed. Unconventional features such as downsampled images and random projections perform just as well as conventional features such as Eigenfaces and Laplacianfaces, as long as the dimension of the feature space surpasses certain threshold, predicted by the theory of sparse representation. This framework can handle errors due to occlusion and corruption uniformly by exploiting the fact that these errors are often sparse with respect to the standard (pixel) basis. The theory of sparse representation helps predict how much occlusion the recognition algorithm can handle and how to choose the training images to maximize robustness to occlusion. We conduct extensive experiments on publicly available databases to verify the efficacy of the proposed algorithm and corroborate the above claims.
Wavelet-based reconstruction of fossil-fuel CO2 emissions from sparse measurements
NASA Astrophysics Data System (ADS)
McKenna, S. A.; Ray, J.; Yadav, V.; Van Bloemen Waanders, B.; Michalak, A. M.
2012-12-01
We present a method to estimate spatially resolved fossil-fuel CO2 (ffCO2) emissions from sparse measurements of time-varying CO2 concentrations. It is based on the wavelet-modeling of the strongly non-stationary spatial distribution of ffCO2 emissions. The dimensionality of the wavelet model is first reduced using images of nightlights, which identify regions of human habitation. Since wavelets are a multiresolution basis set, most of the reduction is accomplished by removing fine-scale wavelets, in the regions with low nightlight radiances. The (reduced) wavelet model of emissions is propagated through an atmospheric transport model (WRF) to predict CO2 concentrations at a handful of measurement sites. The estimation of the wavelet model of emissions i.e., inferring the wavelet weights, is performed by fitting to observations at the measurement sites. This is done using Staggered Orthogonal Matching Pursuit (StOMP), which first identifies (and sets to zero) the wavelet coefficients that cannot be estimated from the observations, before estimating the remaining coefficients. This model sparsification and fitting is performed simultaneously, allowing us to explore multiple wavelet-models of differing complexity. This technique is borrowed from the field of compressive sensing, and is generally used in image and video processing. We test this approach using synthetic observations generated from emissions from the Vulcan database. 35 sensor sites are chosen over the USA. FfCO2 emissions, averaged over 8-day periods, are estimated, at a 1 degree spatial resolutions. We find that only about 40% of the wavelets in emission model can be estimated from the data; however the mix of coefficients that are estimated changes with time. Total US emission can be reconstructed with about ~5% errors. The inferred emissions, if aggregated monthly, have a correlation of 0.9 with Vulcan fluxes. We find that the estimated emissions in the Northeast US are the most accurate. Sandia
Finding nonoverlapping substructures of a sparse matrix
Pinar, Ali; Vassilevska, Virginia
2004-08-09
Many applications of scientific computing rely on computations on sparse matrices, thus the design of efficient implementations of sparse matrix kernels is crucial for the overall efficiency of these applications. Due to the high compute-to-memory ratio and irregular memory access patterns, the performance of sparse matrix kernels is often far away from the peak performance on a modern processor. Alternative data structures have been proposed, which split the original matrix A into A{sub d} and A{sub s}, so that A{sub d} contains all dense blocks of a specified size in the matrix, and A{sub s} contains the remaining entries. This enables the use of dense matrix kernels on the entries of A{sub d} producing better memory performance. In this work, we study the problem of finding a maximum number of non overlapping rectangular dense blocks in a sparse matrix, which has not been studied in the sparse matrix community. We show that the maximum non overlapping dense blocks problem is NP-complete by using a reduction from the maximum independent set problem on cubic planar graphs. We also propose a 2/3-approximation algorithm for 2 times 2 blocks that runs in linear time in the number of nonzeros in the matrix. We discuss alternatives to rectangular blocks such as diagonal blocks and cross blocks and present complexity analysis and approximation algorithms.
Removing sparse noise from hyperspectral images with sparse and low-rank penalties
NASA Astrophysics Data System (ADS)
Tariyal, Snigdha; Aggarwal, Hemant Kumar; Majumdar, Angshul
2016-03-01
In diffraction grating, at times, there are defective pixels on the focal plane array; this results in horizontal lines of corrupted pixels in some channels. Since only a few such pixels exist, the corruption/noise is sparse. Studies on sparse noise removal from hyperspectral noise are parsimonious. To remove such sparse noise, a prior work exploited the interband spectral correlation along with intraband spatial redundancy to yield a sparse representation in transform domains. We improve upon the prior technique. The intraband spatial redundancy is modeled as a sparse set of transform coefficients and the interband spectral correlation is modeled as a rank deficient matrix. The resulting optimization problem is solved using the split Bregman technique. Comparative experimental results show that our proposed approach is better than the previous one.
Fast wavelet based sparse approximate inverse preconditioner
Wan, W.L.
1996-12-31
Incomplete LU factorization is a robust preconditioner for both general and PDE problems but unfortunately not easy to parallelize. Recent study of Huckle and Grote and Chow and Saad showed that sparse approximate inverse could be a potential alternative while readily parallelizable. However, for special class of matrix A that comes from elliptic PDE problems, their preconditioners are not optimal in the sense that independent of mesh size. A reason may be that no good sparse approximate inverse exists for the dense inverse matrix. Our observation is that for this kind of matrices, its inverse entries typically have piecewise smooth changes. We can take advantage of this fact and use wavelet compression techniques to construct a better sparse approximate inverse preconditioner. We shall show numerically that our approach is effective for this kind of matrices.
Sparse regularization for force identification using dictionaries
NASA Astrophysics Data System (ADS)
Qiao, Baijie; Zhang, Xingwu; Wang, Chenxi; Zhang, Hang; Chen, Xuefeng
2016-04-01
The classical function expansion method based on minimizing l2-norm of the response residual employs various basis functions to represent the unknown force. Its difficulty lies in determining the optimum number of basis functions. Considering the sparsity of force in the time domain or in other basis space, we develop a general sparse regularization method based on minimizing l1-norm of the coefficient vector of basis functions. The number of basis functions is adaptively determined by minimizing the number of nonzero components in the coefficient vector during the sparse regularization process. First, according to the profile of the unknown force, the dictionary composed of basis functions is determined. Second, a sparsity convex optimization model for force identification is constructed. Third, given the transfer function and the operational response, Sparse reconstruction by separable approximation (SpaRSA) is developed to solve the sparse regularization problem of force identification. Finally, experiments including identification of impact and harmonic forces are conducted on a cantilever thin plate structure to illustrate the effectiveness and applicability of SpaRSA. Besides the Dirac dictionary, other three sparse dictionaries including Db6 wavelets, Sym4 wavelets and cubic B-spline functions can also accurately identify both the single and double impact forces from highly noisy responses in a sparse representation frame. The discrete cosine functions can also successfully reconstruct the harmonic forces including the sinusoidal, square and triangular forces. Conversely, the traditional Tikhonov regularization method with the L-curve criterion fails to identify both the impact and harmonic forces in these cases.
Tensor methods for large, sparse unconstrained optimization
Bouaricha, A.
1996-11-01
Tensor methods for unconstrained optimization were first introduced by Schnabel and Chow [SIAM J. Optimization, 1 (1991), pp. 293-315], who describe these methods for small to moderate size problems. This paper extends these methods to large, sparse unconstrained optimization problems. This requires an entirely new way of solving the tensor model that makes the methods suitable for solving large, sparse optimization problems efficiently. We present test results for sets of problems where the Hessian at the minimizer is nonsingular and where it is singular. These results show that tensor methods are significantly more efficient and more reliable than standard methods based on Newton`s method.
Analog system for computing sparse codes
Rozell, Christopher John; Johnson, Don Herrick; Baraniuk, Richard Gordon; Olshausen, Bruno A.; Ortman, Robert Lowell
2010-08-24
A parallel dynamical system for computing sparse representations of data, i.e., where the data can be fully represented in terms of a small number of non-zero code elements, and for reconstructing compressively sensed images. The system is based on the principles of thresholding and local competition that solves a family of sparse approximation problems corresponding to various sparsity metrics. The system utilizes Locally Competitive Algorithms (LCAs), nodes in a population continually compete with neighboring units using (usually one-way) lateral inhibition to calculate coefficients representing an input in an over complete dictionary.
Multisnapshot Sparse Bayesian Learning for DOA
NASA Astrophysics Data System (ADS)
Gerstoft, Peter; Mecklenbrauker, Christoph F.; Xenaki, Angeliki; Nannuru, Santosh
2016-10-01
The directions of arrival (DOA) of plane waves are estimated from multi-snapshot sensor array data using Sparse Bayesian Learning (SBL). The prior source amplitudes is assumed independent zero-mean complex Gaussian distributed with hyperparameters the unknown variances (i.e. the source powers). For a complex Gaussian likelihood with hyperparameter the unknown noise variance, the corresponding Gaussian posterior distribution is derived. For a given number of DOAs, the hyperparameters are automatically selected by maximizing the evidence and promote sparse DOA estimates. The SBL scheme for DOA estimation is discussed and evaluated competitively against LASSO ($\\ell_1$-regularization), conventional beamforming, and MUSIC
Efficient quantum circuits for arbitrary sparse unitaries
Jordan, Stephen P.; Wocjan, Pawel
2009-12-15
Arbitrary exponentially large unitaries cannot be implemented efficiently by quantum circuits. However, we show that quantum circuits can efficiently implement any unitary provided it has at most polynomially many nonzero entries in any row or column, and these entries are efficiently computable. One can formulate a model of computation based on the composition of sparse unitaries which includes the quantum Turing machine model, the quantum circuit model, anyonic models, permutational quantum computation, and discrete time quantum walks as special cases. Thus, we obtain a simple unified proof that these models are all contained in BQP. Furthermore, our general method for implementing sparse unitaries simplifies several existing quantum algorithms.
Sparse Density Estimation on the Multinomial Manifold.
Hong, Xia; Gao, Junbin; Chen, Sheng; Zia, Tanveer
2015-11-01
A new sparse kernel density estimator is introduced based on the minimum integrated square error criterion for the finite mixture model. Since the constraint on the mixing coefficients of the finite mixture model is on the multinomial manifold, we use the well-known Riemannian trust-region (RTR) algorithm for solving this problem. The first- and second-order Riemannian geometry of the multinomial manifold are derived and utilized in the RTR algorithm. Numerical examples are employed to demonstrate that the proposed approach is effective in constructing sparse kernel density estimators with an accuracy competitive with those of existing kernel density estimators. PMID:25647665
Infrared image recognition based on structure sparse and atomic sparse parallel
NASA Astrophysics Data System (ADS)
Wu, Yalu; Li, Ruilong; Xu, Yi; Wang, Liping
2015-12-01
Use the redundancy of the super complete dictionary can capture the structural features of the image effectively, can achieving the effective representation of the image. However, the commonly used atomic sparse representation without regard the structure of the dictionary and the unrelated non-zero-term in the process of the computation, though structure sparse consider the structure feature of dictionary, the majority coefficients of the blocks maybe are non-zero, it may affect the identification efficiency. For the disadvantages of these two sparse expressions, a weighted parallel atomic sparse and sparse structure is proposed, and the recognition efficiency is improved by the adaptive computation of the optimal weights. The atomic sparse expression and structure sparse expression are respectively, and the optimal weights are calculated by the adaptive method. Methods are as follows: training by using the less part of the identification sample, the recognition rate is calculated by the increase of the certain step size and t the constraint between weight. The recognition rate as the Z axis, two weight values respectively as X, Y axis, the resulting points can be connected in a straight line in the 3 dimensional coordinate system, by solving the highest recognition rate, the optimal weights can be obtained. Through simulation experiments can be known, the optimal weights based on adaptive method are better in the recognition rate, weights obtained by adaptive computation of a few samples, suitable for parallel recognition calculation, can effectively improve the recognition rate of infrared images.
Second SIAM conference on sparse matrices: Abstracts. Final technical report
1996-12-31
This report contains abstracts on the following topics: invited and long presentations (IP1 & LP1); sparse matrix reordering & graph theory I; sparse matrix tools & environments I; eigenvalue computations I; iterative methods & acceleration techniques I; applications I; parallel algorithms I; sparse matrix reordering & graphy theory II; sparse matrix tool & environments II; least squares & optimization I; iterative methods & acceleration techniques II; applications II; eigenvalue computations II; least squares & optimization II; parallel algorithms II; sparse direct methods; iterative methods & acceleration techniques III; eigenvalue computations III; and sparse matrix reordering & graph theory III.
Facial expression recognition with facial parts based sparse representation classifier
NASA Astrophysics Data System (ADS)
Zhi, Ruicong; Ruan, Qiuqi
2009-10-01
Facial expressions play important role in human communication. The understanding of facial expression is a basic requirement in the development of next generation human computer interaction systems. Researches show that the intrinsic facial features always hide in low dimensional facial subspaces. This paper presents facial parts based facial expression recognition system with sparse representation classifier. Sparse representation classifier exploits sparse representation to select face features and classify facial expressions. The sparse solution is obtained by solving l1 -norm minimization problem with constraint of linear combination equation. Experimental results show that sparse representation is efficient for facial expression recognition and sparse representation classifier obtain much higher recognition accuracies than other compared methods.
Learning Stable Multilevel Dictionaries for Sparse Representations.
Thiagarajan, Jayaraman J; Ramamurthy, Karthikeyan Natesan; Spanias, Andreas
2015-09-01
Sparse representations using learned dictionaries are being increasingly used with success in several data processing and machine learning applications. The increasing need for learning sparse models in large-scale applications motivates the development of efficient, robust, and provably good dictionary learning algorithms. Algorithmic stability and generalizability are desirable characteristics for dictionary learning algorithms that aim to build global dictionaries, which can efficiently model any test data similar to the training samples. In this paper, we propose an algorithm to learn dictionaries for sparse representations from large scale data, and prove that the proposed learning algorithm is stable and generalizable asymptotically. The algorithm employs a 1-D subspace clustering procedure, the K-hyperline clustering, to learn a hierarchical dictionary with multiple levels. We also propose an information-theoretic scheme to estimate the number of atoms needed in each level of learning and develop an ensemble approach to learn robust dictionaries. Using the proposed dictionaries, the sparse code for novel test data can be computed using a low-complexity pursuit procedure. We demonstrate the stability and generalization characteristics of the proposed algorithm using simulations. We also evaluate the utility of the multilevel dictionaries in compressed recovery and subspace learning applications.
Multilevel sparse functional principal component analysis.
Di, Chongzhi; Crainiceanu, Ciprian M; Jank, Wolfgang S
2014-01-29
We consider analysis of sparsely sampled multilevel functional data, where the basic observational unit is a function and data have a natural hierarchy of basic units. An example is when functions are recorded at multiple visits for each subject. Multilevel functional principal component analysis (MFPCA; Di et al. 2009) was proposed for such data when functions are densely recorded. Here we consider the case when functions are sparsely sampled and may contain only a few observations per function. We exploit the multilevel structure of covariance operators and achieve data reduction by principal component decompositions at both between and within subject levels. We address inherent methodological differences in the sparse sampling context to: 1) estimate the covariance operators; 2) estimate the functional principal component scores; 3) predict the underlying curves. Through simulations the proposed method is able to discover dominating modes of variations and reconstruct underlying curves well even in sparse settings. Our approach is illustrated by two applications, the Sleep Heart Health Study and eBay auctions. PMID:24872597
Multilevel sparse functional principal component analysis
Di, Chongzhi; Crainiceanu, Ciprian M.; Jank, Wolfgang S.
2014-01-01
We consider analysis of sparsely sampled multilevel functional data, where the basic observational unit is a function and data have a natural hierarchy of basic units. An example is when functions are recorded at multiple visits for each subject. Multilevel functional principal component analysis (MFPCA; Di et al. 2009) was proposed for such data when functions are densely recorded. Here we consider the case when functions are sparsely sampled and may contain only a few observations per function. We exploit the multilevel structure of covariance operators and achieve data reduction by principal component decompositions at both between and within subject levels. We address inherent methodological differences in the sparse sampling context to: 1) estimate the covariance operators; 2) estimate the functional principal component scores; 3) predict the underlying curves. Through simulations the proposed method is able to discover dominating modes of variations and reconstruct underlying curves well even in sparse settings. Our approach is illustrated by two applications, the Sleep Heart Health Study and eBay auctions. PMID:24872597
Algorithms for sparse nonnegative Tucker decompositions.
Mørup, Morten; Hansen, Lars Kai; Arnfred, Sidse M
2008-08-01
There is a increasing interest in analysis of large-scale multiway data. The concept of multiway data refers to arrays of data with more than two dimensions, that is, taking the form of tensors. To analyze such data, decomposition techniques are widely used. The two most common decompositions for tensors are the Tucker model and the more restricted PARAFAC model. Both models can be viewed as generalizations of the regular factor analysis to data of more than two modalities. Nonnegative matrix factorization (NMF), in conjunction with sparse coding, has recently been given much attention due to its part-based and easy interpretable representation. While NMF has been extended to the PARAFAC model, no such attempt has been done to extend NMF to the Tucker model. However, if the tensor data analyzed are nonnegative, it may well be relevant to consider purely additive (i.e., nonnegative) Tucker decompositions). To reduce ambiguities of this type of decomposition, we develop updates that can impose sparseness in any combination of modalities, hence, proposed algorithms for sparse nonnegative Tucker decompositions (SN-TUCKER). We demonstrate how the proposed algorithms are superior to existing algorithms for Tucker decompositions when the data and interactions can be considered nonnegative. We further illustrate how sparse coding can help identify what model (PARAFAC or Tucker) is more appropriate for the data as well as to select the number of components by turning off excess components. The algorithms for SN-TUCKER can be downloaded from Mørup (2007).
Self-Control in Sparsely Coded Networks
NASA Astrophysics Data System (ADS)
Dominguez, D. R. C.; Bollé, D.
1998-03-01
A complete self-control mechanism is proposed in the dynamics of neural networks through the introduction of a time-dependent threshold, determined in function of both the noise and the pattern activity in the network. Especially for sparsely coded models this mechanism is shown to considerably improve the storage capacity, the basins of attraction, and the mutual information content.
Sparse matrix orderings for factorized inverse preconditioners
Benzi, M.; Tuama, M.
1998-09-01
The effect of reorderings on the performance of factorized sparse approximate inverse preconditioners is considered. It is shown that certain reorderings can be very beneficial both in the preconditioner construction phase and in terms of the rate of convergence of the preconditioned iteration.
STIS Sparse Field CTE test {Cycle 9}
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
2000-07-01
CTE measurements are made using the "sparse field test", along both the serial and parallel axes. This program needs special commanding to provide {a} off-center MSM positionings of some slits, and {b} the ability to read out with any amplifier {A, B, C, or D}. All exposures are internals.
STIS Sparse Field CTE test {Cycle 8}
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
1999-07-01
CTE measurements are made using the "sparse field test", along both the serial and parallel axes. This program needs special commanding to provide {a} off-center MSM positionings of some slits, and {b} the ability to read out with any amplifier {A, B, C, or D}. All exposures are internals.
SAR Image Despeckling Via Structural Sparse Representation
NASA Astrophysics Data System (ADS)
Lu, Ting; Li, Shutao; Fang, Leyuan; Benediktsson, Jón Atli
2016-12-01
A novel synthetic aperture radar (SAR) image despeckling method based on structural sparse representation is introduced. The proposed method utilizes the fact that different regions in SAR images correspond to varying terrain reflectivity. Therefore, SAR images can be split into a heterogeneous class (with a varied terrain reflectivity) and a homogeneous class (with a constant terrain reflectivity). In the proposed method, different sparse representation based despeckling schemes are designed by combining the different region characteristics in SAR images. For heterogeneous regions with rich structure and texture information, structural dictionaries are learned to appropriately represent varied structural characteristics. Specifically, each patch in these regions is sparsely coded with the best fitted structural dictionary, thus good structure preservation can be obtained. For homogenous regions without rich structure and texture information, the highly redundant photometric self-similarity is exploited to suppress speckle noise without introducing artifacts. That is achieved by firstly learning the sub-dictionary, then simultaneously sparsely coding for each group of photometrically similar image patches. Visual and objective experimental results demonstrate the superiority of the proposed method over the-state-of-the-art methods.
A Comparative Study of Sparse Associative Memories
NASA Astrophysics Data System (ADS)
Gripon, Vincent; Heusel, Judith; Löwe, Matthias; Vermet, Franck
2016-07-01
We study various models of associative memories with sparse information, i.e. a pattern to be stored is a random string of 0s and 1s with about log N 1s, only. We compare different synaptic weights, architectures and retrieval mechanisms to shed light on the influence of the various parameters on the storage capacity.
Causal Network Inference Via Group Sparse Regularization.
Bolstad, Andrew; Van Veen, Barry D; Nowak, Robert
2011-06-11
This paper addresses the problem of inferring sparse causal networks modeled by multivariate autoregressive (MAR) processes. Conditions are derived under which the Group Lasso (gLasso) procedure consistently estimates sparse network structure. The key condition involves a "false connection score" ψ. In particular, we show that consistent recovery is possible even when the number of observations of the network is far less than the number of parameters describing the network, provided that ψ < 1. The false connection score is also demonstrated to be a useful metric of recovery in nonasymptotic regimes. The conditions suggest a modified gLasso procedure which tends to improve the false connection score and reduce the chances of reversing the direction of causal influence. Computational experiments and a real network based electrocorticogram (ECoG) simulation study demonstrate the effectiveness of the approach.
Statistical prediction with Kanerva's sparse distributed memory
NASA Technical Reports Server (NTRS)
Rogers, David
1989-01-01
A new viewpoint of the processing performed by Kanerva's sparse distributed memory (SDM) is presented. In conditions of near- or over-capacity, where the associative-memory behavior of the model breaks down, the processing performed by the model can be interpreted as that of a statistical predictor. Mathematical results are presented which serve as the framework for a new statistical viewpoint of sparse distributed memory and for which the standard formulation of SDM is a special case. This viewpoint suggests possible enhancements to the SDM model, including a procedure for improving the predictiveness of the system based on Holland's work with genetic algorithms, and a method for improving the capacity of SDM even when used as an associative memory.
Solving large sparse eigenvalue problems on supercomputers
NASA Technical Reports Server (NTRS)
Philippe, Bernard; Saad, Youcef
1988-01-01
An important problem in scientific computing consists in finding a few eigenvalues and corresponding eigenvectors of a very large and sparse matrix. The most popular methods to solve these problems are based on projection techniques on appropriate subspaces. The main attraction of these methods is that they only require the use of the matrix in the form of matrix by vector multiplications. The implementations on supercomputers of two such methods for symmetric matrices, namely Lanczos' method and Davidson's method are compared. Since one of the most important operations in these two methods is the multiplication of vectors by the sparse matrix, methods of performing this operation efficiently are discussed. The advantages and the disadvantages of each method are compared and implementation aspects are discussed. Numerical experiments on a one processor CRAY 2 and CRAY X-MP are reported. Possible parallel implementations are also discussed.
Sparse representation for color image restoration.
Mairal, Julien; Elad, Michael; Sapiro, Guillermo
2008-01-01
Sparse representations of signals have drawn considerable interest in recent years. The assumption that natural signals, such as images, admit a sparse decomposition over a redundant dictionary leads to efficient algorithms for handling such sources of data. In particular, the design of well adapted dictionaries for images has been a major challenge. The K-SVD has been recently proposed for this task and shown to perform very well for various grayscale image processing tasks. In this paper, we address the problem of learning dictionaries for color images and extend the K-SVD-based grayscale image denoising algorithm that appears in. This work puts forward ways for handling nonhomogeneous noise and missing information, paving the way to state-of-the-art results in applications such as color image denoising, demosaicing, and inpainting, as demonstrated in this paper. PMID:18229804
Sparse brain network using penalized linear regression
NASA Astrophysics Data System (ADS)
Lee, Hyekyoung; Lee, Dong Soo; Kang, Hyejin; Kim, Boong-Nyun; Chung, Moo K.
2011-03-01
Sparse partial correlation is a useful connectivity measure for brain networks when it is difficult to compute the exact partial correlation in the small-n large-p setting. In this paper, we formulate the problem of estimating partial correlation as a sparse linear regression with a l1-norm penalty. The method is applied to brain network consisting of parcellated regions of interest (ROIs), which are obtained from FDG-PET images of the autism spectrum disorder (ASD) children and the pediatric control (PedCon) subjects. To validate the results, we check their reproducibilities of the obtained brain networks by the leave-one-out cross validation and compare the clustered structures derived from the brain networks of ASD and PedCon.
Causal Network Inference Via Group Sparse Regularization
Bolstad, Andrew; Van Veen, Barry D.; Nowak, Robert
2011-01-01
This paper addresses the problem of inferring sparse causal networks modeled by multivariate autoregressive (MAR) processes. Conditions are derived under which the Group Lasso (gLasso) procedure consistently estimates sparse network structure. The key condition involves a “false connection score” ψ. In particular, we show that consistent recovery is possible even when the number of observations of the network is far less than the number of parameters describing the network, provided that ψ < 1. The false connection score is also demonstrated to be a useful metric of recovery in nonasymptotic regimes. The conditions suggest a modified gLasso procedure which tends to improve the false connection score and reduce the chances of reversing the direction of causal influence. Computational experiments and a real network based electrocorticogram (ECoG) simulation study demonstrate the effectiveness of the approach. PMID:21918591
Neural process reconstruction from sparse user scribbles.
Roberts, Mike; Jeong, Won-Ki; Vázquez-Reina, Amelio; Unger, Markus; Bischof, Horst; Lichtman, Jeff; Pfister, Hanspeter
2011-01-01
We present a novel semi-automatic method for segmenting neural processes in large, highly anisotropic EM (electron microscopy) image stacks. Our method takes advantage of sparse scribble annotations provided by the user to guide a 3D variational segmentation model, thereby allowing our method to globally optimally enforce 3D geometric constraints on the segmentation. Moreover, we leverage a novel algorithm for propagating segmentation constraints through the image stack via optimal volumetric pathways, thereby allowing our method to compute highly accurate 3D segmentations from very sparse user input. We evaluate our method by reconstructing 16 neural processes in a 1024 x 1024 x 50 nanometer-scale EM image stack of a mouse hippocampus. We demonstrate that, on average, our method is 68% more accurate than previous state-of-the-art semi-automatic methods. PMID:22003670
Fingerprint Compression Based on Sparse Representation.
Shao, Guangqi; Wu, Yanping; A, Yong; Liu, Xiao; Guo, Tiande
2014-02-01
A new fingerprint compression algorithm based on sparse representation is introduced. Obtaining an overcomplete dictionary from a set of fingerprint patches allows us to represent them as a sparse linear combination of dictionary atoms. In the algorithm, we first construct a dictionary for predefined fingerprint image patches. For a new given fingerprint images, represent its patches according to the dictionary by computing l(0)-minimization and then quantize and encode the representation. In this paper, we consider the effect of various factors on compression results. Three groups of fingerprint images are tested. The experiments demonstrate that our algorithm is efficient compared with several competing compression techniques (JPEG, JPEG 2000, and WSQ), especially at high compression ratios. The experiments also illustrate that the proposed algorithm is robust to extract minutiae.
Inpainting with sparse linear combinations of exemplars
Wohlberg, Brendt
2008-01-01
We introduce a new exemplar-based inpainting algorithm based on representing the region to be inpainted as a sparse linear combination of blocks extracted from similar parts of the image being inpainted. This method is conceptually simple, being computed by functional minimization, and avoids the complexity of correctly ordering the filling in of missing regions of other exemplar-based methods. Initial performance comparisons on small inpainting regions indicate that this method provides similar or better performance than other recent methods.
Sparse image reconstruction for molecular imaging.
Ting, Michael; Raich, Raviv; Hero, Alfred O
2009-06-01
The application that motivates this paper is molecular imaging at the atomic level. When discretized at subatomic distances, the volume is inherently sparse. Noiseless measurements from an imaging technology can be modeled by convolution of the image with the system point spread function (psf). Such is the case with magnetic resonance force microscopy (MRFM), an emerging technology where imaging of an individual tobacco mosaic virus was recently demonstrated with nanometer resolution. We also consider additive white Gaussian noise (AWGN) in the measurements. Many prior works of sparse estimators have focused on the case when H has low coherence; however, the system matrix H in our application is the convolution matrix for the system psf. A typical convolution matrix has high coherence. This paper, therefore, does not assume a low coherence H. A discrete-continuous form of the Laplacian and atom at zero (LAZE) p.d.f. used by Johnstone and Silverman is formulated, and two sparse estimators derived by maximizing the joint p.d.f. of the observation and image conditioned on the hyperparameters. A thresholding rule that generalizes the hard and soft thresholding rule appears in the course of the derivation. This so-called hybrid thresholding rule, when used in the iterative thresholding framework, gives rise to the hybrid estimator, a generalization of the lasso. Estimates of the hyperparameters for the lasso and hybrid estimator are obtained via Stein's unbiased risk estimate (SURE). A numerical study with a Gaussian psf and two sparse images shows that the hybrid estimator outperforms the lasso.
Notes on implementation of sparsely distributed memory
NASA Technical Reports Server (NTRS)
Keeler, J. D.; Denning, P. J.
1986-01-01
The Sparsely Distributed Memory (SDM) developed by Kanerva is an unconventional memory design with very interesting and desirable properties. The memory works in a manner that is closely related to modern theories of human memory. The SDM model is discussed in terms of its implementation in hardware. Two appendices discuss the unconventional approaches of the SDM: Appendix A treats a resistive circuit for fast, parallel address decoding; and Appendix B treats a systolic array for high throughput read and write operations.
The relationship between phytomass, NDVI and vegetation communities on Svalbard
NASA Astrophysics Data System (ADS)
Johansen, Bernt; Tømmervik, Hans
2014-04-01
Several studies have shown a close relationship between vegetation fertility and different vegetation indices extracted from satellite data. The vegetation fertility in Arctic is at overall scales highly related to temperature. At lower scales surface material, snow cover, hydrology and anthropogenic effects (geese, reindeer) are determinant in constituting the different vegetation communities. The extent and occurrence of different vegetation communities are expressed in vegetation maps. On Svalbard a vegetation map covering the entire archipelago has recently been developed. The map is differentiated into 18 map units showing large areas of non- and sparsely vegetated ground. The most favorable vegetation is seen as productive marshes and moss tundra communities in the lowland. Various mathematical combinations of spectral channels in satellite images have been applied as sensitive indicators of the presence and condition of green vegetation. Today the normalized difference vegetation index (NDVI) is mostly used to display this information. NDVI is an indicator of the density of chlorophyll in leaf tissue calculated from the red and near infrared bands: NDVI = (NIR - RED)/(NIR + RED). NDVI gives values between -1 and +1 where vegetated areas in general yield high positive values, while non-vegetated ground is found on the negative side.
Technology Transfer Automated Retrieval System (TEKTRAN)
The wide variety of fermented foods of the world can be classified by the materials obtained from the fermentation, such as alcohol (beer, wine), organic acid such as lactic acid and acetic acid (vegetables, dairy), carbon dioxide (bread), and amino acids or peptides from protein (fish fermentations...
Modified sparse regularization for electrical impedance tomography.
Fan, Wenru; Wang, Huaxiang; Xue, Qian; Cui, Ziqiang; Sun, Benyuan; Wang, Qi
2016-03-01
Electrical impedance tomography (EIT) aims to estimate the electrical properties at the interior of an object from current-voltage measurements on its boundary. It has been widely investigated due to its advantages of low cost, non-radiation, non-invasiveness, and high speed. Image reconstruction of EIT is a nonlinear and ill-posed inverse problem. Therefore, regularization techniques like Tikhonov regularization are used to solve the inverse problem. A sparse regularization based on L1 norm exhibits superiority in preserving boundary information at sharp changes or discontinuous areas in the image. However, the limitation of sparse regularization lies in the time consumption for solving the problem. In order to further improve the calculation speed of sparse regularization, a modified method based on separable approximation algorithm is proposed by using adaptive step-size and preconditioning technique. Both simulation and experimental results show the effectiveness of the proposed method in improving the image quality and real-time performance in the presence of different noise intensities and conductivity contrasts. PMID:27036798
Modified sparse regularization for electrical impedance tomography.
Fan, Wenru; Wang, Huaxiang; Xue, Qian; Cui, Ziqiang; Sun, Benyuan; Wang, Qi
2016-03-01
Electrical impedance tomography (EIT) aims to estimate the electrical properties at the interior of an object from current-voltage measurements on its boundary. It has been widely investigated due to its advantages of low cost, non-radiation, non-invasiveness, and high speed. Image reconstruction of EIT is a nonlinear and ill-posed inverse problem. Therefore, regularization techniques like Tikhonov regularization are used to solve the inverse problem. A sparse regularization based on L1 norm exhibits superiority in preserving boundary information at sharp changes or discontinuous areas in the image. However, the limitation of sparse regularization lies in the time consumption for solving the problem. In order to further improve the calculation speed of sparse regularization, a modified method based on separable approximation algorithm is proposed by using adaptive step-size and preconditioning technique. Both simulation and experimental results show the effectiveness of the proposed method in improving the image quality and real-time performance in the presence of different noise intensities and conductivity contrasts.
Aerial Scene Recognition using Efficient Sparse Representation
Cheriyadat, Anil M
2012-01-01
Advanced scene recognition systems for processing large volumes of high-resolution aerial image data are in great demand today. However, automated scene recognition remains a challenging problem. Efficient encoding and representation of spatial and structural patterns in the imagery are key in developing automated scene recognition algorithms. We describe an image representation approach that uses simple and computationally efficient sparse code computation to generate accurate features capable of producing excellent classification performance using linear SVM kernels. Our method exploits unlabeled low-level image feature measurements to learn a set of basis vectors. We project the low-level features onto the basis vectors and use simple soft threshold activation function to derive the sparse features. The proposed technique generates sparse features at a significantly lower computational cost than other methods~\\cite{Yang10, newsam11}, yet it produces comparable or better classification accuracy. We apply our technique to high-resolution aerial image datasets to quantify the aerial scene classification performance. We demonstrate that the dense feature extraction and representation methods are highly effective for automatic large-facility detection on wide area high-resolution aerial imagery.
Mean-field sparse optimal control
Fornasier, Massimo; Piccoli, Benedetto; Rossi, Francesco
2014-01-01
We introduce the rigorous limit process connecting finite dimensional sparse optimal control problems with ODE constraints, modelling parsimonious interventions on the dynamics of a moving population divided into leaders and followers, to an infinite dimensional optimal control problem with a constraint given by a system of ODE for the leaders coupled with a PDE of Vlasov-type, governing the dynamics of the probability distribution of the followers. In the classical mean-field theory, one studies the behaviour of a large number of small individuals freely interacting with each other, by simplifying the effect of all the other individuals on any given individual by a single averaged effect. In this paper, we address instead the situation where the leaders are actually influenced also by an external policy maker, and we propagate its effect for the number N of followers going to infinity. The technical derivation of the sparse mean-field optimal control is realized by the simultaneous development of the mean-field limit of the equations governing the followers dynamics together with the Γ-limit of the finite dimensional sparse optimal control problems. PMID:25288818
SAR Image despeckling via sparse representation
NASA Astrophysics Data System (ADS)
Wang, Zhongmei; Yang, Xiaomei; Zheng, Liang
2014-11-01
SAR image despeckling is an active research area in image processing due to its importance in improving the quality of image for object detection and classification.In this paper, a new approach is proposed for multiplicative noise in SAR image removal based on nonlocal sparse representation by dictionary learning and collaborative filtering. First, a image is divided into many patches, and then a cluster is formed by clustering log-similar image patches using Fuzzy C-means (FCM). For each cluster, an over-complete dictionary is computed using the K-SVD method that iteratively updates the dictionary and the sparse coefficients. The patches belonging to the same cluster are then reconstructed by a sparse combination of the corresponding dictionary atoms. The reconstructed patches are finally collaboratively aggregated to build the denoised image. The experimental results show that the proposed method achieves much better results than many state-of-the-art algorithms in terms of both objective evaluation index (PSNR and ENL) and subjective visual perception.
NASA Astrophysics Data System (ADS)
Qiu, G. Y.; Feng, Y.
2011-12-01
Separately estimation of soil evaporation (E) and vegetation transpiration (T) is important for water management. In this study, a methodology to estimate the ratio of vegetation cover, soil evaporation and vegetation transpiration in a sparse steppe is developed based on a previous published model, the three-temperatures (3T) model. The input parameters of the model includes the surface temperatures of soil and vegetation (from thermal image), net radiation (estimated from surface temperature and solar radiation), and air temperature. The approach of unsupervised classification was used to separate the bare soil and vegetation pixels from the images. The areas with higher temperature could be regarded as the bare soil and E was estimated by the evaporation sub-model in the 3T model; while the areas with lower temperature could be regarded as pure vegetation and T was estimated by the transpiration sub-model in the 3T model. Afterward, the estimated E and T were converted into daily values and compared with the measured E and T by using Bowen Ratio and micro-lysimeter methods. Results show that the proposed approach is a useful way to separately estimated E and T in sparse steppe.
Efficient visual tracking via low-complexity sparse representation
NASA Astrophysics Data System (ADS)
Lu, Weizhi; Zhang, Jinglin; Kpalma, Kidiyo; Ronsin, Joseph
2015-12-01
Thanks to its good performance on object recognition, sparse representation has recently been widely studied in the area of visual object tracking. Up to now, little attention has been paid to the complexity of sparse representation, while most works are focused on the performance improvement. By reducing the computation load related to sparse representation hundreds of times, this paper proposes by far the most computationally efficient tracking approach based on sparse representation. The proposal simply consists of two stages of sparse representation, one is for object detection and the other for object validation. Experimentally, it achieves better performance than some state-of-the-art methods in both accuracy and speed.
Galaxy redshift surveys with sparse sampling
Chiang, Chi-Ting; Wullstein, Philipp; Komatsu, Eiichiro; Jee, Inh; Jeong, Donghui; Blanc, Guillermo A.; Ciardullo, Robin; Gronwall, Caryl; Hagen, Alex; Schneider, Donald P.; Drory, Niv; Fabricius, Maximilian; Landriau, Martin; Finkelstein, Steven; Jogee, Shardha; Cooper, Erin Mentuch; Tuttle, Sarah; Gebhardt, Karl; Hill, Gary J.
2013-12-01
Survey observations of the three-dimensional locations of galaxies are a powerful approach to measure the distribution of matter in the universe, which can be used to learn about the nature of dark energy, physics of inflation, neutrino masses, etc. A competitive survey, however, requires a large volume (e.g., V{sub survey} ∼ 10Gpc{sup 3}) to be covered, and thus tends to be expensive. A ''sparse sampling'' method offers a more affordable solution to this problem: within a survey footprint covering a given survey volume, V{sub survey}, we observe only a fraction of the volume. The distribution of observed regions should be chosen such that their separation is smaller than the length scale corresponding to the wavenumber of interest. Then one can recover the power spectrum of galaxies with precision expected for a survey covering a volume of V{sub survey} (rather than the volume of the sum of observed regions) with the number density of galaxies given by the total number of observed galaxies divided by V{sub survey} (rather than the number density of galaxies within an observed region). We find that regularly-spaced sampling yields an unbiased power spectrum with no window function effect, and deviations from regularly-spaced sampling, which are unavoidable in realistic surveys, introduce calculable window function effects and increase the uncertainties of the recovered power spectrum. On the other hand, we show that the two-point correlation function (pair counting) is not affected by sparse sampling. While we discuss the sparse sampling method within the context of the forthcoming Hobby-Eberly Telescope Dark Energy Experiment, the method is general and can be applied to other galaxy surveys.
Guided wavefield reconstruction from sparse measurements
NASA Astrophysics Data System (ADS)
Mesnil, Olivier; Ruzzene, Massimo
2016-02-01
Guided wave measurements are at the basis of several Non-Destructive Evaluation (NDE) techniques. Although sparse measurements of guided wave obtained using piezoelectric sensors can efficiently detect and locate defects, extensive informa-tion on the shape and subsurface location of defects can be extracted from full-field measurements acquired by Laser Doppler Vibrometers (LDV). Wavefield acquisition from LDVs is generally a slow operation due to the fact that the wave propagation to record must be repeated for each point measurement and the initial conditions must be reached between each measurement. In this research, a Sparse Wavefield Reconstruction (SWR) process using Compressed Sensing is developed. The goal of this technique is to reduce the number of point measurements needed to apply NDE techniques by at least one order of magnitude by extrapolating the knowledge of a few randomly chosen measured pixels over an over-sampled grid. To achieve this, the Lamb wave propagation equation is used to formulate a basis of shape functions in which the wavefield has a sparse representation, in order to comply with the Compressed Sensing requirements and use l1-minimization solvers. The main assumption of this reconstruction process is that every material point of the studied area is a potential source. The Compressed Sensing matrix is defined as being the contribution that would have been received at a measurement location from each possible source, using the dispersion relations of the specimen computed using a Semi-Analytical Finite Element technique. The measurements are then processed through an l1-minimizer to find a minimum corresponding to the set of active sources and their corresponding excitation functions. This minimum represents the best combination of the parameters of the model matching the sparse measurements. Wavefields are then reconstructed using the propagation equation. The set of active sources found by minimization contains all the wave
Effective dimension reduction for sparse functional data
YAO, F.; LEI, E.; WU, Y.
2015-01-01
Summary We propose a method of effective dimension reduction for functional data, emphasizing the sparse design where one observes only a few noisy and irregular measurements for some or all of the subjects. The proposed method borrows strength across the entire sample and provides a way to characterize the effective dimension reduction space, via functional cumulative slicing. Our theoretical study reveals a bias-variance trade-off associated with the regularizing truncation and decaying structures of the predictor process and the effective dimension reduction space. A simulation study and an application illustrate the superior finite-sample performance of the method. PMID:26566293
A view of Kanerva's sparse distributed memory
NASA Technical Reports Server (NTRS)
Denning, P. J.
1986-01-01
Pentti Kanerva is working on a new class of computers, which are called pattern computers. Pattern computers may close the gap between capabilities of biological organisms to recognize and act on patterns (visual, auditory, tactile, or olfactory) and capabilities of modern computers. Combinations of numeric, symbolic, and pattern computers may one day be capable of sustaining robots. The overview of the requirements for a pattern computer, a summary of Kanerva's Sparse Distributed Memory (SDM), and examples of tasks this computer can be expected to perform well are given.
Distributed memory compiler design for sparse problems
NASA Technical Reports Server (NTRS)
Wu, Janet; Saltz, Joel; Berryman, Harry; Hiranandani, Seema
1991-01-01
A compiler and runtime support mechanism is described and demonstrated. The methods presented are capable of solving a wide range of sparse and unstructured problems in scientific computing. The compiler takes as input a FORTRAN 77 program enhanced with specifications for distributing data, and the compiler outputs a message passing program that runs on a distributed memory computer. The runtime support for this compiler is a library of primitives designed to efficiently support irregular patterns of distributed array accesses and irregular distributed array partitions. A variety of Intel iPSC/860 performance results obtained through the use of this compiler are presented.
Sparse dynamics for partial differential equations
Schaeffer, Hayden; Caflisch, Russel; Hauck, Cory D.; Osher, Stanley
2013-01-01
We investigate the approximate dynamics of several differential equations when the solutions are restricted to a sparse subset of a given basis. The restriction is enforced at every time step by simply applying soft thresholding to the coefficients of the basis approximation. By reducing or compressing the information needed to represent the solution at every step, only the essential dynamics are represented. In many cases, there are natural bases derived from the differential equations, which promote sparsity. We find that our method successfully reduces the dynamics of convection equations, diffusion equations, weak shocks, and vorticity equations with high-frequency source terms. PMID:23533273
Parallel preconditioning techniques for sparse CG solvers
Basermann, A.; Reichel, B.; Schelthoff, C.
1996-12-31
Conjugate gradient (CG) methods to solve sparse systems of linear equations play an important role in numerical methods for solving discretized partial differential equations. The large size and the condition of many technical or physical applications in this area result in the need for efficient parallelization and preconditioning techniques of the CG method. In particular for very ill-conditioned matrices, sophisticated preconditioner are necessary to obtain both acceptable convergence and accuracy of CG. Here, we investigate variants of polynomial and incomplete Cholesky preconditioners that markedly reduce the iterations of the simply diagonally scaled CG and are shown to be well suited for massively parallel machines.
Neonatal Atlas Construction Using Sparse Representation
Shi, Feng; Wang, Li; Wu, Guorong; Li, Gang; Gilmore, John H.; Lin, Weili; Shen, Dinggang
2014-01-01
Atlas construction generally includes first an image registration step to normalize all images into a common space and then an atlas building step to fuse the information from all the aligned images. Although numerous atlas construction studies have been performed to improve the accuracy of the image registration step, unweighted or simply weighted average is often used in the atlas building step. In this article, we propose a novel patch-based sparse representation method for atlas construction after all images have been registered into the common space. By taking advantage of local sparse representation, more anatomical details can be recovered in the built atlas. To make the anatomical structures spatially smooth in the atlas, the anatomical feature constraints on group structure of representations and also the overlapping of neighboring patches are imposed to ensure the anatomical consistency between neighboring patches. The proposed method has been applied to 73 neonatal MR images with poor spatial resolution and low tissue contrast, for constructing a neonatal brain atlas with sharp anatomical details. Experimental results demonstrate that the proposed method can significantly enhance the quality of the constructed atlas by discovering more anatomical details especially in the highly convoluted cortical regions. The resulting atlas demonstrates superior performance of our atlas when applied to spatially normalizing three different neonatal datasets, compared with other start-of-the-art neonatal brain atlases. PMID:24638883
A density functional for sparse matter
NASA Astrophysics Data System (ADS)
Langreth, D. C.; Lundqvist, B. I.; Chakarova-Käck, S. D.; Cooper, V. R.; Dion, M.; Hyldgaard, P.; Kelkkanen, A.; Kleis, J.; Kong, Lingzhu; Li, Shen; Moses, P. G.; Murray, E.; Puzder, A.; Rydberg, H.; Schröder, E.; Thonhauser, T.
2009-02-01
Sparse matter is abundant and has both strong local bonds and weak nonbonding forces, in particular nonlocal van der Waals (vdW) forces between atoms separated by empty space. It encompasses a broad spectrum of systems, like soft matter, adsorption systems and biostructures. Density-functional theory (DFT), long since proven successful for dense matter, seems now to have come to a point, where useful extensions to sparse matter are available. In particular, a functional form, vdW-DF (Dion et al 2004 Phys. Rev. Lett. 92 246401; Thonhauser et al 2007 Phys. Rev. B 76 125112), has been proposed for the nonlocal correlations between electrons and applied to various relevant molecules and materials, including to those layered systems like graphite, boron nitride and molybdenum sulfide, to dimers of benzene, polycyclic aromatic hydrocarbons (PAHs), doped benzene, cytosine and DNA base pairs, to nonbonding forces in molecules, to adsorbed molecules, like benzene, naphthalene, phenol and adenine on graphite, alumina and metals, to polymer and carbon nanotube (CNT) crystals, and hydrogen storage in graphite and metal-organic frameworks (MOFs), and to the structure of DNA and of DNA with intercalators. Comparison with results from wavefunction calculations for the smaller systems and with experimental data for the extended ones show the vdW-DF path to be promising. This could have great ramifications.
Fast Sparse Level Sets on Graphics Hardware.
Jalba, Andrei C; van der Laan, Wladimir J; Roerdink, Jos B T M
2013-01-01
The level-set method is one of the most popular techniques for capturing and tracking deformable interfaces. Although level sets have demonstrated great potential in visualization and computer graphics applications, such as surface editing and physically based modeling, their use for interactive simulations has been limited due to the high computational demands involved. In this paper, we address this computational challenge by leveraging the increased computing power of graphics processors, to achieve fast simulations based on level sets. Our efficient, sparse GPU level-set method is substantially faster than other state-of-the-art, parallel approaches on both CPU and GPU hardware. We further investigate its performance through a method for surface reconstruction, based on GPU level sets. Our novel multiresolution method for surface reconstruction from unorganized point clouds compares favorably with recent, existing techniques and other parallel implementations. Finally, we point out that both level-set computations and rendering of level-set surfaces can be performed at interactive rates, even on large volumetric grids. Therefore, many applications based on level sets can benefit from our sparse level-set method.
Topological sparse learning of dynamic form patterns.
Guthier, T; Willert, V; Eggert, J
2015-01-01
Motion is a crucial source of information for a variety of tasks in social interactions. The process of how humans recognize complex articulated movements such as gestures or face expressions remains largely unclear. There is an ongoing discussion if and how explicit low-level motion information, such as optical flow, is involved in the recognition process. Motivated by this discussion, we introduce a computational model that classifies the spatial configuration of gradient and optical flow patterns. The patterns are learned with an unsupervised learning algorithm based on translation-invariant nonnegative sparse coding called VNMF that extracts prototypical optical flow patterns shaped, for example, as moving heads or limb parts. A key element of the proposed system is a lateral inhibition term that suppresses activations of competing patterns in the learning process, leading to a low number of dominant and topological sparse activations. We analyze the classification performance of the gradient and optical flow patterns on three real-world human action recognition and one face expression recognition data set. The results indicate that the recognition of human actions can be achieved by gradient patterns alone, but adding optical flow patterns increases the classification performance. The combined patterns outperform other biological-inspired models and are competitive with current computer vision approaches. PMID:25248088
Inferring sparse networks for noisy transient processes.
Tran, Hoang M; Bukkapatnam, Satish T S
2016-01-01
Inferring causal structures of real world complex networks from measured time series signals remains an open issue. The current approaches are inadequate to discern between direct versus indirect influences (i.e., the presence or absence of a directed arc connecting two nodes) in the presence of noise, sparse interactions, as well as nonlinear and transient dynamics of real world processes. We report a sparse regression (referred to as the l1-min) approach with theoretical bounds on the constraints on the allowable perturbation to recover the network structure that guarantees sparsity and robustness to noise. We also introduce averaging and perturbation procedures to further enhance prediction scores (i.e., reduce inference errors), and the numerical stability of l1-min approach. Extensive investigations have been conducted with multiple benchmark simulated genetic regulatory network and Michaelis-Menten dynamics, as well as real world data sets from DREAM5 challenge. These investigations suggest that our approach can significantly improve, oftentimes by 5 orders of magnitude over the methods reported previously for inferring the structure of dynamic networks, such as Bayesian network, network deconvolution, silencing and modular response analysis methods based on optimizing for sparsity, transients, noise and high dimensionality issues. PMID:26916813
Inferring sparse networks for noisy transient processes
NASA Astrophysics Data System (ADS)
Tran, Hoang M.; Bukkapatnam, Satish T. S.
2016-02-01
Inferring causal structures of real world complex networks from measured time series signals remains an open issue. The current approaches are inadequate to discern between direct versus indirect influences (i.e., the presence or absence of a directed arc connecting two nodes) in the presence of noise, sparse interactions, as well as nonlinear and transient dynamics of real world processes. We report a sparse regression (referred to as the -min) approach with theoretical bounds on the constraints on the allowable perturbation to recover the network structure that guarantees sparsity and robustness to noise. We also introduce averaging and perturbation procedures to further enhance prediction scores (i.e., reduce inference errors), and the numerical stability of -min approach. Extensive investigations have been conducted with multiple benchmark simulated genetic regulatory network and Michaelis-Menten dynamics, as well as real world data sets from DREAM5 challenge. These investigations suggest that our approach can significantly improve, oftentimes by 5 orders of magnitude over the methods reported previously for inferring the structure of dynamic networks, such as Bayesian network, network deconvolution, silencing and modular response analysis methods based on optimizing for sparsity, transients, noise and high dimensionality issues.
Sparse aperture mask wavefront sensor testbed results
NASA Astrophysics Data System (ADS)
Subedi, Hari; Zimmerman, Neil T.; Kasdin, N. Jeremy; Riggs, A. J. E.
2016-07-01
Coronagraphic exoplanet detection at very high contrast requires the estimation and control of low-order wave- front aberrations. At Princeton High Contrast Imaging Lab (PHCIL), we are working on a new technique that integrates a sparse-aperture mask (SAM) with a shaped pupil coronagraph (SPC) to make precise estimates of these low-order aberrations. We collect the starlight rejected from the coronagraphic image plane and interfere it using a sparse aperture mask (SAM) at the relay pupil to estimate the low-order aberrations. In our previous work we numerically demonstrated the efficacy of the technique, and proposed a method to sense and control these differential aberrations in broadband light. We also presented early testbed results in which the SAM was used to sense pointing errors. In this paper, we will briefly overview the SAM wavefront sensor technique, explain the design of the completed testbed, and report the experimental estimation results of the dominant low-order aberrations such as tip/tit, astigmatism and focus.
OSKI: A Library of Automatically Tuned Sparse Matrix Kernels
Vuduc, R; Demmel, J W; Yelick, K A
2005-07-19
The Optimized Sparse Kernel Interface (OSKI) is a collection of low-level primitives that provide automatically tuned computational kernels on sparse matrices, for use by solver libraries and applications. These kernels include sparse matrix-vector multiply and sparse triangular solve, among others. The primary aim of this interface is to hide the complex decision-making process needed to tune the performance of a kernel implementation for a particular user's sparse matrix and machine, while also exposing the steps and potentially non-trivial costs of tuning at run-time. This paper provides an overview of OSKI, which is based on our research on automatically tuned sparse kernels for modern cache-based superscalar machines.
Feature Selection and Pedestrian Detection Based on Sparse Representation.
Yao, Shihong; Wang, Tao; Shen, Weiming; Pan, Shaoming; Chong, Yanwen; Ding, Fei
2015-01-01
Pedestrian detection have been currently devoted to the extraction of effective pedestrian features, which has become one of the obstacles in pedestrian detection application according to the variety of pedestrian features and their large dimension. Based on the theoretical analysis of six frequently-used features, SIFT, SURF, Haar, HOG, LBP and LSS, and their comparison with experimental results, this paper screens out the sparse feature subsets via sparse representation to investigate whether the sparse subsets have the same description abilities and the most stable features. When any two of the six features are fused, the fusion feature is sparsely represented to obtain its important components. Sparse subsets of the fusion features can be rapidly generated by avoiding calculation of the corresponding index of dimension numbers of these feature descriptors; thus, the calculation speed of the feature dimension reduction is improved and the pedestrian detection time is reduced. Experimental results show that sparse feature subsets are capable of keeping the important components of these six feature descriptors. The sparse features of HOG and LSS possess the same description ability and consume less time compared with their full features. The ratios of the sparse feature subsets of HOG and LSS to their full sets are the highest among the six, and thus these two features can be used to best describe the characteristics of the pedestrian and the sparse feature subsets of the combination of HOG-LSS show better distinguishing ability and parsimony. PMID:26295480
Feature Selection and Pedestrian Detection Based on Sparse Representation
Yao, Shihong; Wang, Tao; Shen, Weiming; Pan, Shaoming; Chong, Yanwen; Ding, Fei
2015-01-01
Pedestrian detection have been currently devoted to the extraction of effective pedestrian features, which has become one of the obstacles in pedestrian detection application according to the variety of pedestrian features and their large dimension. Based on the theoretical analysis of six frequently-used features, SIFT, SURF, Haar, HOG, LBP and LSS, and their comparison with experimental results, this paper screens out the sparse feature subsets via sparse representation to investigate whether the sparse subsets have the same description abilities and the most stable features. When any two of the six features are fused, the fusion feature is sparsely represented to obtain its important components. Sparse subsets of the fusion features can be rapidly generated by avoiding calculation of the corresponding index of dimension numbers of these feature descriptors; thus, the calculation speed of the feature dimension reduction is improved and the pedestrian detection time is reduced. Experimental results show that sparse feature subsets are capable of keeping the important components of these six feature descriptors. The sparse features of HOG and LSS possess the same description ability and consume less time compared with their full features. The ratios of the sparse feature subsets of HOG and LSS to their full sets are the highest among the six, and thus these two features can be used to best describe the characteristics of the pedestrian and the sparse feature subsets of the combination of HOG-LSS show better distinguishing ability and parsimony. PMID:26295480
Dose-shaping using targeted sparse optimization
Sayre, George A.; Ruan, Dan
2013-07-15
Purpose: Dose volume histograms (DVHs) are common tools in radiation therapy treatment planning to characterize plan quality. As statistical metrics, DVHs provide a compact summary of the underlying plan at the cost of losing spatial information: the same or similar dose-volume histograms can arise from substantially different spatial dose maps. This is exactly the reason why physicians and physicists scrutinize dose maps even after they satisfy all DVH endpoints numerically. However, up to this point, little has been done to control spatial phenomena, such as the spatial distribution of hot spots, which has significant clinical implications. To this end, the authors propose a novel objective function that enables a more direct tradeoff between target coverage, organ-sparing, and planning target volume (PTV) homogeneity, and presents our findings from four prostate cases, a pancreas case, and a head-and-neck case to illustrate the advantages and general applicability of our method.Methods: In designing the energy minimization objective (E{sub tot}{sup sparse}), the authors utilized the following robust cost functions: (1) an asymmetric linear well function to allow differential penalties for underdose, relaxation of prescription dose, and overdose in the PTV; (2) a two-piece linear function to heavily penalize high dose and mildly penalize low and intermediate dose in organs-at risk (OARs); and (3) a total variation energy, i.e., the L{sub 1} norm applied to the first-order approximation of the dose gradient in the PTV. By minimizing a weighted sum of these robust costs, general conformity to dose prescription and dose-gradient prescription is achieved while encouraging prescription violations to follow a Laplace distribution. In contrast, conventional quadratic objectives are associated with a Gaussian distribution of violations, which is less forgiving to large violations of prescription than the Laplace distribution. As a result, the proposed objective E{sub tot
Predicting structure in nonsymmetric sparse matrix factorizations
Gilbert, J.R.; Ng, E.G.
1992-10-01
Many computations on sparse matrices have a phase that predicts the nonzero structure of the output, followed by a phase that actually performs the numerical computation. We study structure prediction for computations that involve nonsymmetric row and column permutations and nonsymmetric or non-square matrices. Our tools are bipartite graphs, matchings, and alternating paths. Our main new result concerns LU factorization with partial pivoting. We show that if a square matrix A has the strong Hall property (i.e., is fully indecomposable) then an upper bound due to George and Ng on the nonzero structure of L + U is as tight as possible. To show this, we prove a crucial result about alternating paths in strong Hall graphs. The alternating-paths theorem seems to be of independent interest: it can also be used to prove related results about structure prediction for QR factorization that are due to Coleman, Edenbrandt, Gilbert, Hare, Johnson, Olesky, Pothen, and van den Driessche.
ESTIMATION OF FUNCTIONALS OF SPARSE COVARIANCE MATRICES
Fan, Jianqing; Rigollet, Philippe; Wang, Weichen
2016-01-01
High-dimensional statistical tests often ignore correlations to gain simplicity and stability leading to null distributions that depend on functionals of correlation matrices such as their Frobenius norm and other ℓr norms. Motivated by the computation of critical values of such tests, we investigate the difficulty of estimation the functionals of sparse correlation matrices. Specifically, we show that simple plug-in procedures based on thresholded estimators of correlation matrices are sparsity-adaptive and minimax optimal over a large class of correlation matrices. Akin to previous results on functional estimation, the minimax rates exhibit an elbow phenomenon. Our results are further illustrated in simulated data as well as an empirical study of data arising in financial econometrics. PMID:26806986
Sparse distributed memory and related models
NASA Technical Reports Server (NTRS)
Kanerva, Pentti
1992-01-01
Described here is sparse distributed memory (SDM) as a neural-net associative memory. It is characterized by two weight matrices and by a large internal dimension - the number of hidden units is much larger than the number of input or output units. The first matrix, A, is fixed and possibly random, and the second matrix, C, is modifiable. The SDM is compared and contrasted to (1) computer memory, (2) correlation-matrix memory, (3) feet-forward artificial neural network, (4) cortex of the cerebellum, (5) Marr and Albus models of the cerebellum, and (6) Albus' cerebellar model arithmetic computer (CMAC). Several variations of the basic SDM design are discussed: the selected-coordinate and hyperplane designs of Jaeckel, the pseudorandom associative neural memory of Hassoun, and SDM with real-valued input variables by Prager and Fallside. SDM research conducted mainly at the Research Institute for Advanced Computer Science (RIACS) in 1986-1991 is highlighted.
Eigensolver for a Sparse, Large Hermitian Matrix
NASA Technical Reports Server (NTRS)
Tisdale, E. Robert; Oyafuso, Fabiano; Klimeck, Gerhard; Brown, R. Chris
2003-01-01
A parallel-processing computer program finds a few eigenvalues in a sparse Hermitian matrix that contains as many as 100 million diagonal elements. This program finds the eigenvalues faster, using less memory, than do other, comparable eigensolver programs. This program implements a Lanczos algorithm in the American National Standards Institute/ International Organization for Standardization (ANSI/ISO) C computing language, using the Message Passing Interface (MPI) standard to complement an eigensolver in PARPACK. [PARPACK (Parallel Arnoldi Package) is an extension, to parallel-processing computer architectures, of ARPACK (Arnoldi Package), which is a collection of Fortran 77 subroutines that solve large-scale eigenvalue problems.] The eigensolver runs on Beowulf clusters of computers at the Jet Propulsion Laboratory (JPL).
Compressed Sensing for Reconstructing Sparse Quantum States
NASA Astrophysics Data System (ADS)
Rudinger, Kenneth; Joynt, Robert
2014-03-01
Compressed sensing techniques have been successfully applied to quantum state tomography, enabling the efficient determination of states that are nearly pure, i.e, of low rank. We show how compressed sensing may be used even when the states to be reconstructed are full rank. Instead, the necessary requirement is that the states be sparse in some known basis (e.g. the Pauli basis). Physical systems at high temperatures in thermal equilibrium are important examples of such states. Using this method, we are able to demonstrate that, like for classical signals, compressed sensing for quantum states exhibits the Donoho-Tanner phase transition. This method will be useful for the determination of the Hamiltonians of artificially constructed quantum systems whose purpose is to simulate condensed-matter models, as it requires many fewer measurements than demanded by standard tomographic procedures. This work was supported in part by ARO, DOD (W911NF-09-1-0439) and NSF (CCR-0635355).
Inverse lithography using sparse mask representations
NASA Astrophysics Data System (ADS)
Ionescu, Radu C.; Hurley, Paul; Apostol, Stefan
2015-03-01
We present a novel optimisation algorithm for inverse lithography, based on optimization of the mask derivative, a domain inherently sparse, and for rectilinear polygons, invertible. The method is first developed assuming a point light source, and then extended to general incoherent sources. What results is a fast algorithm, producing manufacturable masks (the search space is constrained to rectilinear polygons), and flexible (specific constraints such as minimal line widths can be imposed). One inherent trick is to treat polygons as continuous entities, thus making aerial image calculation extremely fast and accurate. Requirements for mask manufacturability can be integrated in the optimization without too much added complexity. We also explain how to extend the scheme for phase-changing mask optimization.
Blind source separation by sparse decomposition
NASA Astrophysics Data System (ADS)
Zibulevsky, Michael; Pearlmutter, Barak A.
2000-04-01
The blind source separation problem is to extract the underlying source signals from a set of their linear mixtures, where the mixing matrix is unknown. This situation is common, eg in acoustics, radio, and medical signal processing. We exploit the property of the sources to have a sparse representation in a corresponding signal dictionary. Such a dictionary may consist of wavelets, wavelet packets, etc., or be obtained by learning from a given family of signals. Starting from the maximum a posteriori framework, which is applicable to the case of more sources than mixtures, we derive a few other categories of objective functions, which provide faster and more robust computations, when there are an equal number of sources and mixtures. Our experiments with artificial signals and with musical sounds demonstrate significantly better separation than other known techniques.
Fast generation of sparse random kernel graphs
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in time at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.
Fast generation of sparse random kernel graphs
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in timemore » at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.« less
Integer sparse distributed memory: analysis and results.
Snaider, Javier; Franklin, Stan; Strain, Steve; George, E Olusegun
2013-10-01
Sparse distributed memory is an auto-associative memory system that stores high dimensional Boolean vectors. Here we present an extension of the original SDM, the Integer SDM that uses modular arithmetic integer vectors rather than binary vectors. This extension preserves many of the desirable properties of the original SDM: auto-associativity, content addressability, distributed storage, and robustness over noisy inputs. In addition, it improves the representation capabilities of the memory and is more robust over normalization. It can also be extended to support forgetting and reliable sequence storage. We performed several simulations that test the noise robustness property and capacity of the memory. Theoretical analyses of the memory's fidelity and capacity are also presented. PMID:23747569
Convolutional Sparse Coding for Trajectory Reconstruction.
Zhu, Yingying; Lucey, Simon
2015-03-01
Trajectory basis Non-Rigid Structure from Motion (NRSfM) refers to the process of reconstructing the 3D trajectory of each point of a non-rigid object from just their 2D projected trajectories. Reconstruction relies on two factors: (i) the condition of the composed camera & trajectory basis matrix, and (ii) whether the trajectory basis has enough degrees of freedom to model the 3D point trajectory. These two factors are inherently conflicting. Employing a trajectory basis with small capacity has the positive characteristic of reducing the likelihood of an ill-conditioned system (when composed with the camera) during reconstruction. However, this has the negative characteristic of increasing the likelihood that the basis will not be able to fully model the object's "true" 3D point trajectories. In this paper we draw upon a well known result centering around the Reduced Isometry Property (RIP) condition for sparse signal reconstruction. RIP allow us to relax the requirement that the full trajectory basis composed with the camera matrix must be well conditioned. Further, we propose a strategy for learning an over-complete basis using convolutional sparse coding from naturally occurring point trajectory corpora to increase the likelihood that the RIP condition holds for a broad class of point trajectories and camera motions. Finally, we propose an l1 inspired objective for trajectory reconstruction that is able to "adaptively" select the smallest sub-matrix from an over-complete trajectory basis that balances (i) and (ii). We present more practical 3D reconstruction results compared to current state of the art in trajectory basis NRSfM.
Miniature Laboratory for Detecting Sparse Biomolecules
NASA Technical Reports Server (NTRS)
Lin, Ying; Yu, Nan
2005-01-01
A miniature laboratory system has been proposed for use in the field to detect sparsely distributed biomolecules. By emphasizing concentration and sorting of specimens prior to detection, the underlying system concept would make it possible to attain high detection sensitivities without the need to develop ever more sensitive biosensors. The original purpose of the proposal is to aid the search for signs of life on a remote planet by enabling the detection of specimens as sparse as a few molecules or microbes in a large amount of soil, dust, rocks, water/ice, or other raw sample material. Some version of the system could prove useful on Earth for remote sensing of biological contamination, including agents of biological warfare. Processing in this system would begin with dissolution of the raw sample material in a sample-separation vessel. The solution in the vessel would contain floating microscopic magnetic beads coated with substances that could engage in chemical reactions with various target functional groups that are parts of target molecules. The chemical reactions would cause the targeted molecules to be captured on the surfaces of the beads. By use of a controlled magnetic field, the beads would be concentrated in a specified location in the vessel. Once the beads were thus concentrated, the rest of the solution would be discarded. This procedure would obviate the filtration steps and thereby also eliminate the filter-clogging difficulties of typical prior sample-concentration schemes. For ferrous dust/soil samples, the dissolution would be done first in a separate vessel before the solution is transferred to the microbead-containing vessel.
Zhang, Di; He, Jiazhong; Zhao, Yun; Du, Minghui
2015-03-01
In magnetic resonance (MR) imaging, image spatial resolution is determined by various instrumental limitations and physical considerations. This paper presents a new algorithm for producing a high-resolution version of a low-resolution MR image. The proposed method consists of two consecutive steps: (1) reconstructs a high-resolution MR image from a given low-resolution observation via solving a joint sparse representation and nonlocal similarity L1-norm minimization problem; and (2) applies a sparse derivative prior based post-processing to suppress blurring effects. Extensive experiments on simulated brain MR images and two real clinical MR image datasets validate that the proposed method achieves much better results than many state-of-the-art algorithms in terms of both quantitative measures and visual perception.
Group-based sparse representation for image restoration.
Zhang, Jian; Zhao, Debin; Gao, Wen
2014-08-01
Traditional patch-based sparse representation modeling of natural images usually suffer from two problems. First, it has to solve a large-scale optimization problem with high computational complexity in dictionary learning. Second, each patch is considered independently in dictionary learning and sparse coding, which ignores the relationship among patches, resulting in inaccurate sparse coding coefficients. In this paper, instead of using patch as the basic unit of sparse representation, we exploit the concept of group as the basic unit of sparse representation, which is composed of nonlocal patches with similar structures, and establish a novel sparse representation modeling of natural images, called group-based sparse representation (GSR). The proposed GSR is able to sparsely represent natural images in the domain of group, which enforces the intrinsic local sparsity and nonlocal self-similarity of images simultaneously in a unified framework. In addition, an effective self-adaptive dictionary learning method for each group with low complexity is designed, rather than dictionary learning from natural images. To make GSR tractable and robust, a split Bregman-based technique is developed to solve the proposed GSR-driven ℓ0 minimization problem for image restoration efficiently. Extensive experiments on image inpainting, image deblurring and image compressive sensing recovery manifest that the proposed GSR modeling outperforms many current state-of-the-art schemes in both peak signal-to-noise ratio and visual perception.
Cervigram image segmentation based on reconstructive sparse representations
NASA Astrophysics Data System (ADS)
Zhang, Shaoting; Huang, Junzhou; Wang, Wei; Huang, Xiaolei; Metaxas, Dimitris
2010-03-01
We proposed an approach based on reconstructive sparse representations to segment tissues in optical images of the uterine cervix. Because of large variations in image appearance caused by the changing of the illumination and specular reflection, the color and texture features in optical images often overlap with each other and are not linearly separable. By leveraging sparse representations the data can be transformed to higher dimensions with sparse constraints and become more separated. K-SVD algorithm is employed to find sparse representations and corresponding dictionaries. The data can be reconstructed from its sparse representations and positive and/or negative dictionaries. Classification can be achieved based on comparing the reconstructive errors. In the experiments we applied our method to automatically segment the biomarker AcetoWhite (AW) regions in an archive of 60,000 images of the uterine cervix. Compared with other general methods, our approach showed lower space and time complexity and higher sensitivity.
Sparse Coding on Symmetric Positive Definite Manifolds Using Bregman Divergences.
Harandi, Mehrtash T; Hartley, Richard; Lovell, Brian; Sanderson, Conrad
2016-06-01
This paper introduces sparse coding and dictionary learning for symmetric positive definite (SPD) matrices, which are often used in machine learning, computer vision, and related areas. Unlike traditional sparse coding schemes that work in vector spaces, in this paper, we discuss how SPD matrices can be described by sparse combination of dictionary atoms, where the atoms are also SPD matrices. We propose to seek sparse coding by embedding the space of SPD matrices into the Hilbert spaces through two types of the Bregman matrix divergences. This not only leads to an efficient way of performing sparse coding but also an online and iterative scheme for dictionary learning. We apply the proposed methods to several computer vision tasks where images are represented by region covariance matrices. Our proposed algorithms outperform state-of-the-art methods on a wide range of classification tasks, including face recognition, action recognition, material classification, and texture categorization. PMID:25643414
ERIC Educational Resources Information Center
Sadler, Peter G.
The Institute for the Study of Sparsely Populated Areas is a multidisciplinary research unit which acts to coordinate, further, and initiate studies of the economic and social conditions of sparsely populated areas. Short summaries of the eight studies completed in the session of 1977-78 indicate work in such areas as the study of political life…
Sparse distributed memory: Principles and operation
NASA Technical Reports Server (NTRS)
Flynn, M. J.; Kanerva, P.; Bhadkamkar, N.
1989-01-01
Sparse distributed memory is a generalized random access memory (RAM) for long (1000 bit) binary words. Such words can be written into and read from the memory, and they can also be used to address the memory. The main attribute of the memory is sensitivity to similarity, meaning that a word can be read back not only by giving the original write address but also by giving one close to it as measured by the Hamming distance between addresses. Large memories of this kind are expected to have wide use in speech recognition and scene analysis, in signal detection and verification, and in adaptive control of automated equipment, in general, in dealing with real world information in real time. The memory can be realized as a simple, massively parallel computer. Digital technology has reached a point where building large memories is becoming practical. Major design issues were resolved which were faced in building the memories. The design is described of a prototype memory with 256 bit addresses and from 8 to 128 K locations for 256 bit words. A key aspect of the design is extensive use of dynamic RAM and other standard components.
Approximation and compression with sparse orthonormal transforms.
Sezer, Osman Gokhan; Guleryuz, Onur G; Altunbasak, Yucel
2015-08-01
We propose a new transform design method that targets the generation of compression-optimized transforms for next-generation multimedia applications. The fundamental idea behind transform compression is to exploit regularity within signals such that redundancy is minimized subject to a fidelity cost. Multimedia signals, in particular images and video, are well known to contain a diverse set of localized structures, leading to many different types of regularity and to nonstationary signal statistics. The proposed method designs sparse orthonormal transforms (SOTs) that automatically exploit regularity over different signal structures and provides an adaptation method that determines the best representation over localized regions. Unlike earlier work that is motivated by linear approximation constructs and model-based designs that are limited to specific types of signal regularity, our work uses general nonlinear approximation ideas and a data-driven setup to significantly broaden its reach. We show that our SOT designs provide a safe and principled extension of the Karhunen-Loeve transform (KLT) by reducing to the KLT on Gaussian processes and by automatically exploiting non-Gaussian statistics to significantly improve over the KLT on more general processes. We provide an algebraic optimization framework that generates optimized designs for any desired transform structure (multiresolution, block, lapped, and so on) with significantly better n -term approximation performance. For each structure, we propose a new prototype codec and test over a database of images. Simulation results show consistent increase in compression and approximation performance compared with conventional methods. PMID:25823033
Sparse alignment for robust tensor learning.
Lai, Zhihui; Wong, Wai Keung; Xu, Yong; Zhao, Cairong; Sun, Mingming
2014-10-01
Multilinear/tensor extensions of manifold learning based algorithms have been widely used in computer vision and pattern recognition. This paper first provides a systematic analysis of the multilinear extensions for the most popular methods by using alignment techniques, thereby obtaining a general tensor alignment framework. From this framework, it is easy to show that the manifold learning based tensor learning methods are intrinsically different from the alignment techniques. Based on the alignment framework, a robust tensor learning method called sparse tensor alignment (STA) is then proposed for unsupervised tensor feature extraction. Different from the existing tensor learning methods, L1- and L2-norms are introduced to enhance the robustness in the alignment step of the STA. The advantage of the proposed technique is that the difficulty in selecting the size of the local neighborhood can be avoided in the manifold learning based tensor feature extraction algorithms. Although STA is an unsupervised learning method, the sparsity encodes the discriminative information in the alignment step and provides the robustness of STA. Extensive experiments on the well-known image databases as well as action and hand gesture databases by encoding object images as tensors demonstrate that the proposed STA algorithm gives the most competitive performance when compared with the tensor-based unsupervised learning methods. PMID:25291733
Optimal parallel solution of sparse triangular systems
NASA Technical Reports Server (NTRS)
Alvarado, Fernando L.; Schreiber, Robert
1990-01-01
A method for the parallel solution of triangular sets of equations is described that is appropriate when there are many right-handed sides. By preprocessing, the method can reduce the number of parallel steps required to solve Lx = b compared to parallel forward or backsolve. Applications are to iterative solvers with triangular preconditioners, to structural analysis, or to power systems applications, where there may be many right-handed sides (not all available a priori). The inverse of L is represented as a product of sparse triangular factors. The problem is to find a factored representation of this inverse of L with the smallest number of factors (or partitions), subject to the requirement that no new nonzero elements be created in the formation of these inverse factors. A method from an earlier reference is shown to solve this problem. This method is improved upon by constructing a permutation of the rows and columns of L that preserves triangularity and allow for the best possible such partition. A number of practical examples and algorithmic details are presented. The parallelism attainable is illustrated by means of elimination trees and clique trees.
Visual Exploration of Sparse Traffic Trajectory Data.
Wang, Zuchao; Ye, Tangzhi; Lu, Min; Yuan, Xiaoru; Qu, Huamin; Yuan, Jacky; Wu, Qianliang
2014-12-01
In this paper, we present a visual analysis system to explore sparse traffic trajectory data recorded by transportation cells. Such data contains the movements of nearly all moving vehicles on the major roads of a city. Therefore it is very suitable for macro-traffic analysis. However, the vehicle movements are recorded only when they pass through the cells. The exact tracks between two consecutive cells are unknown. To deal with such uncertainties, we first design a local animation, showing the vehicle movements only in the vicinity of cells. Besides, we ignore the micro-behaviors of individual vehicles, and focus on the macro-traffic patterns. We apply existing trajectory aggregation techniques to the dataset, studying cell status pattern and inter-cell flow pattern. Beyond that, we propose to study the correlation between these two patterns with dynamic graph visualization techniques. It allows us to check how traffic congestion on one cell is correlated with traffic flows on neighbouring links, and with route selection in its neighbourhood. Case studies show the effectiveness of our system. PMID:26356895
Sparse coding for layered neural networks
NASA Astrophysics Data System (ADS)
Katayama, Katsuki; Sakata, Yasuo; Horiguchi, Tsuyoshi
2002-07-01
We investigate storage capacity of two types of fully connected layered neural networks with sparse coding when binary patterns are embedded into the networks by a Hebbian learning rule. One of them is a layered network, in which a transfer function of even layers is different from that of odd layers. The other is a layered network with intra-layer connections, in which the transfer function of inter-layer is different from that of intra-layer, and inter-layered neurons and intra-layered neurons are updated alternately. We derive recursion relations for order parameters by means of the signal-to-noise ratio method, and then apply the self-control threshold method proposed by Dominguez and Bollé to both layered networks with monotonic transfer functions. We find that a critical value αC of storage capacity is about 0.11|a ln a| -1 ( a≪1) for both layered networks, where a is a neuronal activity. It turns out that the basin of attraction is larger for both layered networks when the self-control threshold method is applied.
Partitioning sparse matrices with eigenvectors of graphs
NASA Technical Reports Server (NTRS)
Pothen, Alex; Simon, Horst D.; Liou, Kang-Pu
1990-01-01
The problem of computing a small vertex separator in a graph arises in the context of computing a good ordering for the parallel factorization of sparse, symmetric matrices. An algebraic approach for computing vertex separators is considered in this paper. It is shown that lower bounds on separator sizes can be obtained in terms of the eigenvalues of the Laplacian matrix associated with a graph. The Laplacian eigenvectors of grid graphs can be computed from Kronecker products involving the eigenvectors of path graphs, and these eigenvectors can be used to compute good separators in grid graphs. A heuristic algorithm is designed to compute a vertex separator in a general graph by first computing an edge separator in the graph from an eigenvector of the Laplacian matrix, and then using a maximum matching in a subgraph to compute the vertex separator. Results on the quality of the separators computed by the spectral algorithm are presented, and these are compared with separators obtained from other algorithms for computing separators. Finally, the time required to compute the Laplacian eigenvector is reported, and the accuracy with which the eigenvector must be computed to obtain good separators is considered. The spectral algorithm has the advantage that it can be implemented on a medium-size multiprocessor in a straightforward manner.
Transformer fault diagnosis using continuous sparse autoencoder.
Wang, Lukun; Zhao, Xiaoying; Pei, Jiangnan; Tang, Gongyou
2016-01-01
This paper proposes a novel continuous sparse autoencoder (CSAE) which can be used in unsupervised feature learning. The CSAE adds Gaussian stochastic unit into activation function to extract features of nonlinear data. In this paper, CSAE is applied to solve the problem of transformer fault recognition. Firstly, based on dissolved gas analysis method, IEC three ratios are calculated by the concentrations of dissolved gases. Then IEC three ratios data is normalized to reduce data singularity and improve training speed. Secondly, deep belief network is established by two layers of CSAE and one layer of back propagation (BP) network. Thirdly, CSAE is adopted to unsupervised training and getting features. Then BP network is used for supervised training and getting transformer fault. Finally, the experimental data from IEC TC 10 dataset aims to illustrate the effectiveness of the presented approach. Comparative experiments clearly show that CSAE can extract features from the original data, and achieve a superior correct differentiation rate on transformer fault diagnosis.
Index statistical properties of sparse random graphs
NASA Astrophysics Data System (ADS)
Metz, F. L.; Stariolo, Daniel A.
2015-10-01
Using the replica method, we develop an analytical approach to compute the characteristic function for the probability PN(K ,λ ) that a large N ×N adjacency matrix of sparse random graphs has K eigenvalues below a threshold λ . The method allows to determine, in principle, all moments of PN(K ,λ ) , from which the typical sample-to-sample fluctuations can be fully characterized. For random graph models with localized eigenvectors, we show that the index variance scales linearly with N ≫1 for |λ |>0 , with a model-dependent prefactor that can be exactly calculated. Explicit results are discussed for Erdös-Rényi and regular random graphs, both exhibiting a prefactor with a nonmonotonic behavior as a function of λ . These results contrast with rotationally invariant random matrices, where the index variance scales only as lnN , with an universal prefactor that is independent of λ . Numerical diagonalization results confirm the exactness of our approach and, in addition, strongly support the Gaussian nature of the index fluctuations.
Sparse distributed memory prototype: Principles of operation
NASA Technical Reports Server (NTRS)
Flynn, Michael J.; Kanerva, Pentti; Ahanin, Bahram; Bhadkamkar, Neal; Flaherty, Paul; Hickey, Philip
1988-01-01
Sparse distributed memory is a generalized random access memory (RAM) for long binary words. Such words can be written into and read from the memory, and they can be used to address the memory. The main attribute of the memory is sensitivity to similarity, meaning that a word can be read back not only by giving the original right address but also by giving one close to it as measured by the Hamming distance between addresses. Large memories of this kind are expected to have wide use in speech and scene analysis, in signal detection and verification, and in adaptive control of automated equipment. The memory can be realized as a simple, massively parallel computer. Digital technology has reached a point where building large memories is becoming practical. The research is aimed at resolving major design issues that have to be faced in building the memories. The design of a prototype memory with 256-bit addresses and from 8K to 128K locations for 256-bit words is described. A key aspect of the design is extensive use of dynamic RAM and other standard components.
Optimized design and analysis of sparse-sampling FMRI experiments.
Perrachione, Tyler K; Ghosh, Satrajit S
2013-01-01
Sparse-sampling is an important methodological advance in functional magnetic resonance imaging (fMRI), in which silent delays are introduced between MR volume acquisitions, allowing for the presentation of auditory stimuli without contamination by acoustic scanner noise and for overt vocal responses without motion-induced artifacts in the functional time series. As such, the sparse-sampling technique has become a mainstay of principled fMRI research into the cognitive and systems neuroscience of speech, language, hearing, and music. Despite being in use for over a decade, there has been little systematic investigation of the acquisition parameters, experimental design considerations, and statistical analysis approaches that bear on the results and interpretation of sparse-sampling fMRI experiments. In this report, we examined how design and analysis choices related to the duration of repetition time (TR) delay (an acquisition parameter), stimulation rate (an experimental design parameter), and model basis function (an analysis parameter) act independently and interactively to affect the neural activation profiles observed in fMRI. First, we conducted a series of computational simulations to explore the parameter space of sparse design and analysis with respect to these variables; second, we validated the results of these simulations in a series of sparse-sampling fMRI experiments. Overall, these experiments suggest the employment of three methodological approaches that can, in many situations, substantially improve the detection of neurophysiological response in sparse fMRI: (1) Sparse analyses should utilize a physiologically informed model that incorporates hemodynamic response convolution to reduce model error. (2) The design of sparse fMRI experiments should maintain a high rate of stimulus presentation to maximize effect size. (3) TR delays of short to intermediate length can be used between acquisitions of sparse-sampled functional image volumes to increase
Ward, Andy L.
2007-11-26
Fluor Hanford (FH) is designing and assessing the performance of engineered barriers for final closure of 200-UW-1 waste sites. Engineered barriers must minimize the intrusion and water, plants and animals into the underlying waste to provide protection for human health and the environment. The Pacific Northwest National Laboratory (PNNL) developed Subsurface Transport Over Multiple Phases (STOMP) simulator is being used to optimize the performance of candidate barriers. Simulating barrier performance involves computation of mass and energy transfer within a soil-atmosphere-vegetation continuum and requires a variety of input parameters, some of which are more readily available than others. Required input includes parameter values for the geotechnical, physical, hydraulic, and thermal properties of the materials comprising the barrier and the structural fill on which it will be constructed as well as parameters to allow simulation of plant effects. This report provides a data package of the required parameters as well as the technical basis, rationale and methodology used to obtain the parameter values.
New Asteroid Models Based on Combined Dense and Sparse Photometry
NASA Astrophysics Data System (ADS)
Hanuš, Josef; Durech, J.
2010-10-01
For thousands of asteroids we investigated several ten thousands of sparse photometric data from astrometric projects. These data are available on AstDyS server (Asteroids -- Dynamic Site, http://hamilton.dm.unipi.it). We picked 7 astrometric surveys and used their calibrated photometry in lightcurve inversion method for determination of asteroid's convex shapes and rotational states. We present nearly 100 new asteroid models derived from combined dense and sparse data sets, where sparse photometry is taken from AstDyS server and dense lightcurves are from the Uppsala Asteroid Photometric Catalogue (UAPC) and from several individual observers.
Electromagnetic Formation Flight (EMFF) for Sparse Aperture Arrays
NASA Technical Reports Server (NTRS)
Kwon, Daniel W.; Miller, David W.; Sedwick, Raymond J.
2004-01-01
Traditional methods of actuating spacecraft in sparse aperture arrays use propellant as a reaction mass. For formation flying systems, propellant becomes a critical consumable which can be quickly exhausted while maintaining relative orientation. Additional problems posed by propellant include optical contamination, plume impingement, thermal emission, and vibration excitation. For these missions where control of relative degrees of freedom is important, we consider using a system of electromagnets, in concert with reaction wheels, to replace the consumables. Electromagnetic Formation Flight sparse apertures, powered by solar energy, are designed differently from traditional propulsion systems, which are based on V. This paper investigates the design of sparse apertures both inside and outside the Earth's gravity field.
SPARSKIT: A basic tool kit for sparse matrix computations
NASA Technical Reports Server (NTRS)
Saad, Youcef
1990-01-01
Presented here are the main features of a tool package for manipulating and working with sparse matrices. One of the goals of the package is to provide basic tools to facilitate the exchange of software and data between researchers in sparse matrix computations. The starting point is the Harwell/Boeing collection of matrices for which the authors provide a number of tools. Among other things, the package provides programs for converting data structures, printing simple statistics on a matrix, plotting a matrix profile, and performing linear algebra operations with sparse matrices.
Accelerating Dynamic Cardiac MR Imaging Using Structured Sparse Representation
Cai, Nian; Wang, Shengru; Zhu, Shasha
2013-01-01
Compressed sensing (CS) has produced promising results on dynamic cardiac MR imaging by exploiting the sparsity in image series. In this paper, we propose a new method to improve the CS reconstruction for dynamic cardiac MRI based on the theory of structured sparse representation. The proposed method user the PCA subdictionaries for adaptive sparse representation and suppresses the sparse coding noise to obtain good reconstructions. An accelerated iterative shrinkage algorithm is used to solve the optimization problem and achieve a fast convergence rate. Experimental results demonstrate that the proposed method improves the reconstruction quality of dynamic cardiac cine MRI over the state-of-the-art CS method. PMID:24454528
Method and apparatus for distinguishing actual sparse events from sparse event false alarms
Spalding, Richard E.; Grotbeck, Carter L.
2000-01-01
Remote sensing method and apparatus wherein sparse optical events are distinguished from false events. "Ghost" images of actual optical phenomena are generated using an optical beam splitter and optics configured to direct split beams to a single sensor or segmented sensor. True optical signals are distinguished from false signals or noise based on whether the ghost image is presence or absent. The invention obviates the need for dual sensor systems to effect a false target detection capability, thus significantly reducing system complexity and cost.
Classification of vegetation types in military region
NASA Astrophysics Data System (ADS)
Gonçalves, Miguel; Silva, Jose Silvestre; Bioucas-Dias, Jose
2015-10-01
In decision-making process regarding planning and execution of military operations, the terrain is a determining factor. Aerial photographs are a source of vital information for the success of an operation in hostile region, namely when the cartographic information behind enemy lines is scarce or non-existent. The objective of present work is the development of a tool capable of processing aerial photos. The methodology implemented starts with feature extraction, followed by the application of an automatic selector of features. The next step, using the k-fold cross validation technique, estimates the input parameters for the following classifiers: Sparse Multinomial Logist Regression (SMLR), K Nearest Neighbor (KNN), Linear Classifier using Principal Component Expansion on the Joint Data (PCLDC) and Multi-Class Support Vector Machine (MSVM). These classifiers were used in two different studies with distinct objectives: discrimination of vegetation's density and identification of vegetation's main components. It was found that the best classifier on the first approach is the Sparse Logistic Multinomial Regression (SMLR). On the second approach, the implemented methodology applied to high resolution images showed that the better performance was achieved by KNN classifier and PCLDC. Comparing the two approaches there is a multiscale issue, in which for different resolutions, the best solution to the problem requires different classifiers and the extraction of different features.
Finding One Community in a Sparse Graph
NASA Astrophysics Data System (ADS)
Montanari, Andrea
2015-10-01
We consider a random sparse graph with bounded average degree, in which a subset of vertices has higher connectivity than the background. In particular, the average degree inside this subset of vertices is larger than outside (but still bounded). Given a realization of such graph, we aim at identifying the hidden subset of vertices. This can be regarded as a model for the problem of finding a tightly knitted community in a social network, or a cluster in a relational dataset. In this paper we present two sets of contributions: ( i) We use the cavity method from spin glass theory to derive an exact phase diagram for the reconstruction problem. In particular, as the difference in edge probability increases, the problem undergoes two phase transitions, a static phase transition and a dynamic one. ( ii) We establish rigorous bounds on the dynamic phase transition and prove that, above a certain threshold, a local algorithm (belief propagation) correctly identify most of the hidden set. Below the same threshold no local algorithm can achieve this goal. However, in this regime the subset can be identified by exhaustive search. For small hidden sets and large average degree, the phase transition for local algorithms takes an intriguingly simple form. Local algorithms succeed with high probability for deg _in - deg _out > √{deg _out/e} and fail for deg _in - deg _out < √{deg _out/e} (with deg _in, deg _out the average degrees inside and outside the community). We argue that spectral algorithms are also ineffective in the latter regime. It is an open problem whether any polynomial time algorithms might succeed for deg _in - deg _out < √{deg _out/e}.
Sparse source configurations for asteroid tomography
NASA Astrophysics Data System (ADS)
Pursiainen, S.; Kaasalainen, M.
2014-04-01
The objective of our recent research has been to develop non-invasive imaging techniques for future planetary research and mining activities involving a challenging in situ environment and tight payload limits [1]. This presentation will deal in particular with an approach in which the internal relative permittivity ∈r or the refractive index n = √ ∈r of an asteroid is to be recovered based on radio signal transmitted by a sparse set [2] of fixed or movable landers. To address important aspects of mission planning, we have analyzed different signal source configurations to find the minimal number of source positions needed for robust localization of anomalies, such as internal voids. Characteristic to this inverse problem are the large relative changes in signal speed caused by the high permittivity of typical asteroid minerals (e.g. basalt), leading to strong refractions and reflections of the signal. Finding an appropriate problemspecific signaling arrangement is an important premission goal for successful in situ measurements. This presentation will include inversion results obtained with laboratory-recorded travel time data y of the form in which n δ denotes a perturbation of a refractive index n = n δ + nbg; gi estimates the total noise due to different error sources; (ybg)i = ∫Ci nbg ds is an entry of noiseless background data ybg; and Ci is a signal path. Also simulated time-evolution data will be covered with respect to potential u satisfying the wave equation ∈rδ2/δt2+ ōδu/δt-∆u = f, where ō is a (latent) conductivity distribution and f is a source term. Special interest will be paid to inversion robustness regarding changes of the prior model and source positioning. Among other things, our analysis suggests that strongly refractive anomalies can be detected with three or four sources independently of their positioning.
Multiple kernel learning for sparse representation-based classification.
Shrivastava, Ashish; Patel, Vishal M; Chellappa, Rama
2014-07-01
In this paper, we propose a multiple kernel learning (MKL) algorithm that is based on the sparse representation-based classification (SRC) method. Taking advantage of the nonlinear kernel SRC in efficiently representing the nonlinearities in the high-dimensional feature space, we propose an MKL method based on the kernel alignment criteria. Our method uses a two step training method to learn the kernel weights and sparse codes. At each iteration, the sparse codes are updated first while fixing the kernel mixing coefficients, and then the kernel mixing coefficients are updated while fixing the sparse codes. These two steps are repeated until a stopping criteria is met. The effectiveness of the proposed method is demonstrated using several publicly available image classification databases and it is shown that this method can perform significantly better than many competitive image classification algorithms. PMID:24835226
Ensemble polarimetric SAR image classification based on contextual sparse representation
NASA Astrophysics Data System (ADS)
Zhang, Lamei; Wang, Xiao; Zou, Bin; Qiao, Zhijun
2016-05-01
Polarimetric SAR image interpretation has become one of the most interesting topics, in which the construction of the reasonable and effective technique of image classification is of key importance. Sparse representation represents the data using the most succinct sparse atoms of the over-complete dictionary and the advantages of sparse representation also have been confirmed in the field of PolSAR classification. However, it is not perfect, like the ordinary classifier, at different aspects. So ensemble learning is introduced to improve the issue, which makes a plurality of different learners training and obtained the integrated results by combining the individual learner to get more accurate and ideal learning results. Therefore, this paper presents a polarimetric SAR image classification method based on the ensemble learning of sparse representation to achieve the optimal classification.
STIS Sparse Field CTE test-internal {Cycle 12}
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
2003-07-01
CTE measurements are made using the "internal sparse field test", along the parallel axis. The "POS=" optional parameter, introduced during cycle 11, is used to provide off-center MSM positionings of some slits. All exposures are internals.
STIS Sparse Field CTE test-internal {Cycle 11}
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
2002-07-01
CTE measurements are made using the "internal sparse field test", along the parallel axis. The new "POS=" optional parameter is used to provide off-center MSM positionings of some slits. All exposures are internals.
Testing of Error-Correcting Sparse Permutation Channel Codes
NASA Technical Reports Server (NTRS)
Shcheglov, Kirill, V.; Orlov, Sergei S.
2008-01-01
A computer program performs Monte Carlo direct numerical simulations for testing sparse permutation channel codes, which offer strong error-correction capabilities at high code rates and are considered especially suitable for storage of digital data in holographic and volume memories. A word in a code of this type is characterized by, among other things, a sparseness parameter (M) and a fixed number (K) of 1 or "on" bits in a channel block length of N.
Out-of-Core Solutions of Complex Sparse Linear Equations
NASA Technical Reports Server (NTRS)
Yip, E. L.
1982-01-01
ETCLIB is library of subroutines for obtaining out-of-core solutions of complex sparse linear equations. Routines apply to dense and sparse matrices too large to be stored in core. Useful for solving any set of linear equations, but particularly useful in cases where coefficient matrix has no special properties that guarantee convergence with any of interative processes. The only assumption made is that coefficient matrix is not singular.
Saliency Detection Using Sparse and Nonlinear Feature Representation
Zhao, Qingjie; Manzoor, Muhammad Farhan; Ishaq Khan, Saqib
2014-01-01
An important aspect of visual saliency detection is how features that form an input image are represented. A popular theory supports sparse feature representation, an image being represented with a basis dictionary having sparse weighting coefficient. Another method uses a nonlinear combination of image features for representation. In our work, we combine the two methods and propose a scheme that takes advantage of both sparse and nonlinear feature representation. To this end, we use independent component analysis (ICA) and covariant matrices, respectively. To compute saliency, we use a biologically plausible center surround difference (CSD) mechanism. Our sparse features are adaptive in nature; the ICA basis function are learnt at every image representation, rather than being fixed. We show that Adaptive Sparse Features when used with a CSD mechanism yield better results compared to fixed sparse representations. We also show that covariant matrices consisting of nonlinear integration of color information alone are sufficient to efficiently estimate saliency from an image. The proposed dual representation scheme is then evaluated against human eye fixation prediction, response to psychological patterns, and salient object detection on well-known datasets. We conclude that having two forms of representation compliments one another and results in better saliency detection. PMID:24895644
Vector sparse representation of color image using quaternion matrix analysis.
Xu, Yi; Yu, Licheng; Xu, Hongteng; Zhang, Hao; Nguyen, Truong
2015-04-01
Traditional sparse image models treat color image pixel as a scalar, which represents color channels separately or concatenate color channels as a monochrome image. In this paper, we propose a vector sparse representation model for color images using quaternion matrix analysis. As a new tool for color image representation, its potential applications in several image-processing tasks are presented, including color image reconstruction, denoising, inpainting, and super-resolution. The proposed model represents the color image as a quaternion matrix, where a quaternion-based dictionary learning algorithm is presented using the K-quaternion singular value decomposition (QSVD) (generalized K-means clustering for QSVD) method. It conducts the sparse basis selection in quaternion space, which uniformly transforms the channel images to an orthogonal color space. In this new color space, it is significant that the inherent color structures can be completely preserved during vector reconstruction. Moreover, the proposed sparse model is more efficient comparing with the current sparse models for image restoration tasks due to lower redundancy between the atoms of different color channels. The experimental results demonstrate that the proposed sparse image model avoids the hue bias issue successfully and shows its potential as a general and powerful tool in color image analysis and processing domain. PMID:25643407
Visual tracking based on extreme learning machine and sparse representation.
Wang, Baoxian; Tang, Linbo; Yang, Jinglin; Zhao, Baojun; Wang, Shuigen
2015-01-01
The existing sparse representation-based visual trackers mostly suffer from both being time consuming and having poor robustness problems. To address these issues, a novel tracking method is presented via combining sparse representation and an emerging learning technique, namely extreme learning machine (ELM). Specifically, visual tracking can be divided into two consecutive processes. Firstly, ELM is utilized to find the optimal separate hyperplane between the target observations and background ones. Thus, the trained ELM classification function is able to remove most of the candidate samples related to background contents efficiently, thereby reducing the total computational cost of the following sparse representation. Secondly, to further combine ELM and sparse representation, the resultant confidence values (i.e., probabilities to be a target) of samples on the ELM classification function are used to construct a new manifold learning constraint term of the sparse representation framework, which tends to achieve robuster results. Moreover, the accelerated proximal gradient method is used for deriving the optimal solution (in matrix form) of the constrained sparse tracking model. Additionally, the matrix form solution allows the candidate samples to be calculated in parallel, thereby leading to a higher efficiency. Experiments demonstrate the effectiveness of the proposed tracker. PMID:26506359
Representation-Independent Iteration of Sparse Data Arrays
NASA Technical Reports Server (NTRS)
James, Mark
2007-01-01
An approach is defined that describes a method of iterating over massively large arrays containing sparse data using an approach that is implementation independent of how the contents of the sparse arrays are laid out in memory. What is unique and important here is the decoupling of the iteration over the sparse set of array elements from how they are internally represented in memory. This enables this approach to be backward compatible with existing schemes for representing sparse arrays as well as new approaches. What is novel here is a new approach for efficiently iterating over sparse arrays that is independent of the underlying memory layout representation of the array. A functional interface is defined for implementing sparse arrays in any modern programming language with a particular focus for the Chapel programming language. Examples are provided that show the translation of a loop that computes a matrix vector product into this representation for both the distributed and not-distributed cases. This work is directly applicable to NASA and its High Productivity Computing Systems (HPCS) program that JPL and our current program are engaged in. The goal of this program is to create powerful, scalable, and economically viable high-powered computer systems suitable for use in national security and industry by 2010. This is important to NASA for its computationally intensive requirements for analyzing and understanding the volumes of science data from our returned missions.
Visual Tracking Based on Extreme Learning Machine and Sparse Representation
Wang, Baoxian; Tang, Linbo; Yang, Jinglin; Zhao, Baojun; Wang, Shuigen
2015-01-01
The existing sparse representation-based visual trackers mostly suffer from both being time consuming and having poor robustness problems. To address these issues, a novel tracking method is presented via combining sparse representation and an emerging learning technique, namely extreme learning machine (ELM). Specifically, visual tracking can be divided into two consecutive processes. Firstly, ELM is utilized to find the optimal separate hyperplane between the target observations and background ones. Thus, the trained ELM classification function is able to remove most of the candidate samples related to background contents efficiently, thereby reducing the total computational cost of the following sparse representation. Secondly, to further combine ELM and sparse representation, the resultant confidence values (i.e., probabilities to be a target) of samples on the ELM classification function are used to construct a new manifold learning constraint term of the sparse representation framework, which tends to achieve robuster results. Moreover, the accelerated proximal gradient method is used for deriving the optimal solution (in matrix form) of the constrained sparse tracking model. Additionally, the matrix form solution allows the candidate samples to be calculated in parallel, thereby leading to a higher efficiency. Experiments demonstrate the effectiveness of the proposed tracker. PMID:26506359
Automatic target recognition using group-structured sparse representation
NASA Astrophysics Data System (ADS)
Sun, Bo; Wu, Xuewen; He, Jun; Zhu, Xiaoming; Chen, Chao
2014-06-01
Sparse representation classification method has been increasingly used in the fields of computer vision and pattern analysis, due to its high recognition rate, little dependence on the features, robustness to corruption and occlusion, and etc. However, most of these existing methods aim to find the sparsest representations of the test sample y in an overcomplete dictionary, which do not particularly consider the relevant structure between the atoms in the dictionary. Moreover, sufficient training samples are always required by the sparse representation method for effective recognition. In this paper we formulate the classification as a group-structured sparse representation problem using a sparsity-inducing norm minimization optimization and propose a novel sparse representation-based automatic target recognition (ATR) framework for the practical applications in which the training samples are drawn from the simulation models of real targets. The experimental results show that the proposed approach improves the recognition rate of standard sparse models, and our system can effectively and efficiently recognize targets under real environments, especially, where the good characteristics of the sparse representation based classification method are kept.
Arizona Vegetation Resource Inventory (AVRI) accuracy assessment
Szajgin, John; Pettinger, L.R.; Linden, D.S.; Ohlen, D.O.
1982-01-01
A quantitative accuracy assessment was performed for the vegetation classification map produced as part of the Arizona Vegetation Resource Inventory (AVRI) project. This project was a cooperative effort between the Bureau of Land Management (BLM) and the Earth Resources Observation Systems (EROS) Data Center. The objective of the accuracy assessment was to estimate (with a precision of ?10 percent at the 90 percent confidence level) the comission error in each of the eight level II hierarchical vegetation cover types. A stratified two-phase (double) cluster sample was used. Phase I consisted of 160 photointerpreted plots representing clusters of Landsat pixels, and phase II consisted of ground data collection at 80 of the phase I cluster sites. Ground data were used to refine the phase I error estimates by means of a linear regression model. The classified image was stratified by assigning each 15-pixel cluster to the stratum corresponding to the dominant cover type within each cluster. This method is known as stratified plurality sampling. Overall error was estimated to be 36 percent with a standard error of 2 percent. Estimated error for individual vegetation classes ranged from a low of 10 percent ?6 percent for evergreen woodland to 81 percent ?7 percent for cropland and pasture. Total cost of the accuracy assessment was $106,950 for the one-million-hectare study area. The combination of the stratified plurality sampling (SPS) method of sample allocation with double sampling provided the desired estimates within the required precision levels. The overall accuracy results confirmed that highly accurate digital classification of vegetation is difficult to perform in semiarid environments, due largely to the sparse vegetation cover. Nevertheless, these techniques show promise for providing more accurate information than is presently available for many BLM-administered lands.
Marginality and needs of dietary valine for broilers fed certain all-vegetable diets
Technology Transfer Automated Retrieval System (TEKTRAN)
Valine is likely the fourth limiting amino acid in most diets based of corn and soybean meal. However, its exact needs are not well known, and information regarding it is sparse. A series of studies were conducted to validate valine’s limitation in all-vegetable diets fed to broilers, and subseque...
Sparse modeling of spatial environmental variables associated with asthma
Chang, Timothy S.; Gangnon, Ronald E.; Page, C. David; Buckingham, William R.; Tandias, Aman; Cowan, Kelly J.; Tomasallo, Carrie D.; Arndt, Brian G.; Hanrahan, Lawrence P.; Guilbert, Theresa W.
2014-01-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin’s Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5–50 years over a three-year period. Each patient’s home address was geocoded to one of 3,456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin’s geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. PMID:25533437
Sparse modeling of spatial environmental variables associated with asthma.
Chang, Timothy S; Gangnon, Ronald E; David Page, C; Buckingham, William R; Tandias, Aman; Cowan, Kelly J; Tomasallo, Carrie D; Arndt, Brian G; Hanrahan, Lawrence P; Guilbert, Theresa W
2015-02-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin's Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5-50years over a three-year period. Each patient's home address was geocoded to one of 3456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin's geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors.
Sparse modeling of spatial environmental variables associated with asthma.
Chang, Timothy S; Gangnon, Ronald E; David Page, C; Buckingham, William R; Tandias, Aman; Cowan, Kelly J; Tomasallo, Carrie D; Arndt, Brian G; Hanrahan, Lawrence P; Guilbert, Theresa W
2015-02-01
Geographically distributed environmental factors influence the burden of diseases such as asthma. Our objective was to identify sparse environmental variables associated with asthma diagnosis gathered from a large electronic health record (EHR) dataset while controlling for spatial variation. An EHR dataset from the University of Wisconsin's Family Medicine, Internal Medicine and Pediatrics Departments was obtained for 199,220 patients aged 5-50years over a three-year period. Each patient's home address was geocoded to one of 3456 geographic census block groups. Over one thousand block group variables were obtained from a commercial database. We developed a Sparse Spatial Environmental Analysis (SASEA). Using this method, the environmental variables were first dimensionally reduced with sparse principal component analysis. Logistic thin plate regression spline modeling was then used to identify block group variables associated with asthma from sparse principal components. The addresses of patients from the EHR dataset were distributed throughout the majority of Wisconsin's geography. Logistic thin plate regression spline modeling captured spatial variation of asthma. Four sparse principal components identified via model selection consisted of food at home, dog ownership, household size, and disposable income variables. In rural areas, dog ownership and renter occupied housing units from significant sparse principal components were associated with asthma. Our main contribution is the incorporation of sparsity in spatial modeling. SASEA sequentially added sparse principal components to Logistic thin plate regression spline modeling. This method allowed association of geographically distributed environmental factors with asthma using EHR and environmental datasets. SASEA can be applied to other diseases with environmental risk factors. PMID:25533437
Water Control on Vegetation Growth Pattern in Eurasia from GRACE
NASA Astrophysics Data System (ADS)
A, G.; Velicogna, I.; Kimball, J. S.; Kim, Y.
2014-12-01
High latitude ecosystem productivity is constrained by cold temperature and moisture limitations to plant growth, while these environmental restrictions may be changing with global warming. Satellite data driven assessments indicate that over the past three decades, rapid warming in the northern high latitudes has resulted in earlier and longer potential growing seasons and widespread greening, due to general relaxation of cold temperature constraints to vegetation productivity. However, warming may have also increased water stress limitations to growth. In this study, we use GRACE (Gravity Recovery and Climate Experiment) derived total water storage (TWS), 2-m air temperature (T) from ERA-interim reanalysis, normalized difference vegetation index (NDVI) data from MODIS (Collection 5) and satellite data driven vegetation gross primary productivity (GPP) estimates as surrogates for vegetation growth, for the period August 2002-December 2013 to evaluate terrestrial water supply controls to vegetation growth changes over the three major river basins of northern Eurasia. We find that during the analyzed period, the apparent growth response follows regional vegetation, moisture and temperature gradients and is spatially complex. In the drier southwest characterized by grassland, vegetation growth is mainly controlled by TWS availability. In the central region, dominated by cold temperature and water limited boreal forest, T is the main control on vegetation growth. In the Lena basin, where vegetation includes both boreal forest and water limited grassland, both T and TWS impact vegetation growth. We suggest that GRACE TWS estimates provide reliable observational constraints on water availability to vegetation that supplement sparse soil moisture observations and satellite precipitation estimates with unknown bias.
Ecogeomorphology of Sand Dunes Shaped by Vegetation
NASA Astrophysics Data System (ADS)
Tsoar, H.
2014-12-01
Two dune types associated with vegetation are known: Parabolic and Vegetated Linear Dunes (VLDs), the latters are the dominant dune type in the world deserts. Parabolic dunes are formed in humid, sub-humid and semi-arid environments (rather than arid) where vegetation is nearby. VLDs are known today in semiarid and arid lands where the average yearly rainfall is ≥100 mm, enough to support sparse cover of vegetation. These two dune types are formed by unidirectional winds although they demonstrate a different form and have a distinct dynamics. Conceptual and mathematical models of dunes mobility and stability, based on three control parameters: wind power (DP), average annual precipitation (p), and the human impact parameter (μ) show that where human impact is negligible the effect of wind power (DP) on vegetative cover is substantial. The average yearly rainfall of 60-80 mm is the threshold of annual average rainfall for vegetation growth on dune sand. The model is shown to follow a hysteresis path, which explains the bistability of active and stabilized dunes under the same climatic conditions with respect to wind power. We have discerned formation of parabolic dunes from barchans and transverse dunes in the coastal plain of Israel where a decrease in human activity during the second half of the 20th century caused establishment of vegetation on the crest of the dunes, a process that changed the dynamics of these barchans and transverse dunes and led to a change in the shape of the windward slope from convex to concave. These dunes gradually became parabolic. It seems that VLDs in Australia or the Kalahari have always been vegetated to some degree, though the shrubs were sparser in colder periods when the aeolian erosion was sizeable. Those ancient conditions are characterized by higher wind power and lower rainfall that can reduce, but not completely destroy, the vegetation cover, leading to the formation of lee (shadow) dunes behind each shrub. Formation of
... the non-starchy vegetables . Starchy vegetables like potatoes, corn and peas are included in the " Grains and ... or Chinese spinach Artichoke Artichoke hearts Asparagus Baby corn Bamboo shoots Beans (green, wax, Italian) Bean sprouts ...
A healthy diet includes adding vegetables and fruit every day. Vegetables like broccoli, green beans, leafy greens, zucchini, cauliflower, cabbage, carrots, and tomatoes are low in calories and high in fiber, ...
Online Hierarchical Sparse Representation of Multifeature for Robust Object Tracking
Qu, Shiru
2016-01-01
Object tracking based on sparse representation has given promising tracking results in recent years. However, the trackers under the framework of sparse representation always overemphasize the sparse representation and ignore the correlation of visual information. In addition, the sparse coding methods only encode the local region independently and ignore the spatial neighborhood information of the image. In this paper, we propose a robust tracking algorithm. Firstly, multiple complementary features are used to describe the object appearance; the appearance model of the tracked target is modeled by instantaneous and stable appearance features simultaneously. A two-stage sparse-coded method which takes the spatial neighborhood information of the image patch and the computation burden into consideration is used to compute the reconstructed object appearance. Then, the reliability of each tracker is measured by the tracking likelihood function of transient and reconstructed appearance models. Finally, the most reliable tracker is obtained by a well established particle filter framework; the training set and the template library are incrementally updated based on the current tracking results. Experiment results on different challenging video sequences show that the proposed algorithm performs well with superior tracking accuracy and robustness. PMID:27630710
FPGA implementation of sparse matrix algorithm for information retrieval
NASA Astrophysics Data System (ADS)
Bojanic, Slobodan; Jevtic, Ruzica; Nieto-Taladriz, Octavio
2005-06-01
Information text data retrieval requires a tremendous amount of processing time because of the size of the data and the complexity of information retrieval algorithms. In this paper the solution to this problem is proposed via hardware supported information retrieval algorithms. Reconfigurable computing may adopt frequent hardware modifications through its tailorable hardware and exploits parallelism for a given application through reconfigurable and flexible hardware units. The degree of the parallelism can be tuned for data. In this work we implemented standard BLAS (basic linear algebra subprogram) sparse matrix algorithm named Compressed Sparse Row (CSR) that is showed to be more efficient in terms of storage space requirement and query-processing timing over the other sparse matrix algorithms for information retrieval application. Although inverted index algorithm is treated as the de facto standard for information retrieval for years, an alternative approach to store the index of text collection in a sparse matrix structure gains more attention. This approach performs query processing using sparse matrix-vector multiplication and due to parallelization achieves a substantial efficiency over the sequential inverted index. The parallel implementations of information retrieval kernel are presented in this work targeting the Virtex II Field Programmable Gate Arrays (FPGAs) board from Xilinx. A recent development in scientific applications is the use of FPGA to achieve high performance results. Computational results are compared to implementations on other platforms. The design achieves a high level of parallelism for the overall function while retaining highly optimised hardware within processing unit.
X-ray computed tomography using curvelet sparse regularization
Wieczorek, Matthias Vogel, Jakob; Lasser, Tobias; Frikel, Jürgen; Demaret, Laurent; Eggl, Elena; Pfeiffer, Franz; Kopp, Felix; Noël, Peter B.
2015-04-15
Purpose: Reconstruction of x-ray computed tomography (CT) data remains a mathematically challenging problem in medical imaging. Complementing the standard analytical reconstruction methods, sparse regularization is growing in importance, as it allows inclusion of prior knowledge. The paper presents a method for sparse regularization based on the curvelet frame for the application to iterative reconstruction in x-ray computed tomography. Methods: In this work, the authors present an iterative reconstruction approach based on the alternating direction method of multipliers using curvelet sparse regularization. Results: Evaluation of the method is performed on a specifically crafted numerical phantom dataset to highlight the method’s strengths. Additional evaluation is performed on two real datasets from commercial scanners with different noise characteristics, a clinical bone sample acquired in a micro-CT and a human abdomen scanned in a diagnostic CT. The results clearly illustrate that curvelet sparse regularization has characteristic strengths. In particular, it improves the restoration and resolution of highly directional, high contrast features with smooth contrast variations. The authors also compare this approach to the popular technique of total variation and to traditional filtered backprojection. Conclusions: The authors conclude that curvelet sparse regularization is able to improve reconstruction quality by reducing noise while preserving highly directional features.
Online Hierarchical Sparse Representation of Multifeature for Robust Object Tracking
Qu, Shiru
2016-01-01
Object tracking based on sparse representation has given promising tracking results in recent years. However, the trackers under the framework of sparse representation always overemphasize the sparse representation and ignore the correlation of visual information. In addition, the sparse coding methods only encode the local region independently and ignore the spatial neighborhood information of the image. In this paper, we propose a robust tracking algorithm. Firstly, multiple complementary features are used to describe the object appearance; the appearance model of the tracked target is modeled by instantaneous and stable appearance features simultaneously. A two-stage sparse-coded method which takes the spatial neighborhood information of the image patch and the computation burden into consideration is used to compute the reconstructed object appearance. Then, the reliability of each tracker is measured by the tracking likelihood function of transient and reconstructed appearance models. Finally, the most reliable tracker is obtained by a well established particle filter framework; the training set and the template library are incrementally updated based on the current tracking results. Experiment results on different challenging video sequences show that the proposed algorithm performs well with superior tracking accuracy and robustness.
Robust visual multitask tracking via composite sparse model
NASA Astrophysics Data System (ADS)
Jin, Bo; Jing, Zhongliang; Wang, Meng; Pan, Han
2014-11-01
Recently, multitask learning was applied to visual tracking by learning sparse particle representations in a joint task, which led to the so-called multitask tracking algorithm (MTT). Although MTT shows impressive tracking performances by mining the interdependencies between particles, the individual feature of each particle is underestimated. The utilized L1,q norm regularization assumes all features are shared between all particles and results in nearly identical representation coefficients in nonsparse rows. We propose a composite sparse multitask tracking algorithm (CSMTT). We develop a composite sparse model to formulate the object appearance as a combination of the shared feature component, the individual feature component, and the outlier component. The composite sparsity is achieved via the L and L1,1 norm minimization, and is optimized by the alternating direction method of multipliers, which provides a favorable reconstruction performance and an impressive computational efficiency. Moreover, a dynamical dictionary updating scheme is proposed to capture appearance changes. CSMTT is tested on real-world video sequences under various challenges, and experimental results show that the composite sparse model achieves noticeable lower reconstruction errors and higher computational speeds than traditional sparse models, and CSMTT has consistently better tracking performances against seven state-of-the-art trackers.
Image fusion via nonlocal sparse K-SVD dictionary learning.
Li, Ying; Li, Fangyi; Bai, Bendu; Shen, Qiang
2016-03-01
Image fusion aims to merge two or more images captured via various sensors of the same scene to construct a more informative image by integrating their details. Generally, such integration is achieved through the manipulation of the representations of the images concerned. Sparse representation plays an important role in the effective description of images, offering a great potential in a variety of image processing tasks, including image fusion. Supported by sparse representation, in this paper, an approach for image fusion by the use of a novel dictionary learning scheme is proposed. The nonlocal self-similarity property of the images is exploited, not only at the stage of learning the underlying description dictionary but during the process of image fusion. In particular, the property of nonlocal self-similarity is combined with the traditional sparse dictionary. This results in an improved learned dictionary, hereafter referred to as the nonlocal sparse K-SVD dictionary (where K-SVD stands for the K times singular value decomposition that is commonly used in the literature), and abbreviated to NL_SK_SVD. The performance of the NL_SK_SVD dictionary is applied for image fusion using simultaneous orthogonal matching pursuit. The proposed approach is evaluated with different types of images, and compared with a number of alternative image fusion techniques. The resultant superior fused images using the present approach demonstrates the efficacy of the NL_SK_SVD dictionary in sparse image representation.
Online Hierarchical Sparse Representation of Multifeature for Robust Object Tracking.
Yang, Honghong; Qu, Shiru
2016-01-01
Object tracking based on sparse representation has given promising tracking results in recent years. However, the trackers under the framework of sparse representation always overemphasize the sparse representation and ignore the correlation of visual information. In addition, the sparse coding methods only encode the local region independently and ignore the spatial neighborhood information of the image. In this paper, we propose a robust tracking algorithm. Firstly, multiple complementary features are used to describe the object appearance; the appearance model of the tracked target is modeled by instantaneous and stable appearance features simultaneously. A two-stage sparse-coded method which takes the spatial neighborhood information of the image patch and the computation burden into consideration is used to compute the reconstructed object appearance. Then, the reliability of each tracker is measured by the tracking likelihood function of transient and reconstructed appearance models. Finally, the most reliable tracker is obtained by a well established particle filter framework; the training set and the template library are incrementally updated based on the current tracking results. Experiment results on different challenging video sequences show that the proposed algorithm performs well with superior tracking accuracy and robustness. PMID:27630710
NASA Technical Reports Server (NTRS)
Schrumpf, B. J.
1973-01-01
Unique characteristics of ERTS imagery can be used to inventory natural vegetation. While satellite images can seldom be interpreted and identified directly in terms of vegetation types, such types can be inferred by interpretation of physical terrain features and through an understanding of the ecology of the vegetation.
Geostatistical estimation of signal-to-noise ratios for spectral vegetation indices
Ji, Lei; Zhang, Li; Rover, Jennifer R.; Wylie, Bruce K.; Chen, Xuexia
2014-01-01
In the past 40 years, many spectral vegetation indices have been developed to quantify vegetation biophysical parameters. An ideal vegetation index should contain the maximum level of signal related to specific biophysical characteristics and the minimum level of noise such as background soil influences and atmospheric effects. However, accurate quantification of signal and noise in a vegetation index remains a challenge, because it requires a large number of field measurements or laboratory experiments. In this study, we applied a geostatistical method to estimate signal-to-noise ratio (S/N) for spectral vegetation indices. Based on the sample semivariogram of vegetation index images, we used the standardized noise to quantify the noise component of vegetation indices. In a case study in the grasslands and shrublands of the western United States, we demonstrated the geostatistical method for evaluating S/N for a series of soil-adjusted vegetation indices derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor. The soil-adjusted vegetation indices were found to have higher S/N values than the traditional normalized difference vegetation index (NDVI) and simple ratio (SR) in the sparsely vegetated areas. This study shows that the proposed geostatistical analysis can constitute an efficient technique for estimating signal and noise components in vegetation indices.
Geostatistical estimation of signal-to-noise ratios for spectral vegetation indices
NASA Astrophysics Data System (ADS)
Ji, Lei; Zhang, Li; Rover, Jennifer; Wylie, Bruce K.; Chen, Xuexia
2014-10-01
In the past 40 years, many spectral vegetation indices have been developed to quantify vegetation biophysical parameters. An ideal vegetation index should contain the maximum level of signal related to specific biophysical characteristics and the minimum level of noise such as background soil influences and atmospheric effects. However, accurate quantification of signal and noise in a vegetation index remains a challenge, because it requires a large number of field measurements or laboratory experiments. In this study, we applied a geostatistical method to estimate signal-to-noise ratio (S/N) for spectral vegetation indices. Based on the sample semivariogram of vegetation index images, we used the standardized noise to quantify the noise component of vegetation indices. In a case study in the grasslands and shrublands of the western United States, we demonstrated the geostatistical method for evaluating S/N for a series of soil-adjusted vegetation indices derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor. The soil-adjusted vegetation indices were found to have higher S/N values than the traditional normalized difference vegetation index (NDVI) and simple ratio (SR) in the sparsely vegetated areas. This study shows that the proposed geostatistical analysis can constitute an efficient technique for estimating signal and noise components in vegetation indices.
Sparse representation based face recognition using weighted regions
NASA Astrophysics Data System (ADS)
Bilgazyev, Emil; Yeniaras, E.; Uyanik, I.; Unan, Mahmut; Leiss, E. L.
2013-12-01
Face recognition is a challenging research topic, especially when the training (gallery) and recognition (probe) images are acquired using different cameras under varying conditions. Even a small noise or occlusion in the images can compromise the accuracy of recognition. Lately, sparse encoding based classification algorithms gave promising results for such uncontrollable scenarios. In this paper, we introduce a novel methodology by modeling the sparse encoding with weighted patches to increase the robustness of face recognition even further. In the training phase, we define a mask (i.e., weight matrix) using a sparse representation selecting the facial regions, and in the recognition phase, we perform comparison on selected facial regions. The algorithm was evaluated both quantitatively and qualitatively using two comprehensive surveillance facial image databases, i.e., SCfaceandMFPV, with the results clearly superior to common state-of-the-art methodologies in different scenarios.
Uncovering Transcriptional Regulatory Networks by Sparse Bayesian Factor Model
NASA Astrophysics Data System (ADS)
Meng, Jia; Zhang, Jianqiu(Michelle); Qi, Yuan(Alan); Chen, Yidong; Huang, Yufei
2010-12-01
The problem of uncovering transcriptional regulation by transcription factors (TFs) based on microarray data is considered. A novel Bayesian sparse correlated rectified factor model (BSCRFM) is proposed that models the unknown TF protein level activity, the correlated regulations between TFs, and the sparse nature of TF-regulated genes. The model admits prior knowledge from existing database regarding TF-regulated target genes based on a sparse prior and through a developed Gibbs sampling algorithm, a context-specific transcriptional regulatory network specific to the experimental condition of the microarray data can be obtained. The proposed model and the Gibbs sampling algorithm were evaluated on the simulated systems, and results demonstrated the validity and effectiveness of the proposed approach. The proposed model was then applied to the breast cancer microarray data of patients with Estrogen Receptor positive ([InlineEquation not available: see fulltext.]) status and Estrogen Receptor negative ([InlineEquation not available: see fulltext.]) status, respectively.
Joint sparse representation for robust multimodal biometrics recognition.
Shekhar, Sumit; Patel, Vishal M; Nasrabadi, Nasser M; Chellappa, Rama
2014-01-01
Traditional biometric recognition systems rely on a single biometric signature for authentication. While the advantage of using multiple sources of information for establishing the identity has been widely recognized, computational models for multimodal biometrics recognition have only recently received attention. We propose a multimodal sparse representation method, which represents the test data by a sparse linear combination of training data, while constraining the observations from different modalities of the test subject to share their sparse representations. Thus, we simultaneously take into account correlations as well as coupling information among biometric modalities. A multimodal quality measure is also proposed to weigh each modality as it gets fused. Furthermore, we also kernelize the algorithm to handle nonlinearity in data. The optimization problem is solved using an efficient alternative direction method. Various experiments show that the proposed method compares favorably with competing fusion-based methods.
Sparse nonnegative matrix factorization with ℓ0-constraints
Peharz, Robert; Pernkopf, Franz
2012-01-01
Although nonnegative matrix factorization (NMF) favors a sparse and part-based representation of nonnegative data, there is no guarantee for this behavior. Several authors proposed NMF methods which enforce sparseness by constraining or penalizing the ℓ1-norm of the factor matrices. On the other hand, little work has been done using a more natural sparseness measure, the ℓ0-pseudo-norm. In this paper, we propose a framework for approximate NMF which constrains the ℓ0-norm of the basis matrix, or the coefficient matrix, respectively. For this purpose, techniques for unconstrained NMF can be easily incorporated, such as multiplicative update rules, or the alternating nonnegative least-squares scheme. In experiments we demonstrate the benefits of our methods, which compare to, or outperform existing approaches. PMID:22505792
Sparse representation-based image restoration via nonlocal supervised coding
NASA Astrophysics Data System (ADS)
Li, Ao; Chen, Deyun; Sun, Guanglu; Lin, Kezheng
2016-09-01
Sparse representation (SR) and nonlocal technique (NLT) have shown great potential in low-level image processing. However, due to the degradation of the observed image, SR and NLT may not be accurate enough to obtain a faithful restoration results when they are used independently. To improve the performance, in this paper, a nonlocal supervised coding strategy-based NLT for image restoration is proposed. The novel method has three main contributions. First, to exploit the useful nonlocal patches, a nonnegative sparse representation is introduced, whose coefficients can be utilized as the supervised weights among patches. Second, a novel objective function is proposed, which integrated the supervised weights learning and the nonlocal sparse coding to guarantee a more promising solution. Finally, to make the minimization tractable and convergence, a numerical scheme based on iterative shrinkage thresholding is developed to solve the above underdetermined inverse problem. The extensive experiments validate the effectiveness of the proposed method.
Sparse representation-based image restoration via nonlocal supervised coding
NASA Astrophysics Data System (ADS)
Li, Ao; Chen, Deyun; Sun, Guanglu; Lin, Kezheng
2016-10-01
Sparse representation (SR) and nonlocal technique (NLT) have shown great potential in low-level image processing. However, due to the degradation of the observed image, SR and NLT may not be accurate enough to obtain a faithful restoration results when they are used independently. To improve the performance, in this paper, a nonlocal supervised coding strategy-based NLT for image restoration is proposed. The novel method has three main contributions. First, to exploit the useful nonlocal patches, a nonnegative sparse representation is introduced, whose coefficients can be utilized as the supervised weights among patches. Second, a novel objective function is proposed, which integrated the supervised weights learning and the nonlocal sparse coding to guarantee a more promising solution. Finally, to make the minimization tractable and convergence, a numerical scheme based on iterative shrinkage thresholding is developed to solve the above underdetermined inverse problem. The extensive experiments validate the effectiveness of the proposed method.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
NASA Astrophysics Data System (ADS)
Vitello, Peter; Fried, Laurence E.; Pudliner, Brian; McAbee, Tom
2004-07-01
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamic calculations, CHEETAH uses parallel communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH coupled to an ALE hydrocode. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.
SAR target classification based on multiscale sparse representation
NASA Astrophysics Data System (ADS)
Ruan, Huaiyu; Zhang, Rong; Li, Jingge; Zhan, Yibing
2016-03-01
We propose a novel multiscale sparse representation approach for SAR target classification. It firstly extracts the dense SIFT descriptors on multiple scales, then trains a global multiscale dictionary by sparse coding algorithm. After obtaining the sparse representation, the method applies spatial pyramid matching (SPM) and max pooling to summarize the features for each image. The proposed method can provide more information and descriptive ability than single-scale ones. Moreover, it costs less extra computation than existing multiscale methods which compute a dictionary for each scale. The MSTAR database and ship database collected from TerraSAR-X images are used in classification setup. Results show that the best overall classification rate of the proposed approach can achieve 98.83% on the MSTAR database and 92.67% on the TerraSAR-X ship database.
P-SPARSLIB: A parallel sparse iterative solution package
Saad, Y.
1994-12-31
Iterative methods are gaining popularity in engineering and sciences at a time where the computational environment is changing rapidly. P-SPARSLIB is a project to build a software library for sparse matrix computations on parallel computers. The emphasis is on iterative methods and the use of distributed sparse matrices, an extension of the domain decomposition approach to general sparse matrices. One of the goals of this project is to develop a software package geared towards specific applications. For example, the author will test the performance and usefulness of P-SPARSLIB modules on linear systems arising from CFD applications. Equally important is the goal of portability. In the long run, the author wishes to ensure that this package is portable on a variety of platforms, including SIMD environments and shared memory environments.
A Multiobjective Sparse Feature Learning Model for Deep Neural Networks.
Gong, Maoguo; Liu, Jia; Li, Hao; Cai, Qing; Su, Linzhi
2015-12-01
Hierarchical deep neural networks are currently popular learning models for imitating the hierarchical architecture of human brain. Single-layer feature extractors are the bricks to build deep networks. Sparse feature learning models are popular models that can learn useful representations. But most of those models need a user-defined constant to control the sparsity of representations. In this paper, we propose a multiobjective sparse feature learning model based on the autoencoder. The parameters of the model are learnt by optimizing two objectives, reconstruction error and the sparsity of hidden units simultaneously to find a reasonable compromise between them automatically. We design a multiobjective induced learning procedure for this model based on a multiobjective evolutionary algorithm. In the experiments, we demonstrate that the learning procedure is effective, and the proposed multiobjective model can learn useful sparse features.
Sparse Sensing of Aerodynamic Loads on Insect Wings
NASA Astrophysics Data System (ADS)
Manohar, Krithika; Brunton, Steven; Kutz, J. Nathan
2015-11-01
We investigate how insects use sparse sensors on their wings to detect aerodynamic loading and wing deformation using a coupled fluid-structure model given periodically flapping input motion. Recent observations suggest that insects collect sensor information about their wing deformation to inform control actions for maneuvering and rejecting gust disturbances. Given a small number of point measurements of the chordwise aerodynamic loads from the sparse sensors, we reconstruct the entire chordwise loading using sparsesensing - a signal processing technique that reconstructs a signal from a small number of measurements using l1 norm minimization of sparse modal coefficients in some basis. We compare reconstructions from sensors randomly sampled from probability distributions biased toward different regions along the wing chord. In this manner, we determine the preferred regions along the chord for sensor placement and for estimating chordwise loads to inform control decisions in flight.
Automatic landslide and mudflow detection method via multichannel sparse representation
NASA Astrophysics Data System (ADS)
Chao, Chen; Zhou, Jianjun; Hao, Zhuo; Sun, Bo; He, Jun; Ge, Fengxiang
2015-10-01
Landslide and mudflow detection is an important application of aerial images and high resolution remote sensing images, which is crucial for national security and disaster relief. Since the high resolution images are often large in size, it's necessary to develop an efficient algorithm for landslide and mudflow detection. Based on the theory of sparse representation and, we propose a novel automatic landslide and mudflow detection method in this paper, which combines multi-channel sparse representation and eight neighbor judgment methods. The whole process of the detection is totally automatic. We make the experiment on a high resolution image of ZhouQu district of Gansu province in China on August, 2010 and get a promising result which proved the effective of using sparse representation on landslide and mudflow detection.
Sparse Partial Equilibrium Tables in Chemically Resolved Reactive Flow
Vitello, P; Fried, L E; Pudliner, B; McAbee, T
2003-07-14
The detonation of an energetic material is the result of a complex interaction between kinetic chemical reactions and hydrodynamics. Unfortunately, little is known concerning the detailed chemical kinetics of detonations in energetic materials. CHEETAH uses rate laws to treat species with the slowest chemical reactions, while assuming other chemical species are in equilibrium. CHEETAH supports a wide range of elements and condensed detonation products and can also be applied to gas detonations. A sparse hash table of equation of state values, called the ''cache'' is used in CHEETAH to enhance the efficiency of kinetic reaction calculations. For large-scale parallel hydrodynamic calculations, CHEETAH uses MPI communication to updates to the cache. We present here details of the sparse caching model used in the CHEETAH. To demonstrate the efficiency of modeling using a sparse cache model we consider detonations in energetic materials.
Joint sparse representation for robust multimodal biometrics recognition.
Shekhar, Sumit; Patel, Vishal M; Nasrabadi, Nasser M; Chellappa, Rama
2014-01-01
Traditional biometric recognition systems rely on a single biometric signature for authentication. While the advantage of using multiple sources of information for establishing the identity has been widely recognized, computational models for multimodal biometrics recognition have only recently received attention. We propose a multimodal sparse representation method, which represents the test data by a sparse linear combination of training data, while constraining the observations from different modalities of the test subject to share their sparse representations. Thus, we simultaneously take into account correlations as well as coupling information among biometric modalities. A multimodal quality measure is also proposed to weigh each modality as it gets fused. Furthermore, we also kernelize the algorithm to handle nonlinearity in data. The optimization problem is solved using an efficient alternative direction method. Various experiments show that the proposed method compares favorably with competing fusion-based methods. PMID:24231870
Pryde, E.H.
1982-01-01
Suggested standards for vegetable oils and ester fuels, as well as ASTM specifications for No. 2 diesel oil are given. The following physical properties were discussed: cetane number, cloud point, distillation temperatures, flash point, pour point, turbidity, viscosity, free fatty acids, iodine value, phosphorus, and wax. It was apparent that vegetable oils and their esters cannot meet ASTM specifications D975 for No. 2 diesel oil for use in the diesel engine. Vegetable oil modification or engine design modification may make it possible eventually for vegetable oils to become suitable alternative fuels. Vegetable oils must be recognized as experimental fuels until modifications have been tested thoroughly and generally accepted. 1 table. (DP)
Wang, Li-wen; Wei, Ya-xing; Niu, Zheng
2008-06-01
1 km MODIS NDVI time series data combining with decision tree classification, supervised classification and unsupervised classification was used to classify land cover type of Qinghai Province into 14 classes. In our classification system, sparse grassland and sparse shrub were emphasized, and their spatial distribution locations were labeled. From digital elevation model (DEM) of Qinghai Province, five elevation belts were achieved, and we utilized geographic information system (GIS) software to analyze vegetation cover variation on different elevation belts. Our research result shows that vegetation cover in Qinghai Province has been improved in recent five years. Vegetation cover area increases from 370047 km2 in 2001 to 374576 km2 in 2006, and vegetation cover rate increases by 0.63%. Among five grade elevation belts, vegetation cover ratio of high mountain belt is the highest (67.92%). The area of middle density grassland in high mountain belt is the largest, of which area is 94 003 km2. Increased area of dense grassland in high mountain belt is the greatest (1280 km2). During five years, the biggest variation is the conversion from sparse grassland to middle density grassland in high mountain belt, of which area is 15931 km2.
Content addressable systolic array for sparse matrix computation
Wing, O.
1983-01-01
A systolic array is proposed which is specifically designed to solve a system of sparse linear equations. The array consists of a number of processing elements connected in a ring. Each processing element has its own content addressable memory where the nonzero elements of the sparse matrix are stored. Matrix elements to which elementary operations are applied are extracted from the memory by content addressing. The system of equations is solved in a systolic fashion and the solution is obtained in nz+5n-2 steps where nz is the number of nonzero elements along and below the diagonal and n is the number of equations. 13 references.
Towards an Accurate Performance Modeling of Parallel SparseFactorization
Grigori, Laura; Li, Xiaoye S.
2006-05-26
We present a performance model to analyze a parallel sparseLU factorization algorithm on modern cached-based, high-end parallelarchitectures. Our model characterizes the algorithmic behavior bytakingaccount the underlying processor speed, memory system performance, aswell as the interconnect speed. The model is validated using theSuperLU_DIST linear system solver, the sparse matrices from realapplications, and an IBM POWER3 parallel machine. Our modelingmethodology can be easily adapted to study performance of other types ofsparse factorizations, such as Cholesky or QR.
D Super-Resolution Approach for Sparse Laser Scanner Data
NASA Astrophysics Data System (ADS)
Hosseinyalamdary, S.; Yilmaz, A.
2015-08-01
Laser scanner point cloud has been emerging in Photogrammetry and computer vision to achieve high level tasks such as object tracking, object recognition and scene understanding. However, low cost laser scanners are noisy, sparse and prone to systematic errors. This paper proposes a novel 3D super resolution approach to reconstruct surface of the objects in the scene. This method works on sparse, unorganized point clouds and has superior performance over other surface recovery approaches. Since the proposed approach uses anisotropic diffusion equation, it does not deteriorate the object boundaries and it preserves topology of the object.
Sparse Matrix for ECG Identification with Two-Lead Features
Tseng, Kuo-Kun; Luo, Jiao; Wang, Wenmin; Haiting, Dong
2015-01-01
Electrocardiograph (ECG) human identification has the potential to improve biometric security. However, improvements in ECG identification and feature extraction are required. Previous work has focused on single lead ECG signals. Our work proposes a new algorithm for human identification by mapping two-lead ECG signals onto a two-dimensional matrix then employing a sparse matrix method to process the matrix. And that is the first application of sparse matrix techniques for ECG identification. Moreover, the results of our experiments demonstrate the benefits of our approach over existing methods. PMID:25961074
A LONE code for the sparse control of quantum systems
NASA Astrophysics Data System (ADS)
Ciaramella, G.; Borzì, A.
2016-03-01
In many applications with quantum spin systems, control functions with a sparse and pulse-shaped structure are often required. These controls can be obtained by solving quantum optimal control problems with L1-penalized cost functionals. In this paper, the MATLAB package LONE is presented aimed to solving L1-penalized optimal control problems governed by unitary-operator quantum spin models. This package implements a new strategy that includes a globalized semi-smooth Krylov-Newton scheme and a continuation procedure. Results of numerical experiments demonstrate the ability of the LONE code in computing accurate sparse optimal control solutions.
BIRD: A general interface for sparse distributed memory simulators
NASA Technical Reports Server (NTRS)
Rogers, David
1990-01-01
Kanerva's sparse distributed memory (SDM) has now been implemented for at least six different computers, including SUN3 workstations, the Apple Macintosh, and the Connection Machine. A common interface for input of commands would both aid testing of programs on a broad range of computer architectures and assist users in transferring results from research environments to applications. A common interface also allows secondary programs to generate command sequences for a sparse distributed memory, which may then be executed on the appropriate hardware. The BIRD program is an attempt to create such an interface. Simplifying access to different simulators should assist developers in finding appropriate uses for SDM.
Robust head pose estimation using locality-constrained sparse coding
NASA Astrophysics Data System (ADS)
Kim, Hyunduk; Lee, Sang-Heon; Sohn, Myoung-Kyu
2015-12-01
Sparse coding (SC) method has been shown to deliver successful result in a variety of computer vision applications. However, it does not consider the underlying structure of the data in the feature space. On the other hand, locality constrained linear coding (LLC) utilizes locality constraint to project each input data into its local-coordinate system. Based on the recent success of LLC, we propose a novel locality-constrained sparse coding (LSC) method to overcome the limitation of the SC. In experiments, the proposed algorithms were applied to head pose estimation applications. Experimental results demonstrated that the LSC method is better than state-of-the-art methods.
Reconstruction Techniques for Sparse Multistatic Linear Array Microwave Imaging
Sheen, David M.; Hall, Thomas E.
2014-06-09
Sequentially-switched linear arrays are an enabling technology for a number of near-field microwave imaging applications. Electronically sequencing along the array axis followed by mechanical scanning along an orthogonal axis allows dense sampling of a two-dimensional aperture in near real-time. In this paper, a sparse multi-static array technique will be described along with associated Fourier-Transform-based and back-projection-based image reconstruction algorithms. Simulated and measured imaging results are presented that show the effectiveness of the sparse array technique along with the merits and weaknesses of each image reconstruction approach.
Moving target detection for frequency agility radar by sparse reconstruction
NASA Astrophysics Data System (ADS)
Quan, Yinghui; Li, YaChao; Wu, Yaojun; Ran, Lei; Xing, Mengdao; Liu, Mengqi
2016-09-01
Frequency agility radar, with randomly varied carrier frequency from pulse to pulse, exhibits superior performance compared to the conventional fixed carrier frequency pulse-Doppler radar against the electromagnetic interference. A novel moving target detection (MTD) method is proposed for the estimation of the target's velocity of frequency agility radar based on pulses within a coherent processing interval by using sparse reconstruction. Hardware implementation of orthogonal matching pursuit algorithm is executed on Xilinx Virtex-7 Field Programmable Gata Array (FPGA) to perform sparse optimization. Finally, a series of experiments are performed to evaluate the performance of proposed MTD method for frequency agility radar systems.
[Investigation of Multi-Angle Polarization Properties of Vegetation Based on RSP].
Jiao, Jian-nan; Zhao, Hai-meng; Yang, Bin; Yan, Lei
2016-02-01
Polarization detection provides us with novel information to reflect the target attribute. Compared with traditional remote sensing methods, multi-angle polarization has relatively stable correlation and regularity. RSP(research scanning polarimeter)is an airborne prototype for the APS(aerosol polarimetery sensor) developed by the USA, which can provide with us the polarization detection information of 9 channels. We can get optical properties and physical characteristics of vegetation by analyzing stable multi-angle and multi-band polarization detection information from preprocessing scanning polarization data of flight test. In this paper, after making registration based on flight attitude information, a comparative analysis is made between characteristics of reflectance and polarization reflectance with visible light and near infrared band of the view zenith angles between--30 degree and 65 degree, based on dense area and sparse area(close to bare field) of vegetation. The results show that both dense area and sparse area demonstrate regular characteristics of polarization degree. The area close to hot spot area has highest reflectance energy. In contrast,. it has relatively least energy of polarization degree, which can prevent strong reflectance energy from influencing the stability of detector. Because the degree of polarization in dense area of vegetation is higher than that in sparse area at visible light band while that in concentration area of vegetation is lower than sparse area at near infrared light band, it shows that the visible light band information of dense area of vegetation that the sensor received is dominated by single scattering while the near infrared light band information of dense area of vegetation is dominated by multiple scattering. PMID:27209749
Two demonstrators and a simulator for a sparse, distributed memory
NASA Technical Reports Server (NTRS)
Brown, Robert L.
1987-01-01
Described are two programs demonstrating different aspects of Kanerva's Sparse, Distributed Memory (SDM). These programs run on Sun 3 workstations, one using color, and have straightforward graphically oriented user interfaces and graphical output. Presented are descriptions of the programs, how to use them, and what they show. Additionally, this paper describes the software simulator behind each program.
Remote sensing image fusion via wavelet transform and sparse representation
NASA Astrophysics Data System (ADS)
Cheng, Jian; Liu, Haijun; Liu, Ting; Wang, Feng; Li, Hongsheng
2015-06-01
In this paper, we propose a remote sensing image fusion method which combines the wavelet transform and sparse representation to obtain fusion images with high spectral resolution and high spatial resolution. Firstly, intensity-hue-saturation (IHS) transform is applied to Multi-Spectral (MS) images. Then, wavelet transform is used to the intensity component of MS images and the Panchromatic (Pan) image to construct the multi-scale representation respectively. With the multi-scale representation, different fusion strategies are taken on the low-frequency and the high-frequency sub-images. Sparse representation with training dictionary is introduced into the low-frequency sub-image fusion. The fusion rule for the sparse representation coefficients of the low-frequency sub-images is defined by the spatial frequency maximum. For high-frequency sub-images with prolific detail information, the fusion rule is established by the images information fusion measurement indicator. Finally, the fused results are obtained through inverse wavelet transform and inverse IHS transform. The wavelet transform has the ability to extract the spectral information and the global spatial details from the original pairwise images, while sparse representation can extract the local structures of images effectively. Therefore, our proposed fusion method can well preserve the spectral information and the spatial detail information of the original images. The experimental results on the remote sensing images have demonstrated that our proposed method could well maintain the spectral characteristics of fusion images with a high spatial resolution.
Estimating Population Characteristics from Sparse Matrix Samples of Item Responses.
ERIC Educational Resources Information Center
Mislevy, Robert J.; And Others
1992-01-01
Concepts behind plausible values in estimating population characteristics from sparse matrix samples of item responses are discussed. The use of marginal analyses is described in the context of the National Assessment of Educational Progress, and the approach is illustrated with Scholastic Aptitude Test data for 9,075 high school seniors. (SLD)
Robust Methods for Sensing and Reconstructing Sparse Signals
ERIC Educational Resources Information Center
Carrillo, Rafael E.
2012-01-01
Compressed sensing (CS) is an emerging signal acquisition framework that goes against the traditional Nyquist sampling paradigm. CS demonstrates that a sparse, or compressible, signal can be acquired using a low rate acquisition process. Since noise is always present in practical data acquisition systems, sensing and reconstruction methods are…
Image Super-Resolution via Adaptive Regularization and Sparse Representation.
Cao, Feilong; Cai, Miaomiao; Tan, Yuanpeng; Zhao, Jianwei
2016-07-01
Previous studies have shown that image patches can be well represented as a sparse linear combination of elements from an appropriately selected over-complete dictionary. Recently, single-image super-resolution (SISR) via sparse representation using blurred and downsampled low-resolution images has attracted increasing interest, where the aim is to obtain the coefficients for sparse representation by solving an l0 or l1 norm optimization problem. The l0 optimization is a nonconvex and NP-hard problem, while the l1 optimization usually requires many more measurements and presents new challenges even when the image is the usual size, so we propose a new approach for SISR recovery based on regularization nonconvex optimization. The proposed approach is potentially a powerful method for recovering SISR via sparse representations, and it can yield a sparser solution than the l1 regularization method. We also consider the best choice for lp regularization with all p in (0, 1), where we propose a scheme that adaptively selects the norm value for each image patch. In addition, we provide a method for estimating the best value of the regularization parameter λ adaptively, and we discuss an alternate iteration method for selecting p and λ . We perform experiments, which demonstrates that the proposed regularization nonconvex optimization method can outperform the convex optimization method and generate higher quality images.
Beam hardening correction for sparse-view CT reconstruction
NASA Astrophysics Data System (ADS)
Liu, Wenlei; Rong, Junyan; Gao, Peng; Liao, Qimei; Lu, HongBing
2015-03-01
Beam hardening, which is caused by spectrum polychromatism of the X-ray beam, may result in various artifacts in the reconstructed image and degrade image quality. The artifacts would be further aggravated for the sparse-view reconstruction due to insufficient sampling data. Considering the advantages of the total-variation (TV) minimization in CT reconstruction with sparse-view data, in this paper, we propose a beam hardening correction method for sparse-view CT reconstruction based on Brabant's modeling. In this correction model for beam hardening, the attenuation coefficient of each voxel at the effective energy is modeled and estimated linearly, and can be applied in an iterative framework, such as simultaneous algebraic reconstruction technique (SART). By integrating the correction model into the forward projector of the algebraic reconstruction technique (ART), the TV minimization can recover images when only a limited number of projections are available. The proposed method does not need prior information about the beam spectrum. Preliminary validation using Monte Carlo simulations indicates that the proposed method can provide better reconstructed images from sparse-view projection data, with effective suppression of artifacts caused by beam hardening. With appropriate modeling of other degrading effects such as photon scattering, the proposed framework may provide a new way for low-dose CT imaging.
Hyperspherical Sparse Approximation Techniques for High-Dimensional Discontinuity Detection
Zhang, Guannan; Webster, Clayton G.; Gunzburger, Max; Burkardt, John
2016-08-04
This work proposes a hyperspherical sparse approximation framework for detecting jump discontinuities in functions in high-dimensional spaces. The need for a novel approach results from the theoretical and computational inefficiencies of well-known approaches, such as adaptive sparse grids, for discontinuity detection. Our approach constructs the hyperspherical coordinate representation of the discontinuity surface of a function. Then sparse approximations of the transformed function are built in the hyperspherical coordinate system, with values at each point estimated by solving a one-dimensional discontinuity detection problem. Due to the smoothness of the hypersurface, the new technique can identify jump discontinuities with significantly reduced computationalmore » cost, compared to existing methods. Several approaches are used to approximate the transformed discontinuity surface in the hyperspherical system, including adaptive sparse grid and radial basis function interpolation, discrete least squares projection, and compressed sensing approximation. Moreover, hierarchical acceleration techniques are also incorporated to further reduce the overall complexity. In conclusion, rigorous complexity analyses of the new methods are provided, as are several numerical examples that illustrate the effectiveness of our approach.« less
HYPOTHESIS TESTING FOR HIGH-DIMENSIONAL SPARSE BINARY REGRESSION
Mukherjee, Rajarshi; Pillai, Natesh S.; Lin, Xihong
2015-01-01
In this paper, we study the detection boundary for minimax hypothesis testing in the context of high-dimensional, sparse binary regression models. Motivated by genetic sequencing association studies for rare variant effects, we investigate the complexity of the hypothesis testing problem when the design matrix is sparse. We observe a new phenomenon in the behavior of detection boundary which does not occur in the case of Gaussian linear regression. We derive the detection boundary as a function of two components: a design matrix sparsity index and signal strength, each of which is a function of the sparsity of the alternative. For any alternative, if the design matrix sparsity index is too high, any test is asymptotically powerless irrespective of the magnitude of signal strength. For binary design matrices with the sparsity index that is not too high, our results are parallel to those in the Gaussian case. In this context, we derive detection boundaries for both dense and sparse regimes. For the dense regime, we show that the generalized likelihood ratio is rate optimal; for the sparse regime, we propose an extended Higher Criticism Test and show it is rate optimal and sharp. We illustrate the finite sample properties of the theoretical results using simulation studies. PMID:26246645
Sparse principal component analysis in medical shape modeling
NASA Astrophysics Data System (ADS)
Sjöstrand, Karl; Stegmann, Mikkel B.; Larsen, Rasmus
2006-03-01
Principal component analysis (PCA) is a widely used tool in medical image analysis for data reduction, model building, and data understanding and exploration. While PCA is a holistic approach where each new variable is a linear combination of all original variables, sparse PCA (SPCA) aims at producing easily interpreted models through sparse loadings, i.e. each new variable is a linear combination of a subset of the original variables. One of the aims of using SPCA is the possible separation of the results into isolated and easily identifiable effects. This article introduces SPCA for shape analysis in medicine. Results for three different data sets are given in relation to standard PCA and sparse PCA by simple thresholding of small loadings. Focus is on a recent algorithm for computing sparse principal components, but a review of other approaches is supplied as well. The SPCA algorithm has been implemented using Matlab and is available for download. The general behavior of the algorithm is investigated, and strengths and weaknesses are discussed. The original report on the SPCA algorithm argues that the ordering of modes is not an issue. We disagree on this point and propose several approaches to establish sensible orderings. A method that orders modes by decreasing variance and maximizes the sum of variances for all modes is presented and investigated in detail.
Sparse code of conflict in a primate society
Daniels, Bryan C.; Krakauer, David C.; Flack, Jessica C.
2012-01-01
Animals living in groups collectively produce social structure. In this context individuals make strategic decisions about when to cooperate and compete. This requires that individuals can perceive patterns in collective dynamics, but how this pattern extraction occurs is unclear. Our goal is to identify a model that extracts meaningful social patterns from a behavioral time series while remaining cognitively parsimonious by making the fewest demands on memory. Using fine-grained conflict data from macaques, we show that sparse coding, an important principle of neural compression, is an effective method for compressing collective behavior. The sparse code is shown to be efficient, predictive, and socially meaningful. In our monkey society, the sparse code of conflict is composed of related individuals, the policers, and the alpha female. Our results suggest that sparse coding is a natural technique for pattern extraction when cognitive constraints and small sample sizes limit the complexity of inferential models. Our approach highlights the need for cognitive experiments addressing how individuals perceive collective features of social organization. PMID:22891296
A multithread based new sparse matrix method in bioluminescence tomography
NASA Astrophysics Data System (ADS)
Zhang, Bo; Tian, Jie; Liu, Dan; Sun, Li; Yang, Xin; Han, Dong
2010-03-01
Among many molecular imaging modalities, bioluminescence tomography (BLT) stands out as an effective approach for in vivo imaging because of its noninvasive molecular and cellular level detection ability, high sensitivity and low cost in comparison with other imaging technologies. However, there exists the case that large scale problem with large number of points and elements in the structure of mesh standing for the small animal or phantom. And the large scale problem's system matrix generated by the diffuse approximation (DA) model using finite element method (FEM) is large. So there wouldn't be enough random access memory (RAM) for the program and the related inverse problem couldn't be solved. Considering the sparse property of the BLT system matrix, we've developed a new sparse matrix (ZSM) to overcome the problem. And the related algorithms have all been speeded up by multi-thread technologies. Then the inverse problem is solved by Tikhonov regularization method in adaptive finite element (AFE) framework. Finally, the performance of this method is tested on a heterogeneous phantom and the boundary data is obtained through Monte Carlo simulation. During the process of solving the forward model, the ZSM can save more processing time and memory space than the usual way, such as those not using sparse matrix and those using Triples or Cross Linked sparse matrix. Numerical experiments have shown when more CPU cores are used, the processing speed is increased. By incorporating ZSM, BLT can be applied to large scale problems with large system matrix.
Cerebellar Functional Parcellation Using Sparse Dictionary Learning Clustering.
Wang, Changqing; Kipping, Judy; Bao, Chenglong; Ji, Hui; Qiu, Anqi
2016-01-01
The human cerebellum has recently been discovered to contribute to cognition and emotion beyond the planning and execution of movement, suggesting its functional heterogeneity. We aimed to identify the functional parcellation of the cerebellum using information from resting-state functional magnetic resonance imaging (rs-fMRI). For this, we introduced a new data-driven decomposition-based functional parcellation algorithm, called Sparse Dictionary Learning Clustering (SDLC). SDLC integrates dictionary learning, sparse representation of rs-fMRI, and k-means clustering into one optimization problem. The dictionary is comprised of an over-complete set of time course signals, with which a sparse representation of rs-fMRI signals can be constructed. Cerebellar functional regions were then identified using k-means clustering based on the sparse representation of rs-fMRI signals. We solved SDLC using a multi-block hybrid proximal alternating method that guarantees strong convergence. We evaluated the reliability of SDLC and benchmarked its classification accuracy against other clustering techniques using simulated data. We then demonstrated that SDLC can identify biologically reasonable functional regions of the cerebellum as estimated by their cerebello-cortical functional connectivity. We further provided new insights into the cerebello-cortical functional organization in children.
Discriminative object tracking via sparse representation and online dictionary learning.
Xie, Yuan; Zhang, Wensheng; Li, Cuihua; Lin, Shuyang; Qu, Yanyun; Zhang, Yinghua
2014-04-01
We propose a robust tracking algorithm based on local sparse coding with discriminative dictionary learning and new keypoint matching schema. This algorithm consists of two parts: the local sparse coding with online updated discriminative dictionary for tracking (SOD part), and the keypoint matching refinement for enhancing the tracking performance (KP part). In the SOD part, the local image patches of the target object and background are represented by their sparse codes using an over-complete discriminative dictionary. Such discriminative dictionary, which encodes the information of both the foreground and the background, may provide more discriminative power. Furthermore, in order to adapt the dictionary to the variation of the foreground and background during the tracking, an online learning method is employed to update the dictionary. The KP part utilizes refined keypoint matching schema to improve the performance of the SOD. With the help of sparse representation and online updated discriminative dictionary, the KP part are more robust than the traditional method to reject the incorrect matches and eliminate the outliers. The proposed method is embedded into a Bayesian inference framework for visual tracking. Experimental results on several challenging video sequences demonstrate the effectiveness and robustness of our approach.
Energy Efficient Sparse Connectivity from Imbalanced Synaptic Plasticity Rules
Sacramento, João; Wichert, Andreas; van Rossum, Mark C. W.
2015-01-01
It is believed that energy efficiency is an important constraint in brain evolution. As synaptic transmission dominates energy consumption, energy can be saved by ensuring that only a few synapses are active. It is therefore likely that the formation of sparse codes and sparse connectivity are fundamental objectives of synaptic plasticity. In this work we study how sparse connectivity can result from a synaptic learning rule of excitatory synapses. Information is maximised when potentiation and depression are balanced according to the mean presynaptic activity level and the resulting fraction of zero-weight synapses is around 50%. However, an imbalance towards depression increases the fraction of zero-weight synapses without significantly affecting performance. We show that imbalanced plasticity corresponds to imposing a regularising constraint on the L 1-norm of the synaptic weight vector, a procedure that is well-known to induce sparseness. Imbalanced plasticity is biophysically plausible and leads to more efficient synaptic configurations than a previously suggested approach that prunes synapses after learning. Our framework gives a novel interpretation to the high fraction of silent synapses found in brain regions like the cerebellum. PMID:26046817
Half-quadratic-based iterative minimization for robust sparse representation.
He, Ran; Zheng, Wei-Shi; Tan, Tieniu; Sun, Zhenan
2014-02-01
Robust sparse representation has shown significant potential in solving challenging problems in computer vision such as biometrics and visual surveillance. Although several robust sparse models have been proposed and promising results have been obtained, they are either for error correction or for error detection, and learning a general framework that systematically unifies these two aspects and explores their relation is still an open problem. In this paper, we develop a half-quadratic (HQ) framework to solve the robust sparse representation problem. By defining different kinds of half-quadratic functions, the proposed HQ framework is applicable to performing both error correction and error detection. More specifically, by using the additive form of HQ, we propose an ℓ1-regularized error correction method by iteratively recovering corrupted data from errors incurred by noises and outliers; by using the multiplicative form of HQ, we propose an ℓ1-regularized error detection method by learning from uncorrupted data iteratively. We also show that the ℓ1-regularization solved by soft-thresholding function has a dual relationship to Huber M-estimator, which theoretically guarantees the performance of robust sparse representation in terms of M-estimation. Experiments on robust face recognition under severe occlusion and corruption validate our framework and findings.
Gene network-based cancer prognosis analysis with sparse boosting
Ma, Shuangge; Huang, Yuan; Huang, Jian; Fang, Kuangnan
2013-01-01
Summary High-throughput gene profiling studies have been extensively conducted, searching for markers associated with cancer development and progression. In this study, we analyse cancer prognosis studies with right censored survival responses. With gene expression data, we adopt the weighted gene co-expression network analysis (WGCNA) to describe the interplay among genes. In network analysis, nodes represent genes. There are subsets of nodes, called modules, which are tightly connected to each other. Genes within the same modules tend to have co-regulated biological functions. For cancer prognosis data with gene expression measurements, our goal is to identify cancer markers, while properly accounting for the network module structure. A two-step sparse boosting approach, called Network Sparse Boosting (NSBoost), is proposed for marker selection. In the first step, for each module separately, we use a sparse boosting approach for within-module marker selection and construct module-level ‘super markers ’. In the second step, we use the super markers to represent the effects of all genes within the same modules and conduct module-level selection using a sparse boosting approach. Simulation study shows that NSBoost can more accurately identify cancer-associated genes and modules than alternatives. In the analysis of breast cancer and lymphoma prognosis studies, NSBoost identifies genes with important biological implications. It outperforms alternatives including the boosting and penalization approaches by identifying a smaller number of genes/modules and/or having better prediction performance. PMID:22950901
STIS Sparse Field CTE test-internal {Cycle 10}
NASA Astrophysics Data System (ADS)
Goudfrooij, Paul
2001-07-01
CTE measurements are made using the "sparse field test", along both the serial and parallel axes. This program needs special commanding to provide {a} off-center MSM positionings of some slits, and {b} the ability to read out with any amplifier {A, B, C, or D}. All exposures are internals.
Cerebellar Functional Parcellation Using Sparse Dictionary Learning Clustering
Wang, Changqing; Kipping, Judy; Bao, Chenglong; Ji, Hui; Qiu, Anqi
2016-01-01
The human cerebellum has recently been discovered to contribute to cognition and emotion beyond the planning and execution of movement, suggesting its functional heterogeneity. We aimed to identify the functional parcellation of the cerebellum using information from resting-state functional magnetic resonance imaging (rs-fMRI). For this, we introduced a new data-driven decomposition-based functional parcellation algorithm, called Sparse Dictionary Learning Clustering (SDLC). SDLC integrates dictionary learning, sparse representation of rs-fMRI, and k-means clustering into one optimization problem. The dictionary is comprised of an over-complete set of time course signals, with which a sparse representation of rs-fMRI signals can be constructed. Cerebellar functional regions were then identified using k-means clustering based on the sparse representation of rs-fMRI signals. We solved SDLC using a multi-block hybrid proximal alternating method that guarantees strong convergence. We evaluated the reliability of SDLC and benchmarked its classification accuracy against other clustering techniques using simulated data. We then demonstrated that SDLC can identify biologically reasonable functional regions of the cerebellum as estimated by their cerebello-cortical functional connectivity. We further provided new insights into the cerebello-cortical functional organization in children. PMID:27199650
Sparse PDF Volumes for Consistent Multi-Resolution Volume Rendering
Sicat, Ronell; Krüger, Jens; Möller, Torsten; Hadwiger, Markus
2015-01-01
This paper presents a new multi-resolution volume representation called sparse pdf volumes, which enables consistent multi-resolution volume rendering based on probability density functions (pdfs) of voxel neighborhoods. These pdfs are defined in the 4D domain jointly comprising the 3D volume and its 1D intensity range. Crucially, the computation of sparse pdf volumes exploits data coherence in 4D, resulting in a sparse representation with surprisingly low storage requirements. At run time, we dynamically apply transfer functions to the pdfs using simple and fast convolutions. Whereas standard low-pass filtering and down-sampling incur visible differences between resolution levels, the use of pdfs facilitates consistent results independent of the resolution level used. We describe the efficient out-of-core computation of large-scale sparse pdf volumes, using a novel iterative simplification procedure of a mixture of 4D Gaussians. Finally, our data structure is optimized to facilitate interactive multi-resolution volume rendering on GPUs. PMID:26146475
Efficient Coordinated Recovery of Sparse Channels in Massive MIMO
NASA Astrophysics Data System (ADS)
Masood, Mudassir; Afify, Laila H.; Al-Naffouri, Tareq Y.
2015-01-01
This paper addresses the problem of estimating sparse channels in massive MIMO-OFDM systems. Most wireless channels are sparse in nature with large delay spread. In addition, these channels as observed by multiple antennas in a neighborhood have approximately common support. The sparsity and common support properties are attractive when it comes to the efficient estimation of large number of channels in massive MIMO systems. Moreover, to avoid pilot contamination and to achieve better spectral efficiency, it is important to use a small number of pilots. We present a novel channel estimation approach which utilizes the sparsity and common support properties to estimate sparse channels and require a small number of pilots. Two algorithms based on this approach have been developed which perform Bayesian estimates of sparse channels even when the prior is non-Gaussian or unknown. Neighboring antennas share among each other their beliefs about the locations of active channel taps to perform estimation. The coordinated approach improves channel estimates and also reduces the required number of pilots. Further improvement is achieved by the data-aided version of the algorithm. Extensive simulation results are provided to demonstrate the performance of the proposed algorithms.
An adaptive hierarchical sensing scheme for sparse signals
NASA Astrophysics Data System (ADS)
Schütze, Henry; Barth, Erhardt; Martinetz, Thomas
2014-02-01
In this paper, we present Adaptive Hierarchical Sensing (AHS), a novel adaptive hierarchical sensing algorithm for sparse signals. For a given but unknown signal with a sparse representation in an orthogonal basis, the sensing task is to identify its non-zero transform coefficients by performing only few measurements. A measurement is simply the inner product of the signal and a particular measurement vector. During sensing, AHS partially traverses a binary tree and performs one measurement per visited node. AHS is adaptive in the sense that after each measurement a decision is made whether the entire subtree of the current node is either further traversed or omitted depending on the measurement value. In order to acquire an N -dimensional signal that is K-sparse, AHS performs O(K log N/K) measurements. With AHS, the signal is easily reconstructed by a basis transform without the need to solve an optimization problem. When sensing full-size images, AHS can compete with a state-of-the-art compressed sensing approach in terms of reconstruction performance versus number of measurements. Additionally, we simulate the sensing of image patches by AHS and investigate the impact of the choice of the sparse coding basis as well as the impact of the tree composition.
Satellite-based studies on large-scale vegetation changes in China.
Zhao, Xia; Zhou, Daojing; Fang, Jingyun
2012-10-01
Remotely-sensed vegetation indices, which indicate the density and photosynthetic capacity of vegetation, have been widely used to monitor vegetation dynamics over broad areas. In this paper, we reviewed satellite-based studies on vegetation cover changes, biomass and productivity variations, phenological dynamics, desertification, and grassland degradation in China that occurred over the past 2-3 decades. Our review shows that the satellite-derived index (Normalized Difference Vegetation Index, NDVI) during growing season and the vegetation net primary productivity in major terrestrial ecosystems (for example forests, grasslands, shrubs, and croplands) have significantly increased, while the number of fresh lakes and vegetation coverage in urban regions have experienced a substantial decline. The start of the growing season continually advanced in China's temperate regions until the 1990s, with a large spatial heterogeneity. We also found that the coverage of sparsely-vegetated areas declined, and the NDVI per unit in vegetated areas increased in arid and semi-arid regions because of increased vegetation activity in grassland and oasis areas. However, these results depend strongly not only on the periods chosen for investigation, but also on factors such as data sources, changes in detection methods, and geospatial heterogeneity. Therefore, we should be cautious when applying remote sensing techniques to monitor vegetation structures, functions, and changes.
Towards robust and effective shape modeling: sparse shape composition.
Zhang, Shaoting; Zhan, Yiqiang; Dewan, Maneesh; Huang, Junzhou; Metaxas, Dimitris N; Zhou, Xiang Sean
2012-01-01
Organ shape plays an important role in various clinical practices, e.g., diagnosis, surgical planning and treatment evaluation. It is usually derived from low level appearance cues in medical images. However, due to diseases and imaging artifacts, low level appearance cues might be weak or misleading. In this situation, shape priors become critical to infer and refine the shape derived by image appearances. Effective modeling of shape priors is challenging because: (1) shape variation is complex and cannot always be modeled by a parametric probability distribution; (2) a shape instance derived from image appearance cues (input shape) may have gross errors; and (3) local details of the input shape are difficult to preserve if they are not statistically significant in the training data. In this paper we propose a novel Sparse Shape Composition model (SSC) to deal with these three challenges in a unified framework. In our method, a sparse set of shapes in the shape repository is selected and composed together to infer/refine an input shape. The a priori information is thus implicitly incorporated on-the-fly. Our model leverages two sparsity observations of the input shape instance: (1) the input shape can be approximately represented by a sparse linear combination of shapes in the shape repository; (2) parts of the input shape may contain gross errors but such errors are sparse. Our model is formulated as a sparse learning problem. Using L1 norm relaxation, it can be solved by an efficient expectation-maximization (EM) type of framework. Our method is extensively validated on two medical applications, 2D lung localization in X-ray images and 3D liver segmentation in low-dose CT scans. Compared to state-of-the-art methods, our model exhibits better performance in both studies. PMID:21963296
Deformable segmentation via sparse representation and dictionary learning.
Zhang, Shaoting; Zhan, Yiqiang; Metaxas, Dimitris N
2012-10-01
"Shape" and "appearance", the two pillars of a deformable model, complement each other in object segmentation. In many medical imaging applications, while the low-level appearance information is weak or mis-leading, shape priors play a more important role to guide a correct segmentation, thanks to the strong shape characteristics of biological structures. Recently a novel shape prior modeling method has been proposed based on sparse learning theory. Instead of learning a generative shape model, shape priors are incorporated on-the-fly through the sparse shape composition (SSC). SSC is robust to non-Gaussian errors and still preserves individual shape characteristics even when such characteristics is not statistically significant. Although it seems straightforward to incorporate SSC into a deformable segmentation framework as shape priors, the large-scale sparse optimization of SSC has low runtime efficiency, which cannot satisfy clinical requirements. In this paper, we design two strategies to decrease the computational complexity of SSC, making a robust, accurate and efficient deformable segmentation system. (1) When the shape repository contains a large number of instances, which is often the case in 2D problems, K-SVD is used to learn a more compact but still informative shape dictionary. (2) If the derived shape instance has a large number of vertices, which often appears in 3D problems, an affinity propagation method is used to partition the surface into small sub-regions, on which the sparse shape composition is performed locally. Both strategies dramatically decrease the scale of the sparse optimization problem and hence speed up the algorithm. Our method is applied on a diverse set of biomedical image analysis problems. Compared to the original SSC, these two newly-proposed modules not only significant reduce the computational complexity, but also improve the overall accuracy. PMID:22959839
Disentangling giant component and finite cluster contributions in sparse random matrix spectra.
Kühn, Reimer
2016-04-01
We describe a method for disentangling giant component and finite cluster contributions to sparse random matrix spectra, using sparse symmetric random matrices defined on Erdős-Rényi graphs as an example and test bed. Our methods apply to sparse matrices defined in terms of arbitrary graphs in the configuration model class, as long as they have finite mean degree. PMID:27176257
Disentangling giant component and finite cluster contributions in sparse random matrix spectra.
Kühn, Reimer
2016-04-01
We describe a method for disentangling giant component and finite cluster contributions to sparse random matrix spectra, using sparse symmetric random matrices defined on Erdős-Rényi graphs as an example and test bed. Our methods apply to sparse matrices defined in terms of arbitrary graphs in the configuration model class, as long as they have finite mean degree.
Soil and vegetation surveillance
Antonio, E.J.
1995-06-01
Soil sampling and analysis evaluates long-term contamination trends and monitors environmental radionuclide inventories. This section of the 1994 Hanford Site Environmental Report summarizes the soil and vegetation surveillance programs which were conducted during 1994. Vegetation surveillance is conducted offsite to monitor atmospheric deposition of radioactive materials in areas not under cultivation and onsite at locations adjacent to potential sources of radioactivity.
Sparse logistic regression with Lp penalty for biomarker identification.
Liu, Zhenqiu; Jiang, Feng; Tian, Guoliang; Wang, Suna; Sato, Fumiaki; Meltzer, Stephen J; Tan, Ming
2007-01-01
In this paper, we propose a novel method for sparse logistic regression with non-convex regularization Lp (p <1). Based on smooth approximation, we develop several fast algorithms for learning the classifier that is applicable to high dimensional dataset such as gene expression. To the best of our knowledge, these are the first algorithms to perform sparse logistic regression with an Lp and elastic net (Le) penalty. The regularization parameters are decided through maximizing the area under the ROC curve (AUC) of the test data. Experimental results on methylation and microarray data attest the accuracy, sparsity, and efficiency of the proposed algorithms. Biomarkers identified with our methods are compared with that in the literature. Our computational results show that Lp Logistic regression (p <1) outperforms the L1 logistic regression and SCAD SVM. Software is available upon request from the first author. PMID:17402921
Resistor array infrared nonuniformity correction based on sparse grid
NASA Astrophysics Data System (ADS)
He, Xudong; Qiu, Jiang; Zhang, Qiao; Du, Huijie; Zhao, Hongming
2013-10-01
Resistor array plays a vital role in emulation of the IR control and guide system. However, its serious nonuniformity confines the range of its application. Therefore, in order to obtain an available IR image, nonuniformity correction (NUC) is necessary. The traditional method is sparse grid and flood which only take the array's nonuniformity into account. In this paper we present an improved sparse grid method which considers the whole system which affects the array's nonuniformity by dividing the NUC process into different gray levels. In each gray level, we can take two points or several points to calculate the nonuniformity of every block which is divided before correction. After that, we can have several characteristic curves which will be operated with curve fitting. By this means, we will correct the nonuniformity. At last, through the experiment of a number of images, we find the residual nonuniformity associated with random noise is about 0.2% after the correction.
Sparse matrix methods research using the CSM testbed software system
NASA Technical Reports Server (NTRS)
Chu, Eleanor; George, J. Alan
1989-01-01
Research is described on sparse matrix techniques for the Computational Structural Mechanics (CSM) Testbed. The primary objective was to compare the performance of state-of-the-art techniques for solving sparse systems with those that are currently available in the CSM Testbed. Thus, one of the first tasks was to become familiar with the structure of the testbed, and to install some or all of the SPARSPAK package in the testbed. A suite of subroutines to extract from the data base the relevant structural and numerical information about the matrix equations was written, and all the demonstration problems distributed with the testbed were successfully solved. These codes were documented, and performance studies comparing the SPARSPAK technology to the methods currently in the testbed were completed. In addition, some preliminary studies were done comparing some recently developed out-of-core techniques with the performance of the testbed processor INV.
Sparse model selection in the highly under-sampled regime
NASA Astrophysics Data System (ADS)
Bulso, Nicola; Marsili, Matteo; Roudi, Yasser
2016-09-01
We propose a method for recovering the structure of a sparse undirected graphical model when very few samples are available. The method decides about the presence or absence of bonds between pairs of variable by considering one pair at a time and using a closed form formula, analytically derived by calculating the posterior probability for every possible model explaining a two body system using Jeffreys prior. The approach does not rely on the optimization of any cost functions and consequently is much faster than existing algorithms. Despite this time and computational advantage, numerical results show that for several sparse topologies the algorithm is comparable to the best existing algorithms, and is more accurate in the presence of hidden variables. We apply this approach to the analysis of US stock market data and to neural data, in order to show its efficiency in recovering robust statistical dependencies in real data with non-stationary correlations in time and/or space.
Supervised nonparametric sparse discriminant analysis for hyperspectral imagery classification
NASA Astrophysics Data System (ADS)
Wu, Longfei; Sun, Hao; Ji, Kefeng
2016-03-01
Owing to the high spectral sampling, the spectral information in hyperspectral imagery (HSI) is often highly correlated and contains redundancy. Motivated by the recent success of sparsity preserving based dimensionality reduction (DR) techniques in both computer vision and remote sensing image analysis community, a novel supervised nonparametric sparse discriminant analysis (NSDA) algorithm is presented for HSI classification. The objective function of NSDA aims at preserving the within-class sparse reconstructive relationship for within-class compactness characterization and maximizing the nonparametric between-class scatter simultaneously to enhance discriminative ability of the features in the projected space. Essentially, it seeks for the optimal projection matrix to identify the underlying discriminative manifold structure of a multiclass dataset. Experimental results on one visualization dataset and three recorded HSI dataset demonstrate NSDA outperforms several state-of-the-art feature extraction methods for HSI classification.
Hierarchy-Direction Selective Approach for Locally Adaptive Sparse Grids
Stoyanov, Miroslav K
2013-09-01
We consider the problem of multidimensional adaptive hierarchical interpolation. We use sparse grids points and functions that are induced from a one dimensional hierarchical rule via tensor products. The classical locally adaptive sparse grid algorithm uses an isotropic refinement from the coarser to the denser levels of the hierarchy. However, the multidimensional hierarchy provides a more complex structure that allows for various anisotropic and hierarchy selective refinement techniques. We consider the more advanced refinement techniques and apply them to a number of simple test functions chosen to demonstrate the various advantages and disadvantages of each method. While there is no refinement scheme that is optimal for all functions, the fully adaptive family-direction-selective technique is usually more stable and requires fewer samples.
Optimal Estimation and Rank Detection for Sparse Spiked Covariance Matrices
Cai, Tony; Ma, Zongming; Wu, Yihong
2014-01-01
This paper considers a sparse spiked covariancematrix model in the high-dimensional setting and studies the minimax estimation of the covariance matrix and the principal subspace as well as the minimax rank detection. The optimal rate of convergence for estimating the spiked covariance matrix under the spectral norm is established, which requires significantly different techniques from those for estimating other structured covariance matrices such as bandable or sparse covariance matrices. We also establish the minimax rate under the spectral norm for estimating the principal subspace, the primary object of interest in principal component analysis. In addition, the optimal rate for the rank detection boundary is obtained. This result also resolves the gap in a recent paper by Berthet and Rigollet [2] where the special case of rank one is considered. PMID:26257453
Color Sparse Representations for Image Processing: Review, Models, and Prospects.
Barthélemy, Quentin; Larue, Anthony; Mars, Jérôme I
2015-11-01
Sparse representations have been extended to deal with color images composed of three channels. A review of dictionary-learning-based sparse representations for color images is made here, detailing the differences between the models, and comparing their results on the real and simulated data. These models are considered in a unifying framework that is based on the degrees of freedom of the linear filtering/transformation of the color channels. Moreover, this allows it to be shown that the scalar quaternionic linear model is equivalent to constrained matrix-based color filtering, which highlights the filtering implicitly applied through this model. Based on this reformulation, the new color filtering model is introduced, using unconstrained filters. In this model, spatial morphologies of color images are encoded by atoms, and colors are encoded by color filters. Color variability is no longer captured in increasing the dictionary size, but with color filters, this gives an efficient color representation.
Accelerated Gibbs Sampling for Infinite Sparse Factor Analysis
Andrzejewski, D M
2011-09-12
The Indian Buffet Process (IBP) gives a probabilistic model of sparse binary matrices with an unbounded number of columns. This construct can be used, for example, to model a fixed numer of observed data points (rows) associated with an unknown number of latent features (columns). Markov Chain Monte Carlo (MCMC) methods are often used for IBP inference, and in this technical note, we provide a detailed review of the derivations of collapsed and accelerated Gibbs samplers for the linear-Gaussian infinite latent feature model. We also discuss and explain update equations for hyperparameter resampling in a 'full Bayesian' treatment and present a novel slice sampler capable of extending the accelerated Gibbs sampler to the case of infinite sparse factor analysis by allowing the use of real-valued latent features.
Scene-based nonuniformity correction using sparse prior
NASA Astrophysics Data System (ADS)
Mou, Xingang; Zhang, Guilin; Hu, Ruolan; Zhou, Xiao
2011-11-01
The performance of infrared focal plane array (IRFPA) is known to be affected by the presence of spatial fixed pattern noise (FPN) that is superimposed on the true image. Scene-based nonuniformity correction (NUC) algorithms are widely concerned since they only need the readout infrared data captured by the imaging system during its normal operation. A novel adaptive NUC algorithm is proposed using the sparse prior that when derivative filters are applied to infrared images, the filter outputs tends to be sparse. A change detection module based on results of derivative filters is introduced to avoid stationary object being learned into the background, so the ghosting artifact is eliminated effectively. The performance of the new algorithm is evaluated with both real and simulated imagery.
Estimation of sparse directed acyclic graphs for multivariate counts data.
Han, Sung Won; Zhong, Hua
2016-09-01
The next-generation sequencing data, called high-throughput sequencing data, are recorded as count data, which are generally far from normal distribution. Under the assumption that the count data follow the Poisson log-normal distribution, this article provides an L1-penalized likelihood framework and an efficient search algorithm to estimate the structure of sparse directed acyclic graphs (DAGs) for multivariate counts data. In searching for the solution, we use iterative optimization procedures to estimate the adjacency matrix and the variance matrix of the latent variables. The simulation result shows that our proposed method outperforms the approach which assumes multivariate normal distributions, and the log-transformation approach. It also shows that the proposed method outperforms the rank-based PC method under sparse network or hub network structures. As a real data example, we demonstrate the efficiency of the proposed method in estimating the gene regulatory networks of the ovarian cancer study. PMID:26849781
Efficient particle filtering via sparse kernel density estimation.
Banerjee, Amit; Burlina, Philippe
2010-09-01
Particle filters (PFs) are Bayesian filters capable of modeling nonlinear, non-Gaussian, and nonstationary dynamical systems. Recent research in PFs has investigated ways to appropriately sample from the posterior distribution, maintain multiple hypotheses, and alleviate computational costs while preserving tracking accuracy. To address these issues, a novel utilization of the support vector data description (SVDD) density estimation method within the particle filtering framework is presented. The SVDD density estimate can be integrated into a wide range of PFs to realize several benefits. It yields a sparse representation of the posterior density that reduces the computational complexity of the PF. The proposed approach also provides an analytical expression for the posterior distribution that can be used to identify its modes for maintaining multiple hypotheses and computing the MAP estimate, and to directly sample from the posterior. We present several experiments that demonstrate the advantages of incorporating a sparse kernel density estimate in a particle filter.
Universal regularizers for robust sparse coding and modeling.
Ramírez, Ignacio; Sapiro, Guillermo
2012-09-01
Sparse data models, where data is assumed to be well represented as a linear combination of a few elements from a dictionary, have gained considerable attention in recent years, and their use has led to state-of-the-art results in many signal and image processing tasks. It is now well understood that the choice of the sparsity regularization term is critical in the success of such models. Based on a codelength minimization interpretation of sparse coding, and using tools from universal coding theory, we propose a framework for designing sparsity regularization terms which have theoretical and practical advantages when compared with the more standard l(0) or l(1) ones. The presentation of the framework and theoretical foundations is complemented with examples that show its practical advantages in image denoising, zooming and classification.
Sparse cortical source localization using spatio-temporal atoms.
Korats, Gundars; Ranta, Radu; Le Cam, Steven; Louis-Dorr, Valérie
2015-01-01
This paper addresses the problem of sparse localization of cortical sources from scalp EEG recordings. Localization algorithms use propagation model under spatial and/or temporal constraints, but their performance highly depends on the data signal-to-noise ratio (SNR). In this work we propose a dictionary based sparse localization method which uses a data driven spatio-temporal dictionary to reconstruct the measurements using Single Best Replacement (SBR) and Continuation Single Best Replacement (CSBR) algorithms. We tested and compared our methods with the well-known MUSIC and RAP-MUSIC algorithms on simulated realistic data. Tests were carried out for different noise levels. The results show that our method has a strong advantage over MUSIC-type methods in case of synchronized sources. PMID:26737185
Application of sparse matrix solvers as effective preconditioners
Young, D.P.; Melvin, R.G.; Johnson, F.T.; Bussoletti, J.E.; Wigton, L.B.; Samant, S.S. )
1989-11-01
In this paper the use of a new out-of-core sparse matrix package for the numerical solution of partial differential equations involving complex geometries arising from aerospace applications is discussed. The sparse matrix solver accepts contributions to the matrix elements in random order and assembles the matrix using fast sort/merge routines. Fill-in is reduced through the use of a physically based nested dissection ordering. For very large problems a drop tolerance is used during the matrix decomposition phase. The resulting incomplete factorization is an effective preconditioner for Krylov subspace methods, such as GMRES. Problems involving 200,000 unknowns routinely are solved on the Cray X-MP using 64MW of solid-state storage device (SSD).
Mineral identification in hyperspectral imaging using Sparse-PCA
NASA Astrophysics Data System (ADS)
Yousefi, Bardia; Sojasi, Saeed; Ibarra Castanedo, Clemente; Beaudoin, Georges; Huot, François; Maldague, Xavier P. V.; Chamberland, Martin; Lalonde, Erik
2016-05-01
Hyperspectral imaging has been considerably developed during the recent decades. The application of hyperspectral imagery and infrared thermography, particularly for the automatic identification of minerals from satellite images, has been the subject of several interesting researches. In this study, a method is presented for the automated identification of the mineral grains typically used from satellite imagery and adapted for analyzing collected sample grains in a laboratory environment. For this, an approach involving Sparse Principle Components Analysis (SPCA) based on spectral abundance mapping techniques (i.e. SAM, SID, NormXCorr) is proposed for extraction of the representative spectral features. It develops an approximation of endmember as a reference spectrum process through the highest sparse principle component of the pure mineral grains. Subsequently, the features categorized by kernel Extreme Learning Machine (Kernel- ELM) classify and identify the mineral grains in a supervised manner. Classification is conducted in the binary scenario and the results indicate the dependency to the training spectra.
Sparse Multinomial Logistic Regression via Approximate Message Passing
NASA Astrophysics Data System (ADS)
Byrne, Evan; Schniter, Philip
2016-11-01
For the problem of multi-class linear classification and feature selection, we propose approximate message passing approaches to sparse multinomial logistic regression (MLR). First, we propose two algorithms based on the Hybrid Generalized Approximate Message Passing (HyGAMP) framework: one finds the maximum a posteriori (MAP) linear classifier and the other finds an approximation of the test-error-rate minimizing linear classifier. Then we design computationally simplified variants of these two algorithms. Next, we detail methods to tune the hyperparameters of their assumed statistical models using Stein's unbiased risk estimate (SURE) and expectation-maximization (EM), respectively. Finally, using both synthetic and real-world datasets, we demonstrate improved error-rate and runtime performance relative to existing state-of-the-art approaches to sparse MLR.
Minimum complexity FIR filters and sparse systolic arrays
Ferrari, L.A.; Sankar, P.V.
1988-06-01
The properties of BETA-spline approximation and the integral/derivative properties of convolution lead to efficient algorithms for the implementation of multidimensional FIR filters. The implementations are of minimum time complexity under the Nyquist criterion. The algorithm can easily be implemented using a sparse systolic array architecture. The resulting BETA-spline convolvers have much lower circuit complexity than systolic architectures based on conventional convolution algorithms. A two-dimensional hardware implementation based on simplifications of current architectures is presented.
Inverse sparse tracker with a locally weighted distance metric.
Wang, Dong; Lu, Huchuan; Xiao, Ziyang; Yang, Ming-Hsuan
2015-09-01
Sparse representation has been recently extensively studied for visual tracking and generally facilitates more accurate tracking results than classic methods. In this paper, we propose a sparsity-based tracking algorithm that is featured with two components: 1) an inverse sparse representation formulation and 2) a locally weighted distance metric. In the inverse sparse representation formulation, the target template is reconstructed with particles, which enables the tracker to compute the weights of all particles by solving only one l1 optimization problem and thereby provides a quite efficient model. This is in direct contrast to most previous sparse trackers that entail solving one optimization problem for each particle. However, we notice that this formulation with normal Euclidean distance metric is sensitive to partial noise like occlusion and illumination changes. To this end, we design a locally weighted distance metric to replace the Euclidean one. Similar ideas of using local features appear in other works, but only being supported by popular assumptions like local models could handle partial noise better than holistic models, without any solid theoretical analysis. In this paper, we attempt to explicitly explain it from a mathematical view. On that basis, we further propose a method to assign local weights by exploiting the temporal and spatial continuity. In the proposed method, appearance changes caused by partial occlusion and shape deformation are carefully considered, thereby facilitating accurate similarity measurement and model update. The experimental validation is conducted from two aspects: 1) self validation on key components and 2) comparison with other state-of-the-art algorithms. Results over 15 challenging sequences show that the proposed tracking algorithm performs favorably against the existing sparsity-based trackers and the other state-of-the-art methods. PMID:25935033
Spatially Sparse, Temporally Smooth MEG Via Vector ℓ0 .
Cassidy, Ben; Solo, Victor
2015-06-01
In this paper, we describe a new method for solving the magnetoencephalography inverse problem: temporal vector ℓ0-penalized least squares (TV-L0LS). The method calculates maximally sparse current dipole magnitudes and directions via spatial ℓ0 regularization on a cortically-distributed source grid, while constraining the solution to be smooth with respect to time. We demonstrate the utility of this method on real and simulated data by comparison to existing methods.
Parallel iterative methods for sparse linear and nonlinear equations
NASA Technical Reports Server (NTRS)
Saad, Youcef
1989-01-01
As three-dimensional models are gaining importance, iterative methods will become almost mandatory. Among these, preconditioned Krylov subspace methods have been viewed as the most efficient and reliable, when solving linear as well as nonlinear systems of equations. There has been several different approaches taken to adapt iterative methods for supercomputers. Some of these approaches are discussed and the methods that deal more specifically with general unstructured sparse matrices, such as those arising from finite element methods, are emphasized.
Blind deconvolution of images using optimal sparse representations.
Bronstein, Michael M; Bronstein, Alexander M; Zibulevsky, Michael; Zeevi, Yehoshua Y
2005-06-01
The relative Newton algorithm, previously proposed for quasi-maximum likelihood blind source separation and blind deconvolution of one-dimensional signals is generalized for blind deconvolution of images. Smooth approximation of the absolute value is used as the nonlinear term for sparse sources. In addition, we propose a method of sparsification, which allows blind deconvolution of arbitrary sources, and show how to find optimal sparsifying transformations by supervised learning.
A multi-level method for sparse linear systems
Shapira, Y.
1997-09-01
A multi-level method for the solution of sparse linear systems is introduced. The definition of the method is based on data from the coefficient matrix alone. An upper bound for the condition number is available for certain symmetric positive definite (SPD) problems. Numerical experiments confirm the analysis and illustrate the efficiency of the method for diffusion problems with discontinuous coefficients with discontinuities which are not aligned with the coarse meshes.
Shape prior modeling using sparse representation and online dictionary learning.
Zhang, Shaoting; Zhan, Yiqiang; Zhou, Yan; Uzunbas, Mustafa; Metaxas, Dimitris N
2012-01-01
The recently proposed sparse shape composition (SSC) opens a new avenue for shape prior modeling. Instead of assuming any parametric model of shape statistics, SSC incorporates shape priors on-the-fly by approximating a shape instance (usually derived from appearance cues) by a sparse combination of shapes in a training repository. Theoretically, one can increase the modeling capability of SSC by including as many training shapes in the repository. However, this strategy confronts two limitations in practice. First, since SSC involves an iterative sparse optimization at run-time, the more shape instances contained in the repository, the less run-time efficiency SSC has. Therefore, a compact and informative shape dictionary is preferred to a large shape repository. Second, in medical imaging applications, training shapes seldom come in one batch. It is very time consuming and sometimes infeasible to reconstruct the shape dictionary every time new training shapes appear. In this paper, we propose an online learning method to address these two limitations. Our method starts from constructing an initial shape dictionary using the K-SVD algorithm. When new training shapes come, instead of re-constructing the dictionary from the ground up, we update the existing one using a block-coordinates descent approach. Using the dynamically updated dictionary, sparse shape composition can be gracefully scaled up to model shape priors from a large number of training shapes without sacrificing run-time efficiency. Our method is validated on lung localization in X-Ray and cardiac segmentation in MRI time series. Compared to the original SSC, it shows comparable performance while being significantly more efficient. PMID:23286160
Partitioning Rectangular and Structurally Nonsymmetric Sparse Matrices for Parallel Processing
B. Hendrickson; T.G. Kolda
1998-09-01
A common operation in scientific computing is the multiplication of a sparse, rectangular or structurally nonsymmetric matrix and a vector. In many applications the matrix- transpose-vector product is also required. This paper addresses the efficient parallelization of these operations. We show that the problem can be expressed in terms of partitioning bipartite graphs. We then introduce several algorithms for this partitioning problem and compare their performance on a set of test matrices.
Reconstruction techniques for sparse multistatic linear array microwave imaging
NASA Astrophysics Data System (ADS)
Sheen, David M.; Hall, Thomas E.
2014-06-01
Sequentially-switched linear arrays are an enabling technology for a number of near-field microwave imaging applications. Electronically sequencing along the array axis followed by mechanical scanning along an orthogonal axis allows dense sampling of a two-dimensional aperture in near real-time. The Pacific Northwest National Laboratory (PNNL) has developed this technology for several applications including concealed weapon detection, groundpenetrating radar, and non-destructive inspection and evaluation. These techniques form three-dimensional images by scanning a diverging beam swept frequency transceiver over a two-dimensional aperture and mathematically focusing or reconstructing the data into three-dimensional images. Recently, a sparse multi-static array technology has been developed that reduces the number of antennas required to densely sample the linear array axis of the spatial aperture. This allows a significant reduction in cost and complexity of the linear-array-based imaging system. The sparse array has been specifically designed to be compatible with Fourier-Transform-based image reconstruction techniques; however, there are limitations to the use of these techniques, especially for extreme near-field operation. In the extreme near-field of the array, back-projection techniques have been developed that account for the exact location of each transmitter and receiver in the linear array and the 3-D image location. In this paper, the sparse array technique will be described along with associated Fourier-Transform-based and back-projection-based image reconstruction algorithms. Simulated imaging results are presented that show the effectiveness of the sparse array technique along with the merits and weaknesses of each image reconstruction approach.
Mathematical strategies for filtering complex systems: Regularly spaced sparse observations
Harlim, J. Majda, A.J.
2008-05-01
Real time filtering of noisy turbulent signals through sparse observations on a regularly spaced mesh is a notoriously difficult and important prototype filtering problem. Simpler off-line test criteria are proposed here as guidelines for filter performance for these stiff multi-scale filtering problems in the context of linear stochastic partial differential equations with turbulent solutions. Filtering turbulent solutions of the stochastically forced dissipative advection equation through sparse observations is developed as a stringent test bed for filter performance with sparse regular observations. The standard ensemble transform Kalman filter (ETKF) has poor skill on the test bed and even suffers from filter divergence, surprisingly, at observable times with resonant mean forcing and a decaying energy spectrum in the partially observed signal. Systematic alternative filtering strategies are developed here including the Fourier Domain Kalman Filter (FDKF) and various reduced filters called Strongly Damped Approximate Filter (SDAF), Variance Strongly Damped Approximate Filter (VSDAF), and Reduced Fourier Domain Kalman Filter (RFDKF) which operate only on the primary Fourier modes associated with the sparse observation mesh while nevertheless, incorporating into the approximate filter various features of the interaction with the remaining modes. It is shown below that these much cheaper alternative filters have significant skill on the test bed of turbulent solutions which exceeds ETKF and in various regimes often exceeds FDKF, provided that the approximate filters are guided by the off-line test criteria. The skill of the various approximate filters depends on the energy spectrum of the turbulent signal and the observation time relative to the decorrelation time of the turbulence at a given spatial scale in a precise fashion elucidated here.
Neogene biomarker record of vegetation change in eastern Africa
Polissar, Pratigya J.; Jackson, Kevin E.; deMenocal, Peter B.
2016-01-01
The evolution of C4 grassland ecosystems in eastern Africa has been intensely studied because of the potential influence of vegetation on mammalian evolution, including that of our own lineage, hominins. Although a handful of sparse vegetation records exists from middle and early Miocene terrestrial fossil sites, there is no comprehensive record of vegetation through the Neogene. Here we present a vegetation record spanning the Neogene and Quaternary Periods that documents the appearance and subsequent expansion of C4 grasslands in eastern Africa. Carbon isotope ratios from terrestrial plant wax biomarkers deposited in marine sediments indicate constant C3 vegetation from ∼24 Ma to 10 Ma, when C4 grasses first appeared. From this time forward, C4 vegetation increases monotonically to present, with a coherent signal between marine core sites located in the Somali Basin and the Red Sea. The response of mammalian herbivores to the appearance of C4 grasses at 10 Ma is immediate, as evidenced from existing records of mammalian diets from isotopic analyses of tooth enamel. The expansion of C4 vegetation in eastern Africa is broadly mirrored by increasing proportions of C4-based foods in hominin diets, beginning at 3.8 Ma in Australopithecus and, slightly later, Kenyanthropus. This continues into the late Pleistocene in Paranthropus, whereas Homo maintains a flexible diet. The biomarker vegetation record suggests the increase in open, C4 grassland ecosystems over the last 10 Ma may have operated as a selection pressure for traits and behaviors in Homo such as bipedalism, flexible diets, and complex social structure. PMID:27274042
Neogene biomarker record of vegetation change in eastern Africa.
Uno, Kevin T; Polissar, Pratigya J; Jackson, Kevin E; deMenocal, Peter B
2016-06-01
The evolution of C4 grassland ecosystems in eastern Africa has been intensely studied because of the potential influence of vegetation on mammalian evolution, including that of our own lineage, hominins. Although a handful of sparse vegetation records exists from middle and early Miocene terrestrial fossil sites, there is no comprehensive record of vegetation through the Neogene. Here we present a vegetation record spanning the Neogene and Quaternary Periods that documents the appearance and subsequent expansion of C4 grasslands in eastern Africa. Carbon isotope ratios from terrestrial plant wax biomarkers deposited in marine sediments indicate constant C3 vegetation from ∼24 Ma to 10 Ma, when C4 grasses first appeared. From this time forward, C4 vegetation increases monotonically to present, with a coherent signal between marine core sites located in the Somali Basin and the Red Sea. The response of mammalian herbivores to the appearance of C4 grasses at 10 Ma is immediate, as evidenced from existing records of mammalian diets from isotopic analyses of tooth enamel. The expansion of C4 vegetation in eastern Africa is broadly mirrored by increasing proportions of C4-based foods in hominin diets, beginning at 3.8 Ma in Australopithecus and, slightly later, Kenyanthropus This continues into the late Pleistocene in Paranthropus, whereas Homo maintains a flexible diet. The biomarker vegetation record suggests the increase in open, C4 grassland ecosystems over the last 10 Ma may have operated as a selection pressure for traits and behaviors in Homo such as bipedalism, flexible diets, and complex social structure. PMID:27274042
Neogene biomarker record of vegetation change in eastern Africa.
Uno, Kevin T; Polissar, Pratigya J; Jackson, Kevin E; deMenocal, Peter B
2016-06-01
The evolution of C4 grassland ecosystems in eastern Africa has been intensely studied because of the potential influence of vegetation on mammalian evolution, including that of our own lineage, hominins. Although a handful of sparse vegetation records exists from middle and early Miocene terrestrial fossil sites, there is no comprehensive record of vegetation through the Neogene. Here we present a vegetation record spanning the Neogene and Quaternary Periods that documents the appearance and subsequent expansion of C4 grasslands in eastern Africa. Carbon isotope ratios from terrestrial plant wax biomarkers deposited in marine sediments indicate constant C3 vegetation from ∼24 Ma to 10 Ma, when C4 grasses first appeared. From this time forward, C4 vegetation increases monotonically to present, with a coherent signal between marine core sites located in the Somali Basin and the Red Sea. The response of mammalian herbivores to the appearance of C4 grasses at 10 Ma is immediate, as evidenced from existing records of mammalian diets from isotopic analyses of tooth enamel. The expansion of C4 vegetation in eastern Africa is broadly mirrored by increasing proportions of C4-based foods in hominin diets, beginning at 3.8 Ma in Australopithecus and, slightly later, Kenyanthropus This continues into the late Pleistocene in Paranthropus, whereas Homo maintains a flexible diet. The biomarker vegetation record suggests the increase in open, C4 grassland ecosystems over the last 10 Ma may have operated as a selection pressure for traits and behaviors in Homo such as bipedalism, flexible diets, and complex social structure.
Accelerating sparse linear algebra using graphics processing units
NASA Astrophysics Data System (ADS)
Spagnoli, Kyle E.; Humphrey, John R.; Price, Daniel K.; Kelmelis, Eric J.
2011-06-01
The modern graphics processing unit (GPU) found in many standard personal computers is a highly parallel math processor capable of over 1 TFLOPS of peak computational throughput at a cost similar to a high-end CPU with excellent FLOPS-to-watt ratio. High-level sparse linear algebra operations are computationally intense, often requiring large amounts of parallel operations and would seem a natural fit for the processing power of the GPU. Our work is on a GPU accelerated implementation of sparse linear algebra routines. We present results from both direct and iterative sparse system solvers. The GPU execution model featured by NVIDIA GPUs based on CUDA demands very strong parallelism, requiring between hundreds and thousands of simultaneous operations to achieve high performance. Some constructs from linear algebra map extremely well to the GPU and others map poorly. CPUs, on the other hand, do well at smaller order parallelism and perform acceptably during low-parallelism code segments. Our work addresses this via hybrid a processing model, in which the CPU and GPU work simultaneously to produce results. In many cases, this is accomplished by allowing each platform to do the work it performs most naturally. For example, the CPU is responsible for graph theory portion of the direct solvers while the GPU simultaneously performs the low level linear algebra routines.
Inversion of magnetotelluric data in a sparse model domain
NASA Astrophysics Data System (ADS)
Nittinger, Christian G.; Becken, Michael
2016-06-01
The inversion of magnetotelluric data into subsurface electrical conductivity poses an ill-posed problem. Smoothing constraints are widely employed to estimate a regularized solution. Here, we present an alternative inversion scheme that estimates a sparse representation of the model in a wavelet basis. The objective of the inversion is to determine the few non-zero wavelet coefficients which are required to fit the data. This approach falls into the class of sparsity constrained inversion schemes and minimizes the combination of the data misfit in a least squares ℓ2 sense and of a model coefficient norm in a ℓ1 sense (ℓ2-ℓ1 minimization). The ℓ1 coefficient norm renders the solution sparse in a suitable representation such as the multi-resolution wavelet basis, but does not impose explicit structural penalties on the model as it is the case for ℓ2 regularization. The presented numerical algorithm solves the mixed ℓ2-ℓ1 norm minimization problem for the non-linear magnetotelluric inverse problem. We demonstrate the feasibility of our algorithm on synthetic 2-D MT data as well as on a real data example. We found that sparse models can be estimated by inversion and that the spatial distribution of non-vanishing coefficients indicates regions in the model which are resolved.
Inversion of magnetotelluric data in a sparse model domain
NASA Astrophysics Data System (ADS)
Nittinger, Christian G.; Becken, Michael
2016-08-01
The inversion of magnetotelluric data into subsurface electrical conductivity poses an ill-posed problem. Smoothing constraints are widely employed to estimate a regularized solution. Here, we present an alternative inversion scheme that estimates a sparse representation of the model in a wavelet basis. The objective of the inversion is to determine the few non-zero wavelet coefficients which are required to fit the data. This approach falls into the class of sparsity constrained inversion schemes and minimizes the combination of the data misfit in a least-squares ℓ2 sense and of a model coefficient norm in an ℓ1 sense (ℓ2-ℓ1 minimization). The ℓ1 coefficient norm renders the solution sparse in a suitable representation such as the multiresolution wavelet basis, but does not impose explicit structural penalties on the model as it is the case for ℓ2 regularization. The presented numerical algorithm solves the mixed ℓ2-ℓ1 norm minimization problem for the nonlinear magnetotelluric inverse problem. We demonstrate the feasibility of our algorithm on synthetic 2-D MT data as well as on a real data example. We found that sparse models can be estimated by inversion and that the spatial distribution of non-vanishing coefficients indicates regions in the model which are resolved.
Supervised Discriminative Group Sparse Representation for Mild Cognitive Impairment Diagnosis.
Suk, Heung-Il; Wee, Chong-Yaw; Lee, Seong-Whan; Shen, Dinggang
2015-07-01
Research on an early detection of Mild Cognitive Impairment (MCI), a prodromal stage of Alzheimer's Disease (AD), with resting-state functional Magnetic Resonance Imaging (rs-fMRI) has been of great interest for the last decade. Witnessed by recent studies, functional connectivity is a useful concept in extracting brain network features and finding biomarkers for brain disease diagnosis. However, it still remains challenging for the estimation of functional connectivity from rs-fMRI due to the inevitable high dimensional problem. In order to tackle this problem, we utilize a group sparse representation along with a structural equation model. Unlike the conventional group sparse representation method that does not explicitly consider class-label information, which can help enhance the diagnostic performance, in this paper, we propose a novel supervised discriminative group sparse representation method by penalizing a large within-class variance and a small between-class variance of connectivity coefficients. Thanks to the newly devised penalization terms, we can learn connectivity coefficients that are similar within the same class and distinct between classes, thus helping enhance the diagnostic accuracy. The proposed method also allows the learned common network structure to preserve the network specific and label-related characteristics. In our experiments on the rs-fMRI data of 37 subjects (12 MCI; 25 healthy normal control) with a cross-validation technique, we demonstrated the validity and effectiveness of the proposed method, showing the diagnostic accuracy of 89.19 % and the sensitivity of 0.9167.
Medical image registration using sparse coding of image patches.
Afzali, Maryam; Ghaffari, Aboozar; Fatemizadeh, Emad; Soltanian-Zadeh, Hamid
2016-06-01
Image registration is a basic task in medical image processing applications like group analysis and atlas construction. Similarity measure is a critical ingredient of image registration. Intensity distortion of medical images is not considered in most previous similarity measures. Therefore, in the presence of bias field distortions, they do not generate an acceptable registration. In this paper, we propose a sparse based similarity measure for mono-modal images that considers non-stationary intensity and spatially-varying distortions. The main idea behind this measure is that the aligned image is constructed by an analysis dictionary trained using the image patches. For this purpose, we use "Analysis K-SVD" to train the dictionary and find the sparse coefficients. We utilize image patches to construct the analysis dictionary and then we employ the proposed sparse similarity measure to find a non-rigid transformation using free form deformation (FFD). Experimental results show that the proposed approach is able to robustly register 2D and 3D images in both simulated and real cases. The proposed method outperforms other state-of-the-art similarity measures and decreases the transformation error compared to the previous methods. Even in the presence of bias field distortion, the proposed method aligns images without any preprocessing. PMID:27085311
Medical image registration using sparse coding of image patches.
Afzali, Maryam; Ghaffari, Aboozar; Fatemizadeh, Emad; Soltanian-Zadeh, Hamid
2016-06-01
Image registration is a basic task in medical image processing applications like group analysis and atlas construction. Similarity measure is a critical ingredient of image registration. Intensity distortion of medical images is not considered in most previous similarity measures. Therefore, in the presence of bias field distortions, they do not generate an acceptable registration. In this paper, we propose a sparse based similarity measure for mono-modal images that considers non-stationary intensity and spatially-varying distortions. The main idea behind this measure is that the aligned image is constructed by an analysis dictionary trained using the image patches. For this purpose, we use "Analysis K-SVD" to train the dictionary and find the sparse coefficients. We utilize image patches to construct the analysis dictionary and then we employ the proposed sparse similarity measure to find a non-rigid transformation using free form deformation (FFD). Experimental results show that the proposed approach is able to robustly register 2D and 3D images in both simulated and real cases. The proposed method outperforms other state-of-the-art similarity measures and decreases the transformation error compared to the previous methods. Even in the presence of bias field distortion, the proposed method aligns images without any preprocessing.
Dentate Gyrus Circuitry Features Improve Performance of Sparse Approximation Algorithms
Petrantonakis, Panagiotis C.; Poirazi, Panayiota
2015-01-01
Memory-related activity in the Dentate Gyrus (DG) is characterized by sparsity. Memory representations are seen as activated neuronal populations of granule cells, the main encoding cells in DG, which are estimated to engage 2–4% of the total population. This sparsity is assumed to enhance the ability of DG to perform pattern separation, one of the most valuable contributions of DG during memory formation. In this work, we investigate how features of the DG such as its excitatory and inhibitory connectivity diagram can be used to develop theoretical algorithms performing Sparse Approximation, a widely used strategy in the Signal Processing field. Sparse approximation stands for the algorithmic identification of few components from a dictionary that approximate a certain signal. The ability of DG to achieve pattern separation by sparsifing its representations is exploited here to improve the performance of the state of the art sparse approximation algorithm “Iterative Soft Thresholding” (IST) by adding new algorithmic features inspired by the DG circuitry. Lateral inhibition of granule cells, either direct or indirect, via mossy cells, is shown to enhance the performance of the IST. Apart from revealing the potential of DG-inspired theoretical algorithms, this work presents new insights regarding the function of particular cell types in the pattern separation task of the DG. PMID:25635776
Very sparse LSSVM reductions for large-scale data.
Mall, Raghvendra; Suykens, Johan A K
2015-05-01
Least squares support vector machines (LSSVMs) have been widely applied for classification and regression with comparable performance with SVMs. The LSSVM model lacks sparsity and is unable to handle large-scale data due to computational and memory constraints. A primal fixed-size LSSVM (PFS-LSSVM) introduce sparsity using Nyström approximation with a set of prototype vectors (PVs). The PFS-LSSVM model solves an overdetermined system of linear equations in the primal. However, this solution is not the sparsest. We investigate the sparsity-error tradeoff by introducing a second level of sparsity. This is done by means of L0 -norm-based reductions by iteratively sparsifying LSSVM and PFS-LSSVM models. The exact choice of the cardinality for the initial PV set is not important then as the final model is highly sparse. The proposed method overcomes the problem of memory constraints and high computational costs resulting in highly sparse reductions to LSSVM models. The approximations of the two models allow to scale the models to large-scale datasets. Experiments on real-world classification and regression data sets from the UCI repository illustrate that these approaches achieve sparse models without a significant tradeoff in errors.
Polygenic Modeling with Bayesian Sparse Linear Mixed Models
Zhou, Xiang; Carbonetto, Peter; Stephens, Matthew
2013-01-01
Both linear mixed models (LMMs) and sparse regression models are widely used in genetics applications, including, recently, polygenic modeling in genome-wide association studies. These two approaches make very different assumptions, so are expected to perform well in different situations. However, in practice, for a given dataset one typically does not know which assumptions will be more accurate. Motivated by this, we consider a hybrid of the two, which we refer to as a “Bayesian sparse linear mixed model” (BSLMM) that includes both these models as special cases. We address several key computational and statistical issues that arise when applying BSLMM, including appropriate prior specification for the hyper-parameters and a novel Markov chain Monte Carlo algorithm for posterior inference. We apply BSLMM and compare it with other methods for two polygenic modeling applications: estimating the proportion of variance in phenotypes explained (PVE) by available genotypes, and phenotype (or breeding value) prediction. For PVE estimation, we demonstrate that BSLMM combines the advantages of both standard LMMs and sparse regression modeling. For phenotype prediction it considerably outperforms either of the other two methods, as well as several other large-scale regression methods previously suggested for this problem. Software implementing our method is freely available from http://stephenslab.uchicago.edu/software.html. PMID:23408905
Sparse matrix-vector multiplication on a reconfigurable supercomputer
Dubois, David H; Dubois, Andrew J; Boorman, Thomas M; Connor, Carolyn M; Poole, Steve
2008-01-01
Double precision floating point Sparse Matrix-Vector Multiplication (SMVM) is a critical computational kernel used in iterative solvers for systems of sparse linear equations. The poor data locality exhibited by sparse matrices along with the high memory bandwidth requirements of SMVM result in poor performance on general purpose processors. Field Programmable Gate Arrays (FPGAs) offer a possible alternative with their customizable and application-targeted memory sub-system and processing elements. In this work we investigate two separate implementations of the SMVM on an SRC-6 MAPStation workstation. The first implementation investigates the peak performance capability, while the second implementation balances the amount of instantiated logic with the available sustained bandwidth of the FPGA subsystem. Both implementations yield the same sustained performance with the second producing a much more efficient solution. The metrics of processor and application balance are introduced to help provide some insight into the efficiencies of the FPGA and CPU based solutions explicitly showing the tight coupling of the available bandwidth to peak floating point performance. Due to the FPGA's ability to balance the amount of implemented logic to the available memory bandwidth it can provide a much more efficient solution. Finally, making use of the lessons learned implementing the SMVM, we present an fully implemented nonpreconditioned Conjugate Gradient Algorithm utilizing the second SMVM design.
Efficient MATLAB computations with sparse and factored tensors.
Bader, Brett William; Kolda, Tamara Gibson (Sandia National Lab, Livermore, CA)
2006-12-01
In this paper, the term tensor refers simply to a multidimensional or N-way array, and we consider how specially structured tensors allow for efficient storage and computation. First, we study sparse tensors, which have the property that the vast majority of the elements are zero. We propose storing sparse tensors using coordinate format and describe the computational efficiency of this scheme for various mathematical operations, including those typical to tensor decomposition algorithms. Second, we study factored tensors, which have the property that they can be assembled from more basic components. We consider two specific types: a Tucker tensor can be expressed as the product of a core tensor (which itself may be dense, sparse, or factored) and a matrix along each mode, and a Kruskal tensor can be expressed as the sum of rank-1 tensors. We are interested in the case where the storage of the components is less than the storage of the full tensor, and we demonstrate that many elementary operations can be computed using only the components. All of the efficiencies described in this paper are implemented in the Tensor Toolbox for MATLAB.
Treatment of non-sparse cratering in planetary surface dating
NASA Astrophysics Data System (ADS)
Kneissl, T.; Michael, G. G.; Schmedemann, N.
2016-10-01
We here propose a new technique to derive crater size-frequency distributions (CSFDs) from non-sparsely cratered surfaces, by accounting for the loss of craters due to subsequent crater/ejecta coverage. This approach, which we refer to as the buffered non-sparseness correction (BNSC), relates each crater to a measurement area found by excluding regions in the study area that have been resurfaced by larger craters and their ejecta blankets. The approach includes the well-known buffered crater counting (BCC) technique in order to consider the potential identification of craters whose centers are located outside the counting area. We demonstrate the new approach at two test sites on the Moon, one on the ancient lunar highlands outside the South Pole Aitken basin and the other on the much younger surface of lunar Mare Serenitatis. As expected, the correction has a much stronger effect on ancient, densely cratered surfaces than on younger, sparsely cratered surfaces. Furthermore, these first results indicate that the shapes of CSFDs on ancient terrains are actually very similar to the shapes of CSFDs on younger terrains.
Compressive Sensing Based Design of Sparse Tripole Arrays.
Hawes, Matthew; Liu, Wei; Mihaylova, Lyudmila
2015-12-10
This paper considers the problem of designing sparse linear tripole arrays. In such arrays at each antenna location there are three orthogonal dipoles, allowing full measurement of both the horizontal and vertical components of the received waveform. We formulate this problem from the viewpoint of Compressive Sensing (CS). However, unlike for isotropic array elements (single antenna), we now have three complex valued weight coefficients associated with each potential location (due to the three dipoles), which have to be simultaneously minimised. If this is not done, we may only set the weight coefficients of individual dipoles to be zero valued, rather than complete tripoles, meaning some dipoles may remain at each location. Therefore, the contributions of this paper are to formulate the design of sparse tripole arrays as an optimisation problem, and then we obtain a solution based on the minimisation of a modified l1 norm or a series of iteratively solved reweighted minimisations, which ensure a truly sparse solution. Design examples are provided to verify the effectiveness of the proposed methods and show that a good approximation of a reference pattern can be achieved using fewer tripoles than a Uniform Linear Array (ULA) of equivalent length.
Compressive Sensing Based Design of Sparse Tripole Arrays
Hawes, Matthew; Liu, Wei; Mihaylova, Lyudmila
2015-01-01
This paper considers the problem of designing sparse linear tripole arrays. In such arrays at each antenna location there are three orthogonal dipoles, allowing full measurement of both the horizontal and vertical components of the received waveform. We formulate this problem from the viewpoint of Compressive Sensing (CS). However, unlike for isotropic array elements (single antenna), we now have three complex valued weight coefficients associated with each potential location (due to the three dipoles), which have to be simultaneously minimised. If this is not done, we may only set the weight coefficients of individual dipoles to be zero valued, rather than complete tripoles, meaning some dipoles may remain at each location. Therefore, the contributions of this paper are to formulate the design of sparse tripole arrays as an optimisation problem, and then we obtain a solution based on the minimisation of a modified l1 norm or a series of iteratively solved reweighted minimisations, which ensure a truly sparse solution. Design examples are provided to verify the effectiveness of the proposed methods and show that a good approximation of a reference pattern can be achieved using fewer tripoles than a Uniform Linear Array (ULA) of equivalent length. PMID:26690436
Cellular Adaptation Facilitates Sparse and Reliable Coding in Sensory Pathways
Farkhooi, Farzad; Froese, Anja; Muller, Eilif; Menzel, Randolf; Nawrot, Martin P.
2013-01-01
Most neurons in peripheral sensory pathways initially respond vigorously when a preferred stimulus is presented, but adapt as stimulation continues. It is unclear how this phenomenon affects stimulus coding in the later stages of sensory processing. Here, we show that a temporally sparse and reliable stimulus representation develops naturally in sequential stages of a sensory network with adapting neurons. As a modeling framework we employ a mean-field approach together with an adaptive population density treatment, accompanied by numerical simulations of spiking neural networks. We find that cellular adaptation plays a critical role in the dynamic reduction of the trial-by-trial variability of cortical spike responses by transiently suppressing self-generated fast fluctuations in the cortical balanced network. This provides an explanation for a widespread cortical phenomenon by a simple mechanism. We further show that in the insect olfactory system cellular adaptation is sufficient to explain the emergence of the temporally sparse and reliable stimulus representation in the mushroom body. Our results reveal a generic, biophysically plausible mechanism that can explain the emergence of a temporally sparse and reliable stimulus representation within a sequential processing architecture. PMID:24098101
Scalable Library for the Parallel Solution of Sparse Linear Systems
1993-07-14
BlockSolve is a scalable parallel software library for the solution of large sparse, symmetric systems of linear equations. It runs on a variety of parallel architectures and can easily be ported to others. BlockSovle is primarily intended for the solution of sparse linear systems that arise from physical problems having multiple degrees of freedom at each node point. For example, when the finite element method is used to solve practical problems in structural engineering, eachmore » node will typically have anywhere from 3-6 degrees of freedom associated with it. BlockSolve is written to take advantage of problems of this nature; however, it is still reasonably efficient for problems that have only one degree of freedom associated with each node, such as the three-dimensional Poisson problem. It does not require that the matrices have any particular structure other than being sparse and symmetric. BlockSolve is intended to be used within real application codes. It is designed to work best in the context of our experience which indicated that most application codes solve the same linear systems with several different right-hand sides and/or linear systems with the same structure, but different matrix values multiple times.« less
Crack growth sparse pursuit for wind turbine blade
NASA Astrophysics Data System (ADS)
Li, Xiang; Yang, Zhibo; Zhang, Han; Du, Zhaohui; Chen, Xuefeng
2015-01-01
One critical challenge to achieving reliable wind turbine blade structural health monitoring (SHM) is mainly caused by composite laminates with an anisotropy nature and a hard-to-access property. The typical pitch-catch PZTs approach generally detects structural damage with both measured and baseline signals. However, the accuracy of imaging or tomography by delay-and-sum approaches based on these signals requires improvement in practice. Via the model of Lamb wave propagation and the establishment of a dictionary that corresponds to scatters, a robust sparse reconstruction approach for structural health monitoring comes into view for its promising performance. This paper proposes a neighbor dictionary that identifies the first crack location through sparse reconstruction and then presents a growth sparse pursuit algorithm that can precisely pursue the extension of the crack. An experiment with the goal of diagnosing a composite wind turbine blade with an artificial crack is performed, and it validates the proposed approach. The results give competitively accurate crack detection with the correct locations and extension length.
Sparse matrix approximation method for an active optical control system
NASA Astrophysics Data System (ADS)
Murphy, Timothy P.; Lyon, Richard G.; Dorband, John E.; Hollis, Jan M.
2001-12-01
We develop a sparse matrix approximation method to decompose a wave front into a basis set of actuator influence functions for an active optical system consisting of a deformable mirror and a segmented primary mirror. The wave front used is constructed by Zernike polynomials to simulate the output of a phase-retrieval algorithm. Results of a Monte Carlo simulation of the optical control loop are compared with the standard, nonsparse approach in terms of accuracy and precision, as well as computational speed and memory. The sparse matrix approximation method can yield more than a 50-fold increase in the speed and a 20-fold-reduction in matrix size and a commensurate decrease in required memory, with less than 10% degradation in solution accuracy. Our method is also shown to be better than when elements are selected for the sparse matrix on a magnitude basis alone. We show that the method developed is a viable alternative to use of the full control matrix in a phase-retrieval-based active optical control system.
Sparse matrix approximation method for an active optical control system.
Murphy, T P; Lyon, R G; Dorband, J E; Hollis, J M
2001-12-10
We develop a sparse matrix approximation method to decompose a wave front into a basis set of actuator influence functions for an active optical system consisting of a deformable mirror and a segmented primary mirror. The wave front used is constructed by Zernike polynomials to simulate the output of a phase-retrieval algorithm. Results of a Monte Carlo simulation of the optical control loop are compared with the standard, nonsparse approach in terms of accuracy and precision, as well as computational speed and memory. The sparse matrix approximation method can yield more than a 50-fold increase in the speed and a 20-fold reduction in matrix size and a commensurate decrease in required memory, with less than 10% degradation in solution accuracy. Our method is also shown to be better than when elements are selected for the sparse matrix on a magnitude basis alone. We show that the method developed is a viable alternative to use of the full control matrix in a phase-retrieval-based active optical control system. PMID:18364958
Robust Reconstruction of Complex Networks from Sparse Data
NASA Astrophysics Data System (ADS)
Han, Xiao; Shen, Zhesi; Wang, Wen-Xu; Di, Zengru
2015-01-01
Reconstructing complex networks from measurable data is a fundamental problem for understanding and controlling collective dynamics of complex networked systems. However, a significant challenge arises when we attempt to decode structural information hidden in limited amounts of data accompanied by noise and in the presence of inaccessible nodes. Here, we develop a general framework for robust reconstruction of complex networks from sparse and noisy data. Specifically, we decompose the task of reconstructing the whole network into recovering local structures centered at each node. Thus, the natural sparsity of complex networks ensures a conversion from the local structure reconstruction into a sparse signal reconstruction problem that can be addressed by using the lasso, a convex optimization method. We apply our method to evolutionary games, transportation, and communication processes taking place in a variety of model and real complex networks, finding that universal high reconstruction accuracy can be achieved from sparse data in spite of noise in time series and missing data of partial nodes. Our approach opens new routes to the network reconstruction problem and has potential applications in a wide range of fields.
A comparison of methods for representing sparsely sampled random quantities.
Romero, Vicente Jose; Swiler, Laura Painton; Urbina, Angel; Mullins, Joshua
2013-09-01
This report discusses the treatment of uncertainties stemming from relatively few samples of random quantities. The importance of this topic extends beyond experimental data uncertainty to situations involving uncertainty in model calibration, validation, and prediction. With very sparse data samples it is not practical to have a goal of accurately estimating the underlying probability density function (PDF). Rather, a pragmatic goal is that the uncertainty representation should be conservative so as to bound a specified percentile range of the actual PDF, say the range between 0.025 and .975 percentiles, with reasonable reliability. A second, opposing objective is that the representation not be overly conservative; that it minimally over-estimate the desired percentile range of the actual PDF. The presence of the two opposing objectives makes the sparse-data uncertainty representation problem interesting and difficult. In this report, five uncertainty representation techniques are characterized for their performance on twenty-one test problems (over thousands of trials for each problem) according to these two opposing objectives and other performance measures. Two of the methods, statistical Tolerance Intervals and a kernel density approach specifically developed for handling sparse data, exhibit significantly better overall performance than the others.
Clutter Mitigation in Echocardiography Using Sparse Signal Separation
Turek, Javier S.; Elad, Michael; Yavneh, Irad
2015-01-01
In ultrasound imaging, clutter artifacts degrade images and may cause inaccurate diagnosis. In this paper, we apply a method called Morphological Component Analysis (MCA) for sparse signal separation with the objective of reducing such clutter artifacts. The MCA approach assumes that the two signals in the additive mix have each a sparse representation under some dictionary of atoms (a matrix), and separation is achieved by finding these sparse representations. In our work, an adaptive approach is used for learning the dictionary from the echo data. MCA is compared to Singular Value Filtering (SVF), a Principal Component Analysis- (PCA-) based filtering technique, and to a high-pass Finite Impulse Response (FIR) filter. Each filter is applied to a simulated hypoechoic lesion sequence, as well as experimental cardiac ultrasound data. MCA is demonstrated in both cases to outperform the FIR filter and obtain results comparable to the SVF method in terms of contrast-to-noise ratio (CNR). Furthermore, MCA shows a lower impact on tissue sections while removing the clutter artifacts. In experimental heart data, MCA obtains in our experiments clutter mitigation with an average CNR improvement of 1.33 dB. PMID:26199622
LESS: a model-based classifier for sparse subspaces.
Veenman, Cor J; Tax, David M J
2005-09-01
In this paper, we specifically focus on high-dimensional data sets for which the number of dimensions is an order of magnitude higher than the number of objects. From a classifier design standpoint, such small sample size problems have some interesting challenges. The first challenge is to find, from all hyperplanes that separate the classes, a separating hyperplane which generalizes well for future data. A second important task is to determine which features are required to distinguish the classes. To attack these problems, we propose the LESS (Lowest Error in a Sparse Subspace) classifier that efficiently finds linear discriminants in a sparse subspace. In contrast with most classifiers for high-dimensional data sets, the LESS classifier incorporates a (simple) data model. Further, by means of a regularization parameter, the classifier establishes a suitable trade-off between subspace sparseness and classification accuracy. In the experiments, we show how LESS performs on several high-dimensional data sets and compare its performance to related state-of-the-art classifiers like, among others, linear ridge regression with the LASSO and the Support Vector Machine. It turns out that LESS performs competitively while using fewer dimensions.
EFFECTS OF VEGETATION ON TURBULENCE, SEDIMENT TRANSPORT AND STREAM MORPHOLOGY
Neary, Vincent S
2012-01-01
Vegetation, from an individual stem to multiple stems in various configurations, profoundly alters turbulent flows. These alterations influence sediment transport and stream morphology, but depend on complex interactions and relationships between flow, plants and sediment properties. This is illustrated for three case studies that represent a variety of macrophyte patterns and scales in the environment: flows through simulated uniformly distributed plant stems, emergent and submerged; flows with alternating simulated stem patches; and flow around an isolated stem in a flood plain. The emergent case demonstrates that when density is sparse the mean velocity and turbulence intensities vary horizontally around the stems, which would promote a heterogeneous bedform morphology. However, it is still unclear how density, submergence ratio, and flow Reynolds number, in combination, influence interference effects, vortex shedding and dissipation, and velocity, pressure and lift fluctuations that affect sediment entrainment. The submerged case demonstrates significant reduction of the mean velocity, turbulence intensities, and turbulent shear near the bed compared to an unobstructed flow and supports numerous observations that vegetation promotes deposition or stabilizes bed sediments. The case of alternating emergent vegetation patches illustrates how vegetation adjusts the bed promoting scour in open water and deposition within the patches. The isolated stem case illustrates the variety of coherent structures generated, their complex interaction, and their role in specific sediment transport phenomena observed. Additional research is required, however, to quantify thresholds and relationships for flow-vegetation-sediment interactions so that aquatic macrophyte plantings can be used more effectively in water resource management.
Fast Solution in Sparse LDA for Binary Classification
NASA Technical Reports Server (NTRS)
Moghaddam, Baback
2010-01-01
An algorithm that performs sparse linear discriminant analysis (Sparse-LDA) finds near-optimal solutions in far less time than the prior art when specialized to binary classification (of 2 classes). Sparse-LDA is a type of feature- or variable- selection problem with numerous applications in statistics, machine learning, computer vision, computational finance, operations research, and bio-informatics. Because of its combinatorial nature, feature- or variable-selection problems are NP-hard or computationally intractable in cases involving more than 30 variables or features. Therefore, one typically seeks approximate solutions by means of greedy search algorithms. The prior Sparse-LDA algorithm was a greedy algorithm that considered the best variable or feature to add/ delete to/ from its subsets in order to maximally discriminate between multiple classes of data. The present algorithm is designed for the special but prevalent case of 2-class or binary classification (e.g. 1 vs. 0, functioning vs. malfunctioning, or change versus no change). The present algorithm provides near-optimal solutions on large real-world datasets having hundreds or even thousands of variables or features (e.g. selecting the fewest wavelength bands in a hyperspectral sensor to do terrain classification) and does so in typical computation times of minutes as compared to days or weeks as taken by the prior art. Sparse LDA requires solving generalized eigenvalue problems for a large number of variable subsets (represented by the submatrices of the input within-class and between-class covariance matrices). In the general (fullrank) case, the amount of computation scales at least cubically with the number of variables and thus the size of the problems that can be solved is limited accordingly. However, in binary classification, the principal eigenvalues can be found using a special analytic formula, without resorting to costly iterative techniques. The present algorithm exploits this analytic
An Efficient Scheme for Updating Sparse Cholesky Factors
NASA Technical Reports Server (NTRS)
Raghavan, Padma
2002-01-01
Raghavan had earlier developed the software package DCSPACK which can be used for solving sparse linear systems where the coefficient matrix is symmetric and positive definite (this project was not funded by NASA but by agencies such as NSF). DSCPACK-S is the serial code and DSCPACK-P is a parallel implementation suitable for multiprocessors or networks-of-workstations with message passing using MCI. The main algorithm used is the Cholesky factorization of a sparse symmetric positive positive definite matrix A = LL(T). The code can also compute the factorization A = LDL(T). The complexity of the software arises from several factors relating to the sparsity of the matrix A. A sparse N x N matrix A has typically less that cN nonzeroes where c is a small constant. If the matrix were dense, it would have O(N2) nonzeroes. The most complicated part of such sparse Cholesky factorization relates to fill-in, i.e., zeroes in the original matrix that become nonzeroes in the factor L. An efficient implementation depends to a large extent on complex data structures and on techniques from graph theory to reduce, identify, and manage fill. DSCPACK is based on an efficient multifrontal implementation with fill-managing algorithms and implementation arising from earlier research by Raghavan and others. Sparse Cholesky factorization is typically a four step process: (1) ordering to compute a fill-reducing numbering, (2) symbolic factorization to determine the nonzero structure of L, (3) numeric factorization to compute L, and, (4) triangular solution to solve L(T)x = y and Ly = b. The first two steps are symbolic and are performed using the graph of the matrix. The numeric factorization step is of dominant cost and there are several schemes for improving performance by exploiting the nested and dense structure of groups of columns in the factor. The latter are aimed at better utilization of the cache-memory hierarchy on modem processors to prevent cache-misses and provide execution
Vegetable Production System (Veggie)
NASA Technical Reports Server (NTRS)
Levine, Howard G.; Smith, Trent M.
2016-01-01
The Vegetable Production System (Veggie) was developed by Orbital Technologies Corp. to be a simple, easily stowed, and high growth volume yet low resource facility capable of producing fresh vegetables on the International Space Station (ISS). In addition to growing vegetables in space, Veggie can support a variety of experiments designed to determine how plants respond to microgravity, provide real-time psychological benefits for the crew, and conduct outreach activities. Currently, Veggie provides the largest volume available for plant growth on the ISS.
Lee, Young-Beom; Lee, Jeonghyeon; Tak, Sungho; Lee, Kangjoo; Na, Duk L; Seo, Sang Won; Jeong, Yong; Ye, Jong Chul
2016-01-15
Recent studies of functional connectivity MR imaging have revealed that the default-mode network activity is disrupted in diseases such as Alzheimer's disease (AD). However, there is not yet a consensus on the preferred method for resting-state analysis. Because the brain is reported to have complex interconnected networks according to graph theoretical analysis, the independency assumption, as in the popular independent component analysis (ICA) approach, often does not hold. Here, rather than using the independency assumption, we present a new statistical parameter mapping (SPM)-type analysis method based on a sparse graph model where temporal dynamics at each voxel position are described as a sparse combination of global brain dynamics. In particular, a new concept of a spatially adaptive design matrix has been proposed to represent local connectivity that shares the same temporal dynamics. If we further assume that local network structures within a group are similar, the estimation problem of global and local dynamics can be solved using sparse dictionary learning for the concatenated temporal data across subjects. Moreover, under the homoscedasticity variance assumption across subjects and groups that is often used in SPM analysis, the aforementioned individual and group analyses using sparse dictionary learning can be accurately modeled by a mixed-effect model, which also facilitates a standard SPM-type group-level inference using summary statistics. Using an extensive resting fMRI data set obtained from normal, mild cognitive impairment (MCI), and Alzheimer's disease patient groups, we demonstrated that the changes in the default mode network extracted by the proposed method are more closely correlated with the progression of Alzheimer's disease.
Lee, Young-Beom; Lee, Jeonghyeon; Tak, Sungho; Lee, Kangjoo; Na, Duk L; Seo, Sang Won; Jeong, Yong; Ye, Jong Chul
2016-01-15
Recent studies of functional connectivity MR imaging have revealed that the default-mode network activity is disrupted in diseases such as Alzheimer's disease (AD). However, there is not yet a consensus on the preferred method for resting-state analysis. Because the brain is reported to have complex interconnected networks according to graph theoretical analysis, the independency assumption, as in the popular independent component analysis (ICA) approach, often does not hold. Here, rather than using the independency assumption, we present a new statistical parameter mapping (SPM)-type analysis method based on a sparse graph model where temporal dynamics at each voxel position are described as a sparse combination of global brain dynamics. In particular, a new concept of a spatially adaptive design matrix has been proposed to represent local connectivity that shares the same temporal dynamics. If we further assume that local network structures within a group are similar, the estimation problem of global and local dynamics can be solved using sparse dictionary learning for the concatenated temporal data across subjects. Moreover, under the homoscedasticity variance assumption across subjects and groups that is often used in SPM analysis, the aforementioned individual and group analyses using sparse dictionary learning can be accurately modeled by a mixed-effect model, which also facilitates a standard SPM-type group-level inference using summary statistics. Using an extensive resting fMRI data set obtained from normal, mild cognitive impairment (MCI), and Alzheimer's disease patient groups, we demonstrated that the changes in the default mode network extracted by the proposed method are more closely correlated with the progression of Alzheimer's disease. PMID:26524138
Pothen, A.
1999-02-01
Gary Kumfert and Alex Pothen have improved the quality and run time of two ordering algorithms for minimizing the wavefront and envelope size of sparse matrices and graphs. These algorithms compute orderings for irregular data structures (e.g., unstructured meshes) that reduce the number of cache misses on modern workstation architectures. They have completed the implementation of a parallel solver for sparse, symmetric indefinite systems for distributed memory computers such as the IBM SP-2. The indefiniteness requires one to incorporate block pivoting (2 by 2 blocks) in the algorithm, thus demanding dynamic, parallel data structures. This is the first reported parallel solver for the indefinite problem. Direct methods for solving systems of linear equations employ sophisticated combinatorial and algebraic algorithms that contribute to software complexity, and hence it is natural to consider object-oriented design (OOD) in this context. The authors have continued to create software for solving sparse systems of linear equations by direct methods employing OOD. Fast computation of robust preconditioners is a priority for solving large systems of equations on unstructured grids and in other applications. They have developed new algorithms and software that can compute incomplete factorization preconditioners for high level fill in time proportional to the number of floating point operations and memory accesses.
NASA Technical Reports Server (NTRS)
vanLeeuwen, W. J. D.; Huete, A. R.; Duncan, J.; Franklin, J.
1994-01-01
A shrub savannah landscape in Niger was optically characterized utilizing blue, green, red and near-infrared wavelengths. Selected vegetation indices were evaluated for their performance and sensitivity to describe the complex Sahelian soil/vegetation canopies. Bidirectional reflectance factors (BRF) of plants and soils were measured at several view angles, and used as input to various vegetation indices. Both soil and vegetation targets had strong anisotropic reflectance properties, rendering all vegetation index (6) responses to be a direct function of sun and view geometry. Soil background influences were shown to alter the response of most vegetation indices. N-space greenness had the smallest dynamic range in VI response, but the n-space brightness index provided additional useful information. The global environmental monitoring index (GEMI) showed a large 6 dynamic range for bare soils, which was undesirable for a vegetation index. The view angle response of the normalized difference vegetation index (NDVI), atmosphere resistant vegetation index (ARVI) and soil atmosphere resistant vegetation index (SARVI) were asymmetric about nadir for multiple view angles, and were, except for the SARVI, altered seriously by soil moisture and/or soil brightness effects. The soil adjusted vegetation index (SAVI) was least affected by surface soil moisture and was symmetric about nadir for grass vegetation covers. Overall the SAVI, SARVI and the n-space vegetation index performed best under all adverse conditions and were recommended to monitor vegetation growth in the sparsely vegetated Sahelian zone.
Global Enhanced Vegetation Index
NASA Technical Reports Server (NTRS)
2002-01-01
By carefully measuring the wavelengths and intensity of visible and near-infrared light reflected by the land surface back up into space, the Moderate-resolution Imaging Spectroradiometer (MODIS) Team can quantify the concentrations of green leaf vegetation around the world. The above MODIS Enhanced Vegetation Index (EVI) map shows the density of plant growth over the entire globe. Very low values of EVI (white and brown areas) correspond to barren areas of rock, sand, or snow. Moderate values (light greens) represent shrub and grassland, while high values indicate temperate and tropical rainforests (dark greens). The MODIS EVI gives scientists a new tool for monitoring major fluctuations in vegetation and understanding how they affect, and are affected by, regional climate trends. For more information, read NASA Unveils Spectacular Suite of New Global Data Products from MODIS. Image courtesy MODIS Land Group/Vegetation Indices, Alfredo Huete, Principal Investigator, and Kamel Didan, University of Arizona
Not Available
1982-01-01
Fifty contributions (presentations) involving more than one hundred people worldwide were given at the International Conference on Plant and Vegetable Oils as Fuels. The proceedings were in Fargo, North Dakota, from August 2-4, 1982. The conference helped to promote renewable fuels, bio-oils, from plant and vegetable oils. Separate abstracts were prepared for 44 items for inclusion in the Energy Data Base.
Burke, M.K.; King, S.L.; Eisenbies, M.H.; Gartner, D.
2000-01-01
Intro paragraph: Characterization of bottomland hardwood vegetation in relatively undisturbed forests can provide critical information for developing effective wetland creation and restoration techniques and for assessing the impacts of management and development. Classification is a useful technique in characterizing vegetation because it summarizes complex data sets, assists in hypothesis generation about factors influencing community variation, and helps refine models of community structure. Hierarchical classification of communities is particularly useful for showing relationships among samples (Gauche 1982).
NASA Technical Reports Server (NTRS)
Macdonald, R. B.; Houston, A. G.; Heydorn, R. P.; Botkin, D. B.; Estes, J. E.; Strahler, A. H.
1981-01-01
An attempt is made to identify the need for, and the current capability of, a technology which could aid in monitoring the Earth's vegetation resource on a global scale. Vegetation is one of our most critical natural resources, and accurate timely information on its current status and temporal dynamics is essential to understand many basic and applied environmental interrelationships which exist on the small but complex planet Earth.
Moody, Daniela I.; Brumby, Steven P.; Rowland, Joel C.; Altmann, Garrett L.
2014-12-09
We present results from an ongoing effort to extend neuromimetic machine vision algorithms to multispectral data using adaptive signal processing combined with compressive sensing and machine learning techniques. Our goal is to develop a robust classification methodology that will allow for automated discretization of the landscape into distinct units based on attributes such as vegetation, surface hydrological properties, and topographic/geomorphic characteristics. We use a Hebbian learning rule to build spectral-textural dictionaries that are tailored for classification. We learn our dictionaries from millions of overlapping multispectral image patches and then use a pursuit search to generate classification features. Land cover labelsmore » are automatically generated using unsupervised clustering of sparse approximations (CoSA). We demonstrate our method on multispectral WorldView-2 data from a coastal plain ecosystem in Barrow, Alaska. We explore learning from both raw multispectral imagery and normalized band difference indices. We explore a quantitative metric to evaluate the spectral properties of the clusters in order to potentially aid in assigning land cover categories to the cluster labels. In this study, our results suggest CoSA is a promising approach to unsupervised land cover classification in high-resolution satellite imagery.« less
Moody, Daniela I.; Brumby, Steven P.; Rowland, Joel C.; Altmann, Garrett L.
2014-12-09
We present results from an ongoing effort to extend neuromimetic machine vision algorithms to multispectral data using adaptive signal processing combined with compressive sensing and machine learning techniques. Our goal is to develop a robust classification methodology that will allow for automated discretization of the landscape into distinct units based on attributes such as vegetation, surface hydrological properties, and topographic/geomorphic characteristics. We use a Hebbian learning rule to build spectral-textural dictionaries that are tailored for classification. We learn our dictionaries from millions of overlapping multispectral image patches and then use a pursuit search to generate classification features. Land cover labels are automatically generated using unsupervised clustering of sparse approximations (CoSA). We demonstrate our method on multispectral WorldView-2 data from a coastal plain ecosystem in Barrow, Alaska. We explore learning from both raw multispectral imagery and normalized band difference indices. We explore a quantitative metric to evaluate the spectral properties of the clusters in order to potentially aid in assigning land cover categories to the cluster labels. In this study, our results suggest CoSA is a promising approach to unsupervised land cover classification in high-resolution satellite imagery.
Bessler, T.R.
1986-05-13
A process is described for preparing an injectable vegetable oil selected from the group consisting of soybean oil and sunflower oil and mixtures thereof which comprise: (a) first treating the vegetable oil at a temperature of 80/sup 0/C to about 130/sup 0/C with an acid clay; (b) deodorizing the vegetable oil with steam at a temperature of 220/sup 0/C to about 280/sup 0/C and applying a vacuum to remove volatilized components; (c) treating the deodorized vegetable oil, at a temperature of from about 10/sup 0/C to about 60/sup 0/C, with an acid clay to reduce the content of a member selected from the group consisting of diglycerides, tocopherol components, and trilinolenin and mixtures thereof, wherein the acid clay is added in a weight ratio to the deoderized vegetable oil of from about 1:99 to about 1:1; and (d) thereafter conducting a particulate filtration to remove a substantial portion of the acid clay from the vegetable oil, wherein the filtration is accomplished with filters having a pore size of from about 0.1 to 0.45 microns, thereby obtaining the injectable oil.
Role of vegetation in modulating denudation and topography across the Himalaya
NASA Astrophysics Data System (ADS)
Olen, Stephanie; Bookhagen, Bodo; Strecker, Manfred
2015-04-01
Studies of Himalayan denudation, to date, have primarily focused on the effects of lithology, tectonic activity, and climate in shaping landscape and controlling denudation rates. Climate can impact denudation not only through increased precipitation, runoff, or glaciation, but also via its role in controlling vegetation cover. Since the classical study of Langbein and Schumm [1958] emphasizing the role of vegetation cover in determining erosional efficiency, theoretical and plot-scale studies have highlighted the role of vegetation on surface processes [Collins et al., 2004; Istanbulluoglu and Bras, 2005; Collins and Bras, 2010; Carretier et al., 2013; Jeffery et al., 2014]. Vegetation cover and density vary considerably in the Himalaya, both across and along strike. Across strike, vegetation transitions from dense forest and agriculturally-used plots in the Lesser Himalaya to sparse alpine and arid, virtually non-vegetated regions at high elevation and in the rain shadow north of the Higher Himalaya peaks. Along-strike vegetation densities also differ significantly and show a pronounced E-W gradient. To quantify the along-strike vegetation gradient, we use 14 years of MODIS 13C1 enhanced vegetation index (EVI) data to calculate mean annual, summer (MJJASO), and winter (NDJFMA) for the entire Himalaya. Additionally, we calculate a differential EVI that compares summer versus winter vegetation density (MJJASO/NDJFMA). A decrease in vegetation density is observed from east to west, with the greatest difference in winter vegetation cover (225% higher in the eastern than western Himalaya). In contrast, differential EVI is higher in the western Himalaya, increasing 170% from east to west. To evaluate the effect of vegetation on denudation and landscape evolution, we combine the 14-year EVI data, topographic analysis, and a compilation of >100 published and unpublished 10-Be terrestrial cosmogenic nuclide (TCN) catchment-mean denudation rates from across the Himalaya
Sparse and compositionally robust inference of microbial ecological networks.
Kurtz, Zachary D; Müller, Christian L; Miraldi, Emily R; Littman, Dan R; Blaser, Martin J; Bonneau, Richard A
2015-05-01
16S ribosomal RNA (rRNA) gene and other environmental sequencing techniques provide snapshots of microbial communities, revealing phylogeny and the abundances of microbial populations across diverse ecosystems. While changes in microbial community structure are demonstrably associated with certain environmental conditions (from metabolic and immunological health in mammals to ecological stability in soils and oceans), identification of underlying mechanisms requires new statistical tools, as these datasets present several technical challenges. First, the abundances of microbial operational taxonomic units (OTUs) from amplicon-based datasets are compositional. Counts are normalized to the total number of counts in the sample. Thus, microbial abundances are not independent, and traditional statistical metrics (e.g., correlation) for the detection of OTU-OTU relationships can lead to spurious results. Secondly, microbial sequencing-based studies typically measure hundreds of OTUs on only tens to hundreds of samples; thus, inference of OTU-OTU association networks is severely under-powered, and additional information (or assumptions) are required for accurate inference. Here, we present SPIEC-EASI (SParse InversE Covariance Estimation for Ecological Association Inference), a statistical method for the inference of microbial ecological networks from amplicon sequencing datasets that addresses both of these issues. SPIEC-EASI combines data transformations developed for compositional data analysis with a graphical model inference framework that assumes the underlying ecological association network is sparse. To reconstruct the network, SPIEC-EASI relies on algorithms for sparse neighborhood and inverse covariance selection. To provide a synthetic benchmark in the absence of an experimentally validated gold-standard network, SPIEC-EASI is accompanied by a set of computational tools to generate OTU count data from a set of diverse underlying network topologies. SPIEC
Robust image analysis with sparse representation on quantized visual features.
Bao, Bing-Kun; Zhu, Guangyu; Shen, Jialie; Yan, Shuicheng
2013-03-01
Recent techniques based on sparse representation (SR) have demonstrated promising performance in high-level visual recognition, exemplified by the highly accurate face recognition under occlusion and other sparse corruptions. Most research in this area has focused on classification algorithms using raw image pixels, and very few have been proposed to utilize the quantized visual features, such as the popular bag-of-words feature abstraction. In such cases, besides the inherent quantization errors, ambiguity associated with visual word assignment and misdetection of feature points, due to factors such as visual occlusions and noises, constitutes the major cause of dense corruptions of the quantized representation. The dense corruptions can jeopardize the decision process by distorting the patterns of the sparse reconstruction coefficients. In this paper, we aim to eliminate the corruptions and achieve robust image analysis with SR. Toward this goal, we introduce two transfer processes (ambiguity transfer and mis-detection transfer) to account for the two major sources of corruption as discussed. By reasonably assuming the rarity of the two kinds of distortion processes, we augment the original SR-based reconstruction objective with l(0) norm regularization on the transfer terms to encourage sparsity and, hence, discourage dense distortion/transfer. Computationally, we relax the nonconvex l(0) norm optimization into a convex l(1) norm optimization problem, and employ the accelerated proximal gradient method to optimize the convergence provable updating procedure. Extensive experiments on four benchmark datasets, Caltech-101, Caltech-256, Corel-5k, and CMU pose, illumination, and expression, manifest the necessity of removing the quantization corruptions and the various advantages of the proposed framework.
Classification of transient signals using sparse representations over adaptive dictionaries
NASA Astrophysics Data System (ADS)
Moody, Daniela I.; Brumby, Steven P.; Myers, Kary L.; Pawley, Norma H.
2011-06-01
Automatic classification of broadband transient radio frequency (RF) signals is of particular interest in persistent surveillance applications. Because such transients are often acquired in noisy, cluttered environments, and are characterized by complex or unknown analytical models, feature extraction and classification can be difficult. We propose a fast, adaptive classification approach based on non-analytical dictionaries learned from data. Conventional representations using fixed (or analytical) orthogonal dictionaries, e.g., Short Time Fourier and Wavelet Transforms, can be suboptimal for classification of transients, as they provide a rigid tiling of the time-frequency space, and are not specifically designed for a particular signal class. They do not usually lead to sparse decompositions, and require separate feature selection algorithms, creating additional computational overhead. Pursuit-type decompositions over analytical, redundant dictionaries yield sparse representations by design, and work well for target signals in the same function class as the dictionary atoms. The pursuit search however has a high computational cost, and the method can perform poorly in the presence of realistic noise and clutter. Our approach builds on the image analysis work of Mairal et al. (2008) to learn a discriminative dictionary for RF transients directly from data without relying on analytical constraints or additional knowledge about the signal characteristics. We then use a pursuit search over this dictionary to generate sparse classification features. We demonstrate that our learned dictionary is robust to unexpected changes in background content and noise levels. The target classification decision is obtained in almost real-time via a parallel, vectorized implementation.
Social biases determine spatiotemporal sparseness of ciliate mating heuristics.
Clark, Kevin B
2012-01-01
Ciliates become highly social, even displaying animal-like qualities, in the joint presence of aroused conspecifics and nonself mating pheromones. Pheromone detection putatively helps trigger instinctual and learned courtship and dominance displays from which social judgments are made about the availability, compatibility, and fitness representativeness or likelihood of prospective mates and rivals. In earlier studies, I demonstrated the heterotrich Spirostomum ambiguum improves mating competence by effecting preconjugal strategies and inferences in mock social trials via behavioral heuristics built from Hebbian-like associative learning. Heuristics embody serial patterns of socially relevant action that evolve into ordered, topologically invariant computational networks supporting intra- and intermate selection. S. ambiguum employs heuristics to acquire, store, plan, compare, modify, select, and execute sets of mating propaganda. One major adaptive constraint over formation and use of heuristics involves a ciliate's initial subjective bias, responsiveness, or preparedness, as defined by Stevens' Law of subjective stimulus intensity, for perceiving the meaningfulness of mechanical pressures accompanying cell-cell contacts and additional perimating events. This bias controls durations and valences of nonassociative learning, search rates for appropriate mating strategies, potential net reproductive payoffs, levels of social honesty and deception, successful error diagnosis and correction of mating signals, use of insight or analysis to solve mating dilemmas, bioenergetics expenditures, and governance of mating decisions by classical or quantum statistical mechanics. I now report this same social bias also differentially affects the spatiotemporal sparseness, as measured with metric entropy, of ciliate heuristics. Sparseness plays an important role in neural systems through optimizing the specificity, efficiency, and capacity of memory representations. The present
Social biases determine spatiotemporal sparseness of ciliate mating heuristics.
Clark, Kevin B
2012-01-01
Ciliates become highly social, even displaying animal-like qualities, in the joint presence of aroused conspecifics and nonself mating pheromones. Pheromone detection putatively helps trigger instinctual and learned courtship and dominance displays from which social judgments are made about the availability, compatibility, and fitness representativeness or likelihood of prospective mates and rivals. In earlier studies, I demonstrated the heterotrich Spirostomum ambiguum improves mating competence by effecting preconjugal strategies and inferences in mock social trials via behavioral heuristics built from Hebbian-like associative learning. Heuristics embody serial patterns of socially relevant action that evolve into ordered, topologically invariant computational networks supporting intra- and intermate selection. S. ambiguum employs heuristics to acquire, store, plan, compare, modify, select, and execute sets of mating propaganda. One major adaptive constraint over formation and use of heuristics involves a ciliate's initial subjective bias, responsiveness, or preparedness, as defined by Stevens' Law of subjective stimulus intensity, for perceiving the meaningfulness of mechanical pressures accompanying cell-cell contacts and additional perimating events. This bias controls durations and valences of nonassociative learning, search rates for appropriate mating strategies, potential net reproductive payoffs, levels of social honesty and deception, successful error diagnosis and correction of mating signals, use of insight or analysis to solve mating dilemmas, bioenergetics expenditures, and governance of mating decisions by classical or quantum statistical mechanics. I now report this same social bias also differentially affects the spatiotemporal sparseness, as measured with metric entropy, of ciliate heuristics. Sparseness plays an important role in neural systems through optimizing the specificity, efficiency, and capacity of memory representations. The present
A Comparison of Methods for Ocean Reconstruction from Sparse Observations
NASA Astrophysics Data System (ADS)
Streletz, G. J.; Kronenberger, M.; Weber, C.; Gebbie, G.; Hagen, H.; Garth, C.; Hamann, B.; Kreylos, O.; Kellogg, L. H.; Spero, H. J.
2014-12-01
We present a comparison of two methods for developing reconstructions of oceanic scalar property fields from sparse scattered observations. Observed data from deep sea core samples provide valuable information regarding the properties of oceans in the past. However, because the locations of sample sites are distributed on the ocean floor in a sparse and irregular manner, developing a global ocean reconstruction is a difficult task. Our methods include a flow-based and a moving least squares -based approximation method. The flow-based method augments the process of interpolating or approximating scattered scalar data by incorporating known flow information. The scheme exploits this additional knowledge to define a non-Euclidean distance measure between points in the spatial domain. This distance measure is used to create a reconstruction of the desired scalar field on the spatial domain. The resulting reconstruction thus incorporates information from both the scattered samples and the known flow field. The second method does not assume a known flow field, but rather works solely with the observed scattered samples. It is based on a modification of the moving least squares approach, a weighted least squares approximation method that blends local approximations into a global result. The modifications target the selection of data used for these local approximations and the construction of the weighting function. The definition of distance used in the weighting function is crucial for this method, so we use a machine learning approach to determine a set of near-optimal parameters for the weighting. We have implemented both of the reconstruction methods and have tested them using several sparse oceanographic datasets. Based upon these studies, we discuss the advantages and disadvantages of each method and suggest possible ways to combine aspects of both methods in order to achieve an overall high-quality reconstruction.
Analysing Local Sparseness in the Macaque Brain Network
Singh, Raghavendra; Nagar, Seema; Nanavati, Amit A.
2015-01-01
Understanding the network structure of long distance pathways in the brain is a necessary step towards developing an insight into the brain’s function, organization and evolution. Dense global subnetworks of these pathways have often been studied, primarily due to their functional implications. Instead we study sparse local subnetworks of the pathways to establish the role of a brain area in enabling shortest path communication between its non-adjacent topological neighbours. We propose a novel metric to measure the topological communication load on a vertex due to its immediate neighbourhood, and show that in terms of distribution of this local communication load, a network of Macaque long distance pathways is substantially different from other real world networks and random graph models. Macaque network contains the entire range of local subnetworks, from star-like networks to clique-like networks, while other networks tend to contain a relatively small range of subnetworks. Further, sparse local subnetworks in the Macaque network are not only found across topographical super-areas, e.g., lobes, but also within a super-area, arguing that there is conservation of even relatively short-distance pathways. To establish the communication role of a vertex we borrow the concept of brokerage from social science, and present the different types of brokerage roles that brain areas play, highlighting that not only the thalamus, but also cingulate gyrus and insula often act as “relays” for areas in the neocortex. These and other analysis of communication load and roles of the sparse subnetworks of the Macaque brain provide new insights into the organisation of its pathways. PMID:26437077
A sparse Gaussian process framework for photometric redshift estimation
NASA Astrophysics Data System (ADS)
Almosallam, Ibrahim A.; Lindsay, Sam N.; Jarvis, Matt J.; Roberts, Stephen J.
2016-01-01
Accurate photometric redshifts are a lynchpin for many future experiments to pin down the cosmological model and for studies of galaxy evolution. In this study, a novel sparse regression framework for photometric redshift estimation is presented. Synthetic data set simulating the Euclid survey and real data from SDSS DR12 are used to train and test the proposed models. We show that approaches which include careful data preparation and model design offer a significant improvement in comparison with several competing machine learning algorithms. Standard implementations of most regression algorithms use the minimization of the sum of squared errors as the objective function. For redshift inference, this induces a bias in the posterior mean of the output distribution, which can be problematic. In this paper, we directly minimize the target metric Δz = (zs - zp)/(1 + zs) and address the bias problem via a distribution-based weighting scheme, incorporated as part of the optimization objective. The results are compared with other machine learning algorithms in the field such as artificial neural networks (ANN), Gaussian processes (GPs) and sparse GPs. The proposed framework reaches a mean absolute Δz = 0.0026(1 + zs), over the redshift range of 0 ≤ zs ≤ 2 on the simulated data, and Δz = 0.0178(1 + zs) over the entire redshift range on the SDSS DR12 survey, outperforming the standard ANNz used in the literature. We also investigate how the relative size of the training sample affects the photometric redshift accuracy. We find that a training sample of >30 per cent of total sample size, provides little additional constraint on the photometric redshifts, and note that our GP formalism strongly outperforms ANNz in the sparse data regime for the simulated data set.
Sparse and Compositionally Robust Inference of Microbial Ecological Networks
Kurtz, Zachary D.; Müller, Christian L.; Miraldi, Emily R.; Littman, Dan R.; Blaser, Martin J.; Bonneau, Richard A.
2015-01-01
16S ribosomal RNA (rRNA) gene and other environmental sequencing techniques provide snapshots of microbial communities, revealing phylogeny and the abundances of microbial populations across diverse ecosystems. While changes in microbial community structure are demonstrably associated with certain environmental conditions (from metabolic and immunological health in mammals to ecological stability in soils and oceans), identification of underlying mechanisms requires new statistical tools, as these datasets present several technical challenges. First, the abundances of microbial operational taxonomic units (OTUs) from amplicon-based datasets are compositional. Counts are normalized to the total number of counts in the sample. Thus, microbial abundances are not independent, and traditional statistical metrics (e.g., correlation) for the detection of OTU-OTU relationships can lead to spurious results. Secondly, microbial sequencing-based studies typically measure hundreds of OTUs on only tens to hundreds of samples; thus, inference of OTU-OTU association networks is severely under-powered, and additional information (or assumptions) are required for accurate inference. Here, we present SPIEC-EASI (SParse InversE Covariance Estimation for Ecological Association Inference), a statistical method for the inference of microbial ecological networks from amplicon sequencing datasets that addresses both of these issues. SPIEC-EASI combines data transformations developed for compositional data analysis with a graphical model inference framework that assumes the underlying ecological association network is sparse. To reconstruct the network, SPIEC-EASI relies on algorithms for sparse neighborhood and inverse covariance selection. To provide a synthetic benchmark in the absence of an experimentally validated gold-standard network, SPIEC-EASI is accompanied by a set of computational tools to generate OTU count data from a set of diverse underlying network topologies. SPIEC
Analysing Local Sparseness in the Macaque Brain Network.
Singh, Raghavendra; Nagar, Seema; Nanavati, Amit A
2015-01-01
Understanding the network structure of long distance pathways in the brain is a necessary step towards developing an insight into the brain's function, organization and evolution. Dense global subnetworks of these pathways have often been studied, primarily due to their functional implications. Instead we study sparse local subnetworks of the pathways to establish the role of a brain area in enabling shortest path communication between its non-adjacent topological neighbours. We propose a novel metric to measure the topological communication load on a vertex due to its immediate neighbourhood, and show that in terms of distribution of this local communication load, a network of Macaque long distance pathways is substantially different from other real world networks and random graph models. Macaque network contains the entire range of local subnetworks, from star-like networks to clique-like networks, while other networks tend to contain a relatively small range of subnetworks. Further, sparse local subnetworks in the Macaque network are not only found across topographical super-areas, e.g., lobes, but also within a super-area, arguing that there is conservation of even relatively short-distance pathways. To establish the communication role of a vertex we borrow the concept of brokerage from social science, and present the different types of brokerage roles that brain areas play, highlighting that not only the thalamus, but also cingulate gyrus and insula often act as "relays" for areas in the neocortex. These and other analysis of communication load and roles of the sparse subnetworks of the Macaque brain provide new insights into the organisation of its pathways.
Symbolic Givens reduction in large sparse least squares problems
Ostrouchov, G.
1984-12-01
Orthogonal Givens factorization is a popular method for solving large sparse least squares problems. In order to exploit sparsity and to use a fixed data structure in Givens reduction, a preliminary symbolic factorization needs to be performed. Some results on row-ordering and structure of rows in a partially reduced matrix are obtained using a graph-theoretic representation. These results provide a basis for a symbolic Givens factorization. Column-ordering is also discussed, and an algorithm for symbolic Givens reduction is developed and tested.
Solution of sparse positive definite systems on a hypercube
George, A.; Heath, M.; Liu, J.; Ng, E.
1988-10-01
The solution of large sparse positive definite systems of equations typically involves four steps: ordering, data structure set-up (symbolic factorization), numerical factorization, and triangular solution. This article describes how these four phases are implemented on a hypercube multiprocessor. The role of elimination trees in the exploitation of sparsity and the identification of parallelism is explained, and pseudo-code algorithms are provided for some of the important algorithms. Numerical experiments run on an Intel iPSC multiprocessor are presented in order to provide some indication of the performance of the various algorithms. 36 refs., 23 figs., 5 tabs.
Dynamic Stochastic Superresolution of sparsely observed turbulent systems
Branicki, M.; Majda, A.J.
2013-05-15
Real-time capture of the relevant features of the unresolved turbulent dynamics of complex natural systems from sparse noisy observations and imperfect models is a notoriously difficult problem. The resulting lack of observational resolution and statistical accuracy in estimating the important turbulent processes, which intermittently send significant energy to the large-scale fluctuations, hinders efficient parameterization and real-time prediction using discretized PDE models. This issue is particularly subtle and important when dealing with turbulent geophysical systems with an vast range of interacting spatio-temporal scales and rough energy spectra near the mesh scale of numerical models. Here, we introduce and study a suite of general Dynamic Stochastic Superresolution (DSS) algorithms and show that, by appropriately filtering sparse regular observations with the help of cheap stochastic exactly solvable models, one can derive stochastically ‘superresolved’ velocity fields and gain insight into the important characteristics of the unresolved dynamics, including the detection of the so-called black swans. The DSS algorithms operate in Fourier domain and exploit the fact that the coarse observation network aliases high-wavenumber information into the resolved waveband. It is shown that these cheap algorithms are robust and have significant skill on a test bed of turbulent solutions from realistic nonlinear turbulent spatially extended systems in the presence of a significant model error. In particular, the DSS algorithms are capable of successfully capturing time-localized extreme events in the unresolved modes, and they provide good and robust skill for recovery of the unresolved processes in terms of pattern correlation. Moreover, we show that DSS improves the skill for recovering the primary modes associated with the sparse observation mesh which is equally important in applications. The skill of the various DSS algorithms depends on the energy spectrum
Universal Collaboration Strategies for Signal Detection: A Sparse Learning Approach
NASA Astrophysics Data System (ADS)
Khanduri, Prashant; Kailkhura, Bhavya; Thiagarajan, Jayaraman J.; Varshney, Pramod K.
2016-10-01
This paper considers the problem of high dimensional signal detection in a large distributed network whose nodes can collaborate with their one-hop neighboring nodes (spatial collaboration). We assume that only a small subset of nodes communicate with the Fusion Center (FC). We design optimal collaboration strategies which are universal for a class of deterministic signals. By establishing the equivalence between the collaboration strategy design problem and sparse PCA, we solve the problem efficiently and evaluate the impact of collaboration on detection performance.
An image restoration method based on sparse constraint
NASA Astrophysics Data System (ADS)
Qiang, Zhenping; Liu, Hui; Chen, Xu; Shang, Zhenhong; Zeng, Lingjun
2013-07-01
In this paper, proposed an image restoration method which base on the sparse constraint. Based on the principle of Compressed Sensing, the observed image is transformed into the wavelet domain, and then converted the image restoration problem to a convex set unrestricted optimization problem by limiting the number of non-zero elements of the wavelet domain, using the gradient projection method for solving the optimization problem to achieve the restoration of the input image. Experiments show that the method presented has the fast convergence and good robustness compared to the traditional total variation regularization restoration method.
Dynamic Stochastic Superresolution of sparsely observed turbulent systems
NASA Astrophysics Data System (ADS)
Branicki, M.; Majda, A. J.
2013-05-01
Real-time capture of the relevant features of the unresolved turbulent dynamics of complex natural systems from sparse noisy observations and imperfect models is a notoriously difficult problem. The resulting lack of observational resolution and statistical accuracy in estimating the important turbulent processes, which intermittently send significant energy to the large-scale fluctuations, hinders efficient parameterization and real-time prediction using discretized PDE models. This issue is particularly subtle and important when dealing with turbulent geophysical systems with an vast range of interacting spatio-temporal scales and rough energy spectra near the mesh scale of numerical models. Here, we introduce and study a suite of general Dynamic Stochastic Superresolution (DSS) algorithms and show that, by appropriately filtering sparse regular observations with the help of cheap stochastic exactly solvable models, one can derive stochastically 'superresolved' velocity fields and gain insight into the important characteristics of the unresolved dynamics, including the detection of the so-called black swans. The DSS algorithms operate in Fourier domain and exploit the fact that the coarse observation network aliases high-wavenumber information into the resolved waveband. It is shown that these cheap algorithms are robust and have significant skill on a test bed of turbulent solutions from realistic nonlinear turbulent spatially extended systems in the presence of a significant model error. In particular, the DSS algorithms are capable of successfully capturing time-localized extreme events in the unresolved modes, and they provide good and robust skill for recovery of the unresolved processes in terms of pattern correlation. Moreover, we show that DSS improves the skill for recovering the primary modes associated with the sparse observation mesh which is equally important in applications. The skill of the various DSS algorithms depends on the energy spectrum of
An information theoretic approach of designing sparse kernel adaptive filters.
Liu, Weifeng; Park, Il; Principe, José C
2009-12-01
This paper discusses an information theoretic approach of designing sparse kernel adaptive filters. To determine useful data to be learned and remove redundant ones, a subjective information measure called surprise is introduced. Surprise captures the amount of information a datum contains which is transferable to a learning system. Based on this concept, we propose a systematic sparsification scheme, which can drastically reduce the time and space complexity without harming the performance of kernel adaptive filters. Nonlinear regression, short term chaotic time-series prediction, and long term time-series forecasting examples are presented. PMID:19923047
Kanerva's sparse distributed memory with multiple hamming thresholds
NASA Technical Reports Server (NTRS)
Pohja, Seppo; Kaski, Kimmo
1992-01-01
If the stored input patterns of Kanerva's Sparse Distributed Memory (SDM) are highly correlated, utilization of the storage capacity is very low compared to the case of uniformly distributed random input patterns. We consider a variation of SDM that has a better storage capacity utilization for correlated input patterns. This approach uses a separate selection threshold for each physical storage address or hard location. The selection of the hard locations for reading or writing can be done in parallel of which SDM implementations can benefit.
Parallel solution of sparse one-dimensional dynamic programming problems
NASA Technical Reports Server (NTRS)
Nicol, David M.
1989-01-01
Parallel computation offers the potential for quickly solving large computational problems. However, it is often a non-trivial task to effectively use parallel computers. Solution methods must sometimes be reformulated to exploit parallelism; the reformulations are often more complex than their slower serial counterparts. We illustrate these points by studying the parallelization of sparse one-dimensional dynamic programming problems, those which do not obviously admit substantial parallelization. We propose a new method for parallelizing such problems, develop analytic models which help us to identify problems which parallelize well, and compare the performance of our algorithm with existing algorithms on a multiprocessor.
An information theoretic approach of designing sparse kernel adaptive filters.
Liu, Weifeng; Park, Il; Principe, José C
2009-12-01
This paper discusses an information theoretic approach of designing sparse kernel adaptive filters. To determine useful data to be learned and remove redundant ones, a subjective information measure called surprise is introduced. Surprise captures the amount of information a datum contains which is transferable to a learning system. Based on this concept, we propose a systematic sparsification scheme, which can drastically reduce the time and space complexity without harming the performance of kernel adaptive filters. Nonlinear regression, short term chaotic time-series prediction, and long term time-series forecasting examples are presented.
Object-Oriented Design for Sparse Direct Solvers
NASA Technical Reports Server (NTRS)
Dobrian, Florin; Kumfert, Gary; Pothen, Alex
1999-01-01
We discuss the object-oriented design of a software package for solving sparse, symmetric systems of equations (positive definite and indefinite) by direct methods. At the highest layers, we decouple data structure classes from algorithmic classes for flexibility. We describe the important structural and algorithmic classes in our design, and discuss the trade-offs we made for high performance. The kernels at the lower layers were optimized by hand. Our results show no performance loss from our object-oriented design, while providing flexibility, case of use, and extensibility over solvers using procedural design.
Multi-frame blind deconvolution using sparse priors
NASA Astrophysics Data System (ADS)
Dong, Wende; Feng, Huajun; Xu, Zhihai; Li, Qi
2012-05-01
In this paper, we propose a method for multi-frame blind deconvolution. Two sparse priors, i.e., the natural image gradient prior and an l1-norm based prior are used to regularize the latent image and point spread functions (PSFs) respectively. An alternating minimization approach is adopted to solve the resulted optimization problem. We use both gray scale blurred frames from a data set and some colored ones which are captured by a digital camera to verify the robustness of our approach. Experimental results show that the proposed method can accurately reconstruct PSFs with complex structures and the restored images are of high quality.
A LANDSAT study of ephemeral and perennial rangeland vegetation and soils
NASA Technical Reports Server (NTRS)
Bentley, R. G., Jr. (Principal Investigator); Salmon-Drexler, B. C.; Bonner, W. J.; Vincent, R. K.
1976-01-01
The author has identified the following significant results. Several methods of computer processing were applied to LANDSAT data for mapping vegetation characteristics of perennial rangeland in Montana and ephemeral rangeland in Arizona. The choice of optimal processing technique was dependent on prescribed mapping and site condition. Single channel level slicing and ratioing of channels were used for simple enhancement. Predictive models for mapping percent vegetation cover based on data from field spectra and LANDSAT data were generated by multiple linear regression of six unique LANDSAT spectral ratios. Ratio gating logic and maximum likelihood classification were applied successfully to recognize plant communities in Montana. Maximum likelihood classification did little to improve recognition of terrain features when compared to a single channel density slice in sparsely vegetated Arizona. LANDSAT was found to be more sensitive to differences between plant communities based on percentages of vigorous vegetation than to actual physical or spectral differences among plant species.
Effect of vegetation on soil moisture sensing observed from orbiting microwave radiometers
NASA Technical Reports Server (NTRS)
Wang, J. R.
1985-01-01
The microwave radiometric measurements made by the Skylab 1.4 GHz radiometer and by the 6.6 GHz and 10.7 GHz channels of the Nimbus-7 Scanning Multichannel Microwave Radiometer were analyzed to study the large-area soil moisture variations of land surfaces. Two regions in Texas, one with sparse and the other with dense vegetation covers, were selected for the study. The results gave a confirmation of the vegetation effect observed by ground-level microwave radiometers. Based on the statistics of the satellite data, it was possible to estimate surface soil moisture in about five different levels from dry to wet conditions with a 1.4 GHz radiometer, provided that the biomass of the vegetation cover could be independently measured. At frequencies greater than about 6.6 GHz, the radiometric measurements showed little sensitivity to moisture variation for vegetation-covered soils. The effects of polarization in microwave emission were studied also.
Sparse Downscaling and Adaptive Fusion of Multi-sensor Precipitation
NASA Astrophysics Data System (ADS)
Ebtehaj, M.; Foufoula, E.
2011-12-01
The past decades have witnessed a remarkable emergence of new sources of multiscale multi-sensor precipitation data including data from global spaceborne active and passive sensors, regional ground based weather surveillance radars and local rain-gauges. Resolution enhancement of remotely sensed rainfall and optimal integration of multi-sensor data promise a posteriori estimates of precipitation fluxes with increased accuracy and resolution to be used in hydro-meteorological applications. In this context, new frameworks are proposed for resolution enhancement and multiscale multi-sensor precipitation data fusion, which capitalize on two main observations: (1) sparseness of remotely sensed precipitation fields in appropriately chosen transformed domains, (e.g., in wavelet space) which promotes the use of the newly emerged theory of sparse representation and compressive sensing for resolution enhancement; (2) a conditionally Gaussian Scale Mixture (GSM) parameterization in the wavelet domain which allows exploiting the efficient linear estimation methodologies, while capturing the non-Gaussian data structure of rainfall. The proposed methodologies are demonstrated using a data set of coincidental observations of precipitation reflectivity images by the spaceborne precipitation radar (PR) aboard the Tropical Rainfall Measurement Mission (TRMM) satellite and ground-based NEXRAD weather surveillance Doppler radars. Uniqueness and stability of the solution, capturing non-Gaussian singular structure of rainfall, reduced uncertainty of estimation and efficiency of computation are the main advantages of the proposed methodologies over the commonly used standard Gaussian techniques.
Dynamic Facial Expression Recognition With Atlas Construction and Sparse Representation.
Guo, Yimo; Zhao, Guoying; Pietikainen, Matti
2016-05-01
In this paper, a new dynamic facial expression recognition method is proposed. Dynamic facial expression recognition is formulated as a longitudinal groupwise registration problem. The main contributions of this method lie in the following aspects: 1) subject-specific facial feature movements of different expressions are described by a diffeomorphic growth model; 2) salient longitudinal facial expression atlas is built for each expression by a sparse groupwise image registration method, which can describe the overall facial feature changes among the whole population and can suppress the bias due to large intersubject facial variations; and 3) both the image appearance information in spatial domain and topological evolution information in temporal domain are used to guide recognition by a sparse representation method. The proposed framework has been extensively evaluated on five databases for different applications: the extended Cohn-Kanade, MMI, FERA, and AFEW databases for dynamic facial expression recognition, and UNBC-McMaster database for spontaneous pain expression monitoring. This framework is also compared with several state-of-the-art dynamic facial expression recognition methods. The experimental results demonstrate that the recognition rates of the new method are consistently higher than other methods under comparison. PMID:26955032
Guided-mode resonance nanophotonics in materially sparse architectures
NASA Astrophysics Data System (ADS)
Magnusson, Robert; Niraula, Manoj; Yoon, Jae W.; Ko, Yeong H.; Lee, Kyu J.
2016-03-01
The guided-mode resonance (GMR) concept refers to lateral quasi-guided waveguide modes induced in periodic layers. Whereas these effects have been known for a long time, new attributes and innovations continue to appear. Here, we review some recent progress in this field with emphasis on sparse, or minimal, device embodiments. We discuss properties of wideband resonant reflectors designed with gratings in which the grating ridges are matched to an identical material to eliminate local reflections and phase changes. This critical interface therefore possesses zero refractive-index contrast; hence we call them "zero-contrast gratings." Applying this architecture, we present single-layer, wideband reflectors that are robust under experimentally realistic parametric variations. We introduce a new class of reflectors and polarizers fashioned with dielectric nanowire grids that are mostly empty space. Computed results predict high reflection and attendant polarization extinction for these sparse lattices. Experimental verification with Si nanowire grids yields ~200-nm-wide band of high reflection for one polarization state and free transmission of the orthogonal state. Finally, we present bandpass filters using all-dielectric resonant gratings. We design, fabricate, and test nanostructured single layer filters exhibiting high efficiency and sub-nanometer-wide passbands surrounded by 100-nm-wide stopbands.
Configurable hardware integrate and fire neurons for sparse approximation.
Shapero, Samuel; Rozell, Christopher; Hasler, Paul
2013-09-01
Sparse approximation is an important optimization problem in signal and image processing applications. A Hopfield-Network-like system of integrate and fire (IF) neurons is proposed as a solution, using the Locally Competitive Algorithm (LCA) to solve an overcomplete L1 sparse approximation problem. A scalable system architecture is described, including IF neurons with a nonlinear firing function, and current-based synapses to provide linear computation. A network of 18 neurons with 12 inputs is implemented on the RASP 2.9v chip, a Field Programmable Analog Array (FPAA) with directly programmable floating gate elements. Said system uses over 1400 floating gates, the largest system programmed on a FPAA to date. The circuit successfully reproduced the outputs of a digital optimization program, converging to within 4.8% RMS, and an objective cost only 1.7% higher on average. The active circuit consumed 559 μA of current at 2.4 V and converges on solutions in 25 μs, with measurement of the converged spike rate taking an additional 1 ms. Extrapolating the scaling trends to a N=1000 node system, the spiking LCA compares favorably with state-of-the-art digital solutions, and analog solutions using a non-spiking approach.
Sparse expression bases in cancer reveal tumor drivers
Logsdon, Benjamin A.; Gentles, Andrew J.; Miller, Chris P.; Blau, C. Anthony; Becker, Pamela S.; Lee, Su-In
2015-01-01
We define a new category of candidate tumor drivers in cancer genome evolution: ‘selected expression regulators’ (SERs)—genes driving dysregulated transcriptional programs in cancer evolution. The SERs are identified from genome-wide tumor expression data with a novel method, namely SPARROW (SPARse selected expRessiOn regulators identified With penalized regression). SPARROW uncovers a previously unknown connection between cancer expression variation and driver events, by using a novel sparse regression technique. Our results indicate that SPARROW is a powerful complementary approach to identify candidate genes containing driver events that are hard to detect from sequence data, due to a large number of passenger mutations and lack of comprehensive sequence information from a sufficiently large number of samples. SERs identified by SPARROW reveal known driver mutations in multiple human cancers, along with known cancer-associated processes and survival-associated genes, better than popular methods for inferring gene expression networks. We demonstrate that when applied to acute myeloid leukemia expression data, SPARROW identifies an apoptotic biomarker (PYCARD) for an investigational drug obatoclax. The PYCARD and obatoclax association is validated in 30 AML patient samples. PMID:25583238
Inpainting of historical seismograms using sparse representation method
NASA Astrophysics Data System (ADS)
Wang, Lifu; Sun, Yi; Cai, Xiaogang
2015-01-01
This paper presents a method of inpainting historical seismograms recorded by a pen and paper drum-type seismograph. In the seismogram, some portions of the wave may be lost or distorted owing to time marks or violent shaking. In this study, the seismic waveform is divided into several frames of equal length, and the lost or distorted portions are restored frame by frame. Because a seismogram contains several repetitive patterns in the entire waveform, each frame can be sparsely represented on the basis of these patterns. Therefore, the sparse representation model is employed to represent historical seismograms. In addition, an inpainting model that employs sparsity as a prior is formulated, and it is used to restore the lost portions by solving a L0-norm minimization problem. However, this minimization problem may be ill posed and result in an incorrect outcome if the missing interval duration of the wave is very large. Therefore, to solve this ill-posed problem, a prior based on the Fourier spectrum of the waveform is added to the inpainting method. Simulation results prove that the proposed inpainting method can restore the missing wave well.
Online Signature Verification Based on DCT and Sparse Representation.
Liu, Yishu; Yang, Zhihua; Yang, Lihua
2015-11-01
In this paper, a novel online signature verification technique based on discrete cosine transform (DCT) and sparse representation is proposed. We find a new property of DCT, which can be used to obtain a compact representation of an online signature using a fixed number of coefficients, leading to simple matching procedures and providing an effective alternative to deal with time series of different lengths. The property is also used to extract energy features. Furthermore, a new attempt to apply sparse representation to online signature verification is made, and a novel task-specific method for building overcomplete dictionaries is proposed, then sparsity features are extracted. Finally, energy features and sparsity features are concatenated to form a feature vector. Experiments are conducted on the Sabancı University's Signature Database (SUSIG)-Visual and SVC2004 databases, and the results show that our proposed method authenticates persons very reliably with a verification performance which is better than those of state-of-the-art methods on the same databases.
Sparse regression and marginal testing using cluster prototypes.
Reid, Stephen; Tibshirani, Robert
2016-04-01
We propose a new approach for sparse regression and marginal testing, for data with correlated features. Our procedure first clusters the features, and then chooses as the cluster prototype the most informative feature in that cluster. Then we apply either sparse regression (lasso) or marginal significance testing to these prototypes. While this kind of strategy is not entirely new, a key feature of our proposal is its use of the post-selection inference theory of Taylor and others (2014, Exact post-selection inference for forward stepwise and least angle regression, Preprint, arXiv:1401.3889) and Lee and others (2014, Exact post-selection inference with the lasso, Preprint, arXiv:1311.6238v5) to compute exact [Formula: see text]-values and confidence intervals that properly account for the selection of prototypes. We also apply the recent "knockoff" idea of Barber and Candès (2014, Controlling the false discovery rate via knockoffs, Preprint, arXiv:1404.5609) to provide exact finite sample control of the FDR of our regression procedure. We illustrate our proposals on both real and simulated data.
Synthesizing spatiotemporally sparse smartphone sensor data for bridge modal identification
NASA Astrophysics Data System (ADS)
Ozer, Ekin; Feng, Maria Q.
2016-08-01
Smartphones as vibration measurement instruments form a large-scale, citizen-induced, and mobile wireless sensor network (WSN) for system identification and structural health monitoring (SHM) applications. Crowdsourcing-based SHM is possible with a decentralized system granting citizens with operational responsibility and control. Yet, citizen initiatives introduce device mobility, drastically changing SHM results due to uncertainties in the time and the space domains. This paper proposes a modal identification strategy that fuses spatiotemporally sparse SHM data collected by smartphone-based WSNs. Multichannel data sampled with the time and the space independence is used to compose the modal identification parameters such as frequencies and mode shapes. Structural response time history can be gathered by smartphone accelerometers and converted into Fourier spectra by the processor units. Timestamp, data length, energy to power conversion address temporal variation, whereas spatial uncertainties are reduced by geolocation services or determining node identity via QR code labels. Then, parameters collected from each distributed network component can be extended to global behavior to deduce modal parameters without the need of a centralized and synchronous data acquisition system. The proposed method is tested on a pedestrian bridge and compared with a conventional reference monitoring system. The results show that the spatiotemporally sparse mobile WSN data can be used to infer modal parameters despite non-overlapping sensor operation schedule.
Pedestrian detection from thermal images: A sparse representation based approach
NASA Astrophysics Data System (ADS)
Qi, Bin; John, Vijay; Liu, Zheng; Mita, Seiichi
2016-05-01
Pedestrian detection, a key technology in computer vision, plays a paramount role in the applications of advanced driver assistant systems (ADASs) and autonomous vehicles. The objective of pedestrian detection is to identify and locate people in a dynamic environment so that accidents can be avoided. With significant variations introduced by illumination, occlusion, articulated pose, and complex background, pedestrian detection is a challenging task for visual perception. Different from visible images, thermal images are captured and presented with intensity maps based objects' emissivity, and thus have an enhanced spectral range to make human beings perceptible from the cool background. In this study, a sparse representation based approach is proposed for pedestrian detection from thermal images. We first adopted the histogram of sparse code to represent image features and then detect pedestrian with the extracted features in an unimodal and a multimodal framework respectively. In the unimodal framework, two types of dictionaries, i.e. joint dictionary and individual dictionary, are built by learning from prepared training samples. In the multimodal framework, a weighted fusion scheme is proposed to further highlight the contributions from features with higher separability. To validate the proposed approach, experiments were conducted to compare with three widely used features: Haar wavelets (HWs), histogram of oriented gradients (HOG), and histogram of phase congruency (HPC) as well as two classification methods, i.e. AdaBoost and support vector machine (SVM). Experimental results on a publicly available data set demonstrate the superiority of the proposed approach.
Robust Pedestrian Classification Based on Hierarchical Kernel Sparse Representation.
Sun, Rui; Zhang, Guanghai; Yan, Xiaoxing; Gao, Jun
2016-08-16
Vision-based pedestrian detection has become an active topic in computer vision and autonomous vehicles. It aims at detecting pedestrians appearing ahead of the vehicle using a camera so that autonomous vehicles can assess the danger and take action. Due to varied illumination and appearance, complex background and occlusion pedestrian detection in outdoor environments is a difficult problem. In this paper, we propose a novel hierarchical feature extraction and weighted kernel sparse representation model for pedestrian classification. Initially, hierarchical feature extraction based on a CENTRIST descriptor is used to capture discriminative structures. A max pooling operation is used to enhance the invariance of varying appearance. Then, a kernel sparse representation model is proposed to fully exploit the discrimination information embedded in the hierarchical local features, and a Gaussian weight function as the measure to effectively handle the occlusion in pedestrian images. Extensive experiments are conducted on benchmark databases, including INRIA, Daimler, an artificially generated dataset and a real occluded dataset, demonstrating the more robust performance of the proposed method compared to state-of-the-art pedestrian classification methods.
Music emotion detection using hierarchical sparse kernel machines.
Chin, Yu-Hao; Lin, Chang-Hong; Siahaan, Ernestasia; Wang, Jia-Ching
2014-01-01
For music emotion detection, this paper presents a music emotion verification system based on hierarchical sparse kernel machines. With the proposed system, we intend to verify if a music clip possesses happiness emotion or not. There are two levels in the hierarchical sparse kernel machines. In the first level, a set of acoustical features are extracted, and principle component analysis (PCA) is implemented to reduce the dimension. The acoustical features are utilized to generate the first-level decision vector, which is a vector with each element being a significant value of an emotion. The significant values of eight main emotional classes are utilized in this paper. To calculate the significant value of an emotion, we construct its 2-class SVM with calm emotion as the global (non-target) side of the SVM. The probability distributions of the adopted acoustical features are calculated and the probability product kernel is applied in the first-level SVMs to obtain first-level decision vector feature. In the second level of the hierarchical system, we merely construct a 2-class relevance vector machine (RVM) with happiness as the target side and other emotions as the background side of the RVM. The first-level decision vector is used as the feature with conventional radial basis function kernel. The happiness verification threshold is built on the probability value. In the experimental results, the detection error tradeoff (DET) curve shows that the proposed system has a good performance on verifying if a music clip reveals happiness emotion.
Sparse Gamma Rhythms Arising through Clustering in Adapting Neuronal Networks
Kilpatrick, Zachary P.; Ermentrout, Bard
2011-01-01
Gamma rhythms (30–100 Hz) are an extensively studied synchronous brain state responsible for a number of sensory, memory, and motor processes. Experimental evidence suggests that fast-spiking interneurons are responsible for carrying the high frequency components of the rhythm, while regular-spiking pyramidal neurons fire sparsely. We propose that a combination of spike frequency adaptation and global inhibition may be responsible for this behavior. Excitatory neurons form several clusters that fire every few cycles of the fast oscillation. This is first shown in a detailed biophysical network model and then analyzed thoroughly in an idealized model. We exploit the fact that the timescale of adaptation is much slower than that of the other variables. Singular perturbation theory is used to derive an approximate periodic solution for a single spiking unit. This is then used to predict the relationship between the number of clusters arising spontaneously in the network as it relates to the adaptation time constant. We compare this to a complementary analysis that employs a weak coupling assumption to predict the first Fourier mode to destabilize from the incoherent state of an associated phase model as the external noise is reduced. Both approaches predict the same scaling of cluster number with respect to the adaptation time constant, which is corroborated in numerical simulations of the full system. Thus, we develop several testable predictions regarding the formation and characteristics of gamma rhythms with sparsely firing excitatory neurons. PMID:22125486
Optimization-based multiple-point geostatistics: A sparse way
NASA Astrophysics Data System (ADS)
Kalantari, Sadegh; Abdollahifard, Mohammad Javad
2016-10-01
In multiple-point simulation the image should be synthesized consistent with the given training image and hard conditioning data. Existing sequential simulation methods usually lead to error accumulation which is hardly manageable in future steps. Optimization-based methods are capable of handling inconsistencies by iteratively refining the simulation grid. In this paper, the multiple-point stochastic simulation problem is formulated in an optimization-based framework using a sparse model. Sparse model allows each patch to be constructed as a superposition of a few atoms of a dictionary formed using training patterns, leading to a significant increase in the variability of the patches. To control the creativity of the model, a local histogram matching method is proposed. Furthermore, effective solutions are proposed for different issues arisen in multiple-point simulation. In order to handle hard conditioning data a weighted matching pursuit method is developed in this paper. Moreover, a simple and efficient thresholding method is developed which allows working with categorical variables. The experiments show that the proposed method produces acceptable realizations in terms of pattern reproduction, increases the variability of the realizations, and properly handles numerous conditioning data.
Asteroid models from combined sparse and dense photometric data
NASA Astrophysics Data System (ADS)
Durech, J.; Kaasalainen, M.; Warner, B. D.; Fauerbach, M.; Marks, S. A.; Fauvaud, S.; Fauvaud, M.; Vugnon, J.-M.; Pilcher, F.; Bernasconi, L.; Behrend, R.
2009-01-01
Aims: Shape and spin state are basic physical characteristics of an asteroid. They can be derived from disc-integrated photometry by the lightcurve inversion method. Increasing the number of asteroids with known basic physical properties is necessary to better understand the nature of individual objects as well as for studies of the whole asteroid population. Methods: We use the lightcurve inversion method to obtain rotation parameters and coarse shape models of selected asteroids. We combine sparse photometric data from the US Naval Observatory with ordinary lightcurves from the Uppsala Asteroid Photometric Catalogue and the Palmer Divide Observatory archive, and show that such combined data sets are in many cases sufficient to derive a model even if neither sparse photometry nor lightcurves can be used alone. Our approach is tested on multiple-apparition lightcurve inversion models and we show that the method produces consistent results. Results: We present new shape models and spin parameters for 24 asteroids. The shape models are only coarse but describe the global shape characteristics well. The typical error in the pole direction is ~10-20°. For a further 18 asteroids, inversion led to a unique determination of the rotation period but the pole direction was not well constrained. In these cases we give only an estimate of the ecliptic latitude of the pole.
One-step Sparse Estimates in Nonconcave Penalized Likelihood Models.
Zou, Hui; Li, Runze
2008-08-01
Fan & Li (2001) propose a family of variable selection methods via penalized likelihood using concave penalty functions. The nonconcave penalized likelihood estimators enjoy the oracle properties, but maximizing the penalized likelihood function is computationally challenging, because the objective function is nondifferentiable and nonconcave. In this article we propose a new unified algorithm based on the local linear approximation (LLA) for maximizing the penalized likelihood for a broad class of concave penalty functions. Convergence and other theoretical properties of the LLA algorithm are established. A distinguished feature of the LLA algorithm is that at each LLA step, the LLA estimator can naturally adopt a sparse representation. Thus we suggest using the one-step LLA estimator from the LLA algorithm as the final estimates. Statistically, we show that if the regularization parameter is appropriately chosen, the one-step LLA estimates enjoy the oracle properties with good initial estimators. Computationally, the one-step LLA estimation methods dramatically reduce the computational cost in maximizing the nonconcave penalized likelihood. We conduct some Monte Carlo simulation to assess the finite sample performance of the one-step sparse estimation methods. The results are very encouraging.
Synaptic mechanisms underlying sparse coding of active touch.
Crochet, Sylvain; Poulet, James F A; Kremer, Yves; Petersen, Carl C H
2011-03-24
Sensory information is actively gathered by animals, but the synaptic mechanisms driving neuronal circuit function during active sensory processing are poorly understood. Here, we investigated the synaptically driven membrane potential dynamics during active whisker sensation using whole-cell recordings from layer 2/3 pyramidal neurons in the primary somatosensory barrel cortex of behaving mice. Although whisker contact with an object evoked rapid depolarization in all neurons, these touch responses only drove action potentials in ∼10% of the cells. Such sparse coding was ensured by cell-specific reversal potentials of the touch-evoked response that were hyperpolarized relative to action potential threshold for most neurons. Intercontact interval profoundly influenced touch-evoked postsynaptic potentials, interestingly without affecting the peak membrane potential of the touch response. Dual whole-cell recordings indicated highly correlated membrane potential dynamics during active touch. Sparse action potential firing within synchronized cortical layer 2/3 microcircuits therefore appears to robustly signal each active touch response.
Tissue-specific sparse deconvolution for brain CT perfusion.
Fang, Ruogu; Jiang, Haodi; Huang, Junzhou
2015-12-01
Enhancing perfusion maps in low-dose computed tomography perfusion (CTP) for cerebrovascular disease diagnosis is a challenging task, especially for low-contrast tissue categories where infarct core and ischemic penumbra usually occur. Sparse perfusion deconvolution has been recently proposed to effectively improve the image quality and diagnostic accuracy of low-dose perfusion CT by extracting the complementary information from the high-dose perfusion maps to restore the low-dose using a joint spatio-temporal model. However the low-contrast tissue classes where infarct core and ischemic penumbra are likely to occur in cerebral perfusion CT tend to be over-smoothed, leading to loss of essential biomarkers. In this paper, we propose a tissue-specific sparse deconvolution approach to preserve the subtle perfusion information in the low-contrast tissue classes. We first build tissue-specific dictionaries from segmentations of high-dose perfusion maps using online dictionary learning, and then perform deconvolution-based hemodynamic parameters estimation for block-wise tissue segments on the low-dose CTP data. Extensive validation on clinical datasets of patients with cerebrovascular disease demonstrates the superior performance of our proposed method compared to state-of-art, and potentially improve diagnostic accuracy by increasing the differentiation between normal and ischemic tissues in the brain. PMID:26055434
3D ear identification based on sparse representation.
Zhang, Lin; Ding, Zhixuan; Li, Hongyu; Shen, Ying
2014-01-01
Biometrics based personal authentication is an effective way for automatically recognizing, with a high confidence, a person's identity. Recently, 3D ear shape has attracted tremendous interests in research field due to its richness of feature and ease of acquisition. However, the existing ICP (Iterative Closet Point)-based 3D ear matching methods prevalent in the literature are not quite efficient to cope with the one-to-many identification case. In this paper, we aim to fill this gap by proposing a novel effective fully automatic 3D ear identification system. We at first propose an accurate and efficient template-based ear detection method. By utilizing such a method, the extracted ear regions are represented in a common canonical coordinate system determined by the ear contour template, which facilitates much the following stages of feature extraction and classification. For each extracted 3D ear, a feature vector is generated as its representation by making use of a PCA-based local feature descriptor. At the stage of classification, we resort to the sparse representation based classification approach, which actually solves an l1-minimization problem. To the best of our knowledge, this is the first work introducing the sparse representation framework into the field of 3D ear identification. Extensive experiments conducted on a benchmark dataset corroborate the effectiveness and efficiency of the proposed approach. The associated Matlab source code and the evaluation results have been made publicly online available at http://sse.tongji.edu.cn/linzhang/ear/srcear/srcear.htm.
Improved beamforming using curved sparse 2D arrays in ultrasound.
Kirkebø, Jan Egil; Austeng, Andreas
2007-05-01
In this work we have investigated the effect of curving phase-steered sparse periodic two-dimensional arrays in one direction, and relate this effect to the geometry of the arrays. We have shown that curving is equivalent to removing some of the element periodicity, thus adding some "randomness" to the layout. Compared to flat phase-steered periodically sparse two-dimensional arrays, curving offers an even greater suppression of grating lobes located at directions along the curvature. The class of arrays yielding improved performance due to this suppression of grating lobes has been characterized. The point spread functions of some previously proposed array layouts, shown to be promising for ultrasonic imaging, have been simulated. The arrays have been simulated with various number of elements as well as various focal points, with array and field parameters typical to those in volumetric cardiac imaging. On a 48 x 48 element grid with a transducer center frequency of 3 MHz and the target at 40 mm, reductions in the peak sidelobe level of up to 12dB were recorded for some critical steering directions, without significant differences in the beamwidth. The integrated sidelobe ratio was also examined, showing an almost equivalent performance as the flat array. This study shows that, without adding any complexity to the system, the overall image quality of a volumetric imaging system can be improved significantly by curving the array in one direction. PMID:17313966
Robust Pedestrian Classification Based on Hierarchical Kernel Sparse Representation
Sun, Rui; Zhang, Guanghai; Yan, Xiaoxing; Gao, Jun
2016-01-01
Vision-based pedestrian detection has become an active topic in computer vision and autonomous vehicles. It aims at detecting pedestrians appearing ahead of the vehicle using a camera so that autonomous vehicles can assess the danger and take action. Due to varied illumination and appearance, complex background and occlusion pedestrian detection in outdoor environments is a difficult problem. In this paper, we propose a novel hierarchical feature extraction and weighted kernel sparse representation model for pedestrian classification. Initially, hierarchical feature extraction based on a CENTRIST descriptor is used to capture discriminative structures. A max pooling operation is used to enhance the invariance of varying appearance. Then, a kernel sparse representation model is proposed to fully exploit the discrimination information embedded in the hierarchical local features, and a Gaussian weight function as the measure to effectively handle the occlusion in pedestrian images. Extensive experiments are conducted on benchmark databases, including INRIA, Daimler, an artificially generated dataset and a real occluded dataset, demonstrating the more robust performance of the proposed method compared to state-of-the-art pedestrian classification methods. PMID:27537888
Threshold partitioning of sparse matrices and applications to Markov chains
Choi, Hwajeong; Szyld, D.B.
1996-12-31
It is well known that the order of the variables and equations of a large, sparse linear system influences the performance of classical iterative methods. In particular if, after a symmetric permutation, the blocks in the diagonal have more nonzeros, classical block methods have a faster asymptotic rate of convergence. In this paper, different ordering and partitioning algorithms for sparse matrices are presented. They are modifications of PABLO. In the new algorithms, in addition to the location of the nonzeros, the values of the entries are taken into account. The matrix resulting after the symmetric permutation has dense blocks along the diagonal, and small entries in the off-diagonal blocks. Parameters can be easily adjusted to obtain, for example, denser blocks, or blocks with elements of larger magnitude. In particular, when the matrices represent Markov chains, the permuted matrices are well suited for block iterative methods that find the corresponding probability distribution. Applications to three types of methods are explored: (1) Classical block methods, such as Block Gauss Seidel. (2) Preconditioned GMRES, where a block diagonal preconditioner is used. (3) Iterative aggregation method (also called aggregation/disaggregation) where the partition obtained from the ordering algorithm with certain parameters is used as an aggregation scheme. In all three cases, experiments are presented which illustrate the performance of the methods with the new orderings. The complexity of the new algorithms is linear in the number of nonzeros and the order of the matrix, and thus adding little computational effort to the overall solution.
Structure-preserving sparse decomposition for facial expression analysis.
Taheri, Sima; Qiang Qiu; Chellappa, Rama
2014-08-01
Although facial expressions can be decomposed in terms of action units (AUs) as suggested by the facial action coding system, there have been only a few attempts that recognize expression using AUs and their composition rules. In this paper, we propose a dictionary-based approach for facial expression analysis by decomposing expressions in terms of AUs. First, we construct an AU-dictionary using domain experts' knowledge of AUs. To incorporate the high-level knowledge regarding expression decomposition and AUs, we then perform structure-preserving sparse coding by imposing two layers of grouping over AU-dictionary atoms as well as over the test image matrix columns. We use the computed sparse code matrix for each expressive face to perform expression decomposition and recognition. Since domain experts' knowledge may not always be available for constructing an AU-dictionary, we also propose a structure-preserving dictionary learning algorithm, which we use to learn a structured dictionary as well as divide expressive faces into several semantic regions. Experimental results on publicly available expression data sets demonstrate the effectiveness of the proposed approach for facial expression analysis.
Framelet-Based Sparse Unmixing of Hyperspectral Images.
Zhang, Guixu; Xu, Yingying; Fang, Faming
2016-04-01
Spectral unmixing aims at estimating the proportions (abundances) of pure spectrums (endmembers) in each mixed pixel of hyperspectral data. Recently, a semi-supervised approach, which takes the spectral library as prior knowledge, has been attracting much attention in unmixing. In this paper, we propose a new semi-supervised unmixing model, termed framelet-based sparse unmixing (FSU), which promotes the abundance sparsity in framelet domain and discriminates the approximation and detail components of hyperspectral data after framelet decomposition. Due to the advantages of the framelet representations, e.g., images have good sparse approximations in framelet domain, and most of the additive noises are included in the detail coefficients, the FSU model has a better antinoise capability, and accordingly leads to more desirable unmixing performance. The existence and uniqueness of the minimizer of the FSU model are then discussed, and the split Bregman algorithm and its convergence property are presented to obtain the minimal solution. Experimental results on both simulated data and real data demonstrate that the FSU model generally performs better than the compared methods. PMID:26849863
Method and apparatus for optimized processing of sparse matrices
Taylor, Valerie E.
1993-01-01
A computer architecture for processing a sparse matrix is disclosed. The apparatus stores a value-row vector corresponding to nonzero values of a sparse matrix. Each of the nonzero values is located at a defined row and column position in the matrix. The value-row vector includes a first vector including nonzero values and delimiting characters indicating a transition from one column to another. The value-row vector also includes a second vector which defines row position values in the matrix corresponding to the nonzero values in the first vector and column position values in the matrix corresponding to the column position of the nonzero values in the first vector. The architecture also includes a circuit for detecting a special character within the value-row vector. Matrix-vector multiplication is executed on the value-row vector. This multiplication is performed by multiplying an index value of the first vector value by a column value from a second matrix to form a matrix-vector product which is added to a previous matrix-vector product.
Fault feature extraction of rolling element bearings using sparse representation
NASA Astrophysics Data System (ADS)
He, Guolin; Ding, Kang; Lin, Huibin
2016-03-01
Influenced by factors such as speed fluctuation, rolling element sliding and periodical variation of load distribution and impact force on the measuring direction of sensor, the impulse response signals caused by defective rolling bearing are non-stationary, and the amplitudes of the impulse may even drop to zero when the fault is out of load zone. The non-stationary characteristic and impulse missing phenomenon reduce the effectiveness of the commonly used demodulation method on rolling element bearing fault diagnosis. Based on sparse representation theories, a new approach for fault diagnosis of rolling element bearing is proposed. The over-complete dictionary is constructed by the unit impulse response function of damped second-order system, whose natural frequencies and relative damping ratios are directly identified from the fault signal by correlation filtering method. It leads to a high similarity between atoms and defect induced impulse, and also a sharply reduction of the redundancy of the dictionary. To improve the matching accuracy and calculation speed of sparse coefficient solving, the fault signal is divided into segments and the matching pursuit algorithm is carried out by segments. After splicing together all the reconstructed signals, the fault feature is extracted successfully. The simulation and experimental results show that the proposed method is effective for the fault diagnosis of rolling element bearing in large rolling element sliding and low signal to noise ratio circumstances.
Sparse Representation of Electrodermal Activity With Knowledge-Driven Dictionaries
Tsiartas, Andreas; Stein, Leah I.; Cermak, Sharon A.; Narayanan, Shrikanth S.
2015-01-01
Biometric sensors and portable devices are being increasingly embedded into our everyday life, creating the need for robust physiological models that efficiently represent, analyze, and interpret the acquired signals. We propose a knowledge-driven method to represent electrodermal activity (EDA), a psychophysiological signal linked to stress, affect, and cognitive processing. We build EDA-specific dictionaries that accurately model both the slow varying tonic part and the signal fluctuations, called skin conductance responses (SCR), and use greedy sparse representation techniques to decompose the signal into a small number of atoms from the dictionary. Quantitative evaluation of our method considers signal reconstruction, compression rate, and information retrieval measures, that capture the ability of the model to incorporate the main signal characteristics, such as SCR occurrences. Compared to previous studies fitting a predetermined structure to the signal, results indicate that our approach provides benefits across all aforementioned criteria. This paper demonstrates the ability of appropriate dictionaries along with sparse decomposition methods to reliably represent EDA signals and provides a foundation for automatic measurement of SCR characteristics and the extraction of meaningful EDA features. PMID:25494494
Population coding in sparsely connected networks of noisy neurons
Tripp, Bryan P.; Orchard, Jeff
2012-01-01
This study examines the relationship between population coding and spatial connection statistics in networks of noisy neurons. Encoding of sensory information in the neocortex is thought to require coordinated neural populations, because individual cortical neurons respond to a wide range of stimuli, and exhibit highly variable spiking in response to repeated stimuli. Population coding is rooted in network structure, because cortical neurons receive information only from other neurons, and because the information they encode must be decoded by other neurons, if it is to affect behavior. However, population coding theory has often ignored network structure, or assumed discrete, fully connected populations (in contrast with the sparsely connected, continuous sheet of the cortex). In this study, we modeled a sheet of cortical neurons with sparse, primarily local connections, and found that a network with this structure could encode multiple internal state variables with high signal-to-noise ratio. However, we were unable to create high-fidelity networks by instantiating connections at random according to spatial connection probabilities. In our models, high-fidelity networks required additional structure, with higher cluster factors and correlations between the inputs to nearby neurons. PMID:22586391
Hyperspectral anomaly detection using sparse kernel-based ensemble learning
NASA Astrophysics Data System (ADS)
Gurram, Prudhvi; Han, Timothy; Kwon, Heesung
2011-06-01
In this paper, sparse kernel-based ensemble learning for hyperspectral anomaly detection is proposed. The proposed technique is aimed to optimize an ensemble of kernel-based one class classifiers, such as Support Vector Data Description (SVDD) classifiers, by estimating optimal sparse weights. In this method, hyperspectral signatures are first randomly sub-sampled into a large number of spectral feature subspaces. An enclosing hypersphere that defines the support of spectral data, corresponding to the normalcy/background data, in the Reproducing Kernel Hilbert Space (RKHS) of each respective feature subspace is then estimated using regular SVDD. The enclosing hypersphere basically represents the spectral characteristics of the background data in the respective feature subspace. The joint hypersphere is learned by optimally combining the hyperspheres from the individual RKHS, while imposing the l1 constraint on the combining weights. The joint hypersphere representing the most optimal compact support of the local hyperspectral data in the joint feature subspaces is then used to test each pixel in hyperspectral image data to determine if it belongs to the local background data or not. The outliers are considered to be targets. The performance comparison between the proposed technique and the regular SVDD is provided using the HYDICE hyperspectral images.
Sparse MEG source imaging in Landau-Kleffner syndrome.
Zhu, Min; Zhang, Wenbo; Dickens, Deanna; Ding, Lei
2011-01-01
Epilepsy patients with Landau-Kleffner syndrome (LKS) usually have a normal brain structure, which makes it a challenge to identify the epileptogenic zone only based on magnetic resonance imaging (MRI) data. A sparse source imaging technique called variation based sparse cortical current density (VB-SCCD) imaging was adopted here to reconstruct cortical sources of magnetoencephalography (MEG) interictal spikes from an LKS patient. Realistic boundary element (BE) head and cortex models were built by segmenting structural MRI. 148-channel MEG was recorded for 10 minutes between seizures. Total 29 epileptiform spikes were selected for analysis. The primary cortical sources were observed locating at the left intra- and perisylvian cortex. Multiple extrasylvian sources were identified as the secondary sources. The spatio-temporal patterns of cortical sources provide more insights about the neuronal synchrony and propagation of epileptic discharges. Our observations were consistent with presurgical diagnosis for this patient and observation of aphasia in LKS. The present results suggest that the promising of VB-SCCD technique in assisting with presurgical planning and studying the neural network for LKS in determining the lateralization of epileptic origins. It can further be applied to non-invasively localize and/or lateralize eloquent cortex for language for epilepsy patients in general in the future.
Magnetic resonance brain tissue segmentation based on sparse representations
NASA Astrophysics Data System (ADS)
Rueda, Andrea
2015-12-01
Segmentation or delineation of specific organs and structures in medical images is an important task in the clinical diagnosis and treatment, since it allows to characterize pathologies through imaging measures (biomarkers). In brain imaging, segmentation of main tissues or specific structures is challenging, due to the anatomic variability and complexity, and the presence of image artifacts (noise, intensity inhomogeneities, partial volume effect). In this paper, an automatic segmentation strategy is proposed, based on sparse representations and coupled dictionaries. Image intensity patterns are singly related to tissue labels at the level of small patches, gathering this information in coupled intensity/segmentation dictionaries. This dictionaries are used within a sparse representation framework to find the projection of a new intensity image onto the intensity dictionary, and the same projection can be used with the segmentation dictionary to estimate the corresponding segmentation. Preliminary results obtained with two publicly available datasets suggest that the proposal is capable of estimating adequate segmentations for gray matter (GM) and white matter (WM) tissues, with an average overlapping of 0:79 for GM and 0:71 for WM (with respect to original segmentations).
Improved terahertz imaging with a sparse synthetic aperture array
NASA Astrophysics Data System (ADS)
Zhang, Zhuopeng; Buma, Takashi
2010-02-01
Sparse arrays are highly attractive for implementing two-dimensional arrays, but come at the cost of degraded image quality. We demonstrate significantly improved performance by exploiting the coherent ultrawideband nature of singlecycle THz pulses. We compute two weighting factors to each time-delayed signal before final summation to form the reconstructed image. The first factor employs cross-correlation analysis to measure the degree of walk-off between timedelayed signals of neighboring elements. The second factor measures the spatial coherence of the time-delayed delayed signals. Synthetic aperture imaging experiments are performed with a THz time-domain system employing a mechanically scanned single transceiver element. Cross-sectional imaging of wire targets is performed with a onedimensional sparse array with an inter-element spacing of 1.36 mm (over four λ at 1 THz). The proposed image reconstruction technique improves image contrast by 15 dB, which is impressive considering the relatively few elements in the array. En-face imaging of a razor blade is also demonstrated with a 56 x 56 element two-dimensional array, showing reduced image artifacts with adaptive reconstruction. These encouraging results suggest that the proposed image reconstruction technique can be highly beneficial to the development of large area two-dimensional THz arrays.
Dictionary learning and sparse recovery for electrodermal activity analysis
NASA Astrophysics Data System (ADS)
Kelsey, Malia; Dallal, Ahmed; Eldeeb, Safaa; Akcakaya, Murat; Kleckner, Ian; Gerard, Christophe; Quigley, Karen S.; Goodwin, Matthew S.
2016-05-01
Measures of electrodermal activity (EDA) have advanced research in a wide variety of areas including psychophysiology; however, the majority of this research is typically undertaken in laboratory settings. To extend the ecological validity of laboratory assessments, researchers are taking advantage of advances in wireless biosensors to gather EDA data in ambulatory settings, such as in school classrooms. While measuring EDA in naturalistic contexts may enhance ecological validity, it also introduces analytical challenges that current techniques cannot address. One limitation is the limited efficiency and automation of analysis techniques. Many groups either analyze their data by hand, reviewing each individual record, or use computationally inefficient software that limits timely analysis of large data sets. To address this limitation, we developed a method to accurately and automatically identify SCRs using curve fitting methods. Curve fitting has been shown to improve the accuracy of SCR amplitude and location estimations, but have not yet been used to reduce computational complexity. In this paper, sparse recovery and dictionary learning methods are combined to improve computational efficiency of analysis and decrease run time, while maintaining a high degree of accuracy in detecting SCRs. Here, a dictionary is first created using curve fitting methods for a standard SCR shape. Then, orthogonal matching pursuit (OMP) is used to detect SCRs within a dataset using the dictionary to complete sparse recovery. Evaluation of our method, including a comparison to for speed and accuracy with existing software, showed an accuracy of 80% and a reduced run time.
Tissue-specific sparse deconvolution for brain CT perfusion.
Fang, Ruogu; Jiang, Haodi; Huang, Junzhou
2015-12-01
Enhancing perfusion maps in low-dose computed tomography perfusion (CTP) for cerebrovascular disease diagnosis is a challenging task, especially for low-contrast tissue categories where infarct core and ischemic penumbra usually occur. Sparse perfusion deconvolution has been recently proposed to effectively improve the image quality and diagnostic accuracy of low-dose perfusion CT by extracting the complementary information from the high-dose perfusion maps to restore the low-dose using a joint spatio-temporal model. However the low-contrast tissue classes where infarct core and ischemic penumbra are likely to occur in cerebral perfusion CT tend to be over-smoothed, leading to loss of essential biomarkers. In this paper, we propose a tissue-specific sparse deconvolution approach to preserve the subtle perfusion information in the low-contrast tissue classes. We first build tissue-specific dictionaries from segmentations of high-dose perfusion maps using online dictionary learning, and then perform deconvolution-based hemodynamic parameters estimation for block-wise tissue segments on the low-dose CTP data. Extensive validation on clinical datasets of patients with cerebrovascular disease demonstrates the superior performance of our proposed method compared to state-of-art, and potentially improve diagnostic accuracy by increasing the differentiation between normal and ischemic tissues in the brain.
Super-resolution imaging with radio interferometry using sparse modeling
NASA Astrophysics Data System (ADS)
Honma, Mareki; Akiyama, Kazunori; Uemura, Makoto; Ikeda, Shiro
2014-10-01
We propose a new technique to obtain super-resolution images with radio interferometry using sparse modeling. In standard radio interferometry, sampling of (u, v) is quite often incomplete and thus obtaining an image from observed visibilities becomes an underdetermined problem, and a technique of so-called "zero-padding" is often used to fill up unsampled grids in the (u, v) plane, resulting in image degradation by finite beam size as well as numerous side-lobes. In this paper we show that directly solving such an underdetermined problem based on sparse modeling (in this paper, Least Absolute Shrinkage and Selection Operator, known as LASSO) avoids the above problems introduced by zero-padding, leading to super-resolution images in which structure finer than the standard beam size (diffraction limit) can be reproduced. We present results of one-dimensional and two-dimensional simulations of interferometric imaging, and discuss its implications for super-resolution imaging, particularly focusing on imaging of black hole shadows with millimeter VLBI (Very Long Baseline Interferometry).
Sparse aperture mask for low order wavefront sensing
NASA Astrophysics Data System (ADS)
Subedi, Hari; Zimmerman, Neil T.; Kasdin, N. Jeremy; Cavanagh, Kathleen; Riggs, A. J. E.
2015-09-01
A high contrast is required for direct imaging of exoplanets. Ideally, the level of contrast required for direct imaging of exoplanets can be achieved by coronagraphic imaging, but in practice, the contrast is degraded by wavefront aberrations. To achieve the required contrast, low-order wavefront aberrations such as tip-tilt, defocus and coma must be determined and corrected. In this paper, we present a technique that integrates a sparse- aperture mask (SAM) with a shaped pupil coronagraph (SPC) to make precise estimates of these low-order aberrations. Starlight rejected by the coronagraph's focal plane stop is collimated to a relay pupil, where the mask forms an interference fringe pattern on a detector. Using numerical simulations, we show that the SAM can estimate rapidly varying tip-tilt errors in space telescopes arising from line-of-sight pointing oscillations as well as other higher-order modes. We also show that a Kalman filter can be used with the SAM to improve the estimation. At Princetons High Contrast Imaging Laboratory, we have recently created a testbed devoted to low-order wavefront sensing experiments. The testbed incorporates custom-fabricated masks (shaped pupil, focal plane, and sparse aperture) with a deformable mirror and a CCD camera to demonstrate the estimation and correction of low-order aberrations. Our first experiments aim to replicate the results of the SAM wavefront sensor (SAM WFS) Fourier propagation models.
Robot-Assisted Measurements in Data Sparse Regions
NASA Astrophysics Data System (ADS)
Peschel, J.; Young, S. N.; Penny, G.; Thompson, S. E.; Srinivasan, V.
2015-12-01
This work presents a methodology for topographic and bathymetric data collection using multiple robot platforms in the data sparse Arkavathy region around Bangalore, India. In the late 20th century, Arkavathy River flows began declining; consequently, a dependence on the Cauvery River has occurred while the reasons for the drying of the Arkavathy remain unknown. Understanding this shift is critical for managing local water resources, specifically for quantifying the socio-hydrologic effects of human intervention through the use of tanks which serve as a controlled method of irrigation for farmers. Determining the potential volume of water capable of being stored in these tanks can aid investigators to better understand hydrologic parameters such as recharge and streamflow. At present, satellite and LiDAR data are the two methods to collect topographic and bathymetric data for this region, but both options are either too poor of resolution or too costly. Small unmanned aerial vehicles (UAVs) and unmanned surface vehicles (USVs) were demonstrated as low-cost and reliable, high-resolution alternatives for surface data gathering at two locations in the Arkavathy basin during a Summer 2015 field campaign: i) Hadonahalli, and ii) SM Gollahalli. This robot-assisted approach for data gathering will be of interest to investigators in the geophysical sciences, especially those operating with budget constraints in data sparse regions.
Robust Pedestrian Classification Based on Hierarchical Kernel Sparse Representation.
Sun, Rui; Zhang, Guanghai; Yan, Xiaoxing; Gao, Jun
2016-01-01
Vision-based pedestrian detection has become an active topic in computer vision and autonomous vehicles. It aims at detecting pedestrians appearing ahead of the vehicle using a camera so that autonomous vehicles can assess the danger and take action. Due to varied illumination and appearance, complex background and occlusion pedestrian detection in outdoor environments is a difficult problem. In this paper, we propose a novel hierarchical feature extraction and weighted kernel sparse representation model for pedestrian classification. Initially, hierarchical feature extraction based on a CENTRIST descriptor is used to capture discriminative structures. A max pooling operation is used to enhance the invariance of varying appearance. Then, a kernel sparse representation model is proposed to fully exploit the discrimination information embedded in the hierarchical local features, and a Gaussian weight function as the measure to effectively handle the occlusion in pedestrian images. Extensive experiments are conducted on benchmark databases, including INRIA, Daimler, an artificially generated dataset and a real occluded dataset, demonstrating the more robust performance of the proposed method compared to state-of-the-art pedestrian classification methods. PMID:27537888
Sparse, Decorrelated Odor Coding in the Mushroom Body Enhances Learned Odor Discrimination
Lin, Andrew C.; Bygrave, Alexei; de Calignon, Alix; Lee, Tzumin; Miesenböck, Gero
2014-01-01
Summary Sparse coding may be a general strategy of neural systems to augment memory capacity. In Drosophila, sparse odor coding by the Kenyon cells of the mushroom body is thought to generate a large number of precisely addressable locations for the storage of odor-specific memories. However, it remains untested how sparse coding relates to behavioral performance. Here we demonstrate that sparseness is controlled by a negative feedback circuit between Kenyon cells and the GABAergic anterior paired lateral (APL) neuron. Systematic activation and blockade of each leg of this feedback circuit show that Kenyon cells activate APL and APL inhibits Kenyon cells. Disrupting the Kenyon cell-APL feedback loop decreases the sparseness of Kenyon cell odor responses, increases inter-odor correlations, and prevents flies from learning to discriminate similar, but not dissimilar, odors. These results suggest that feedback inhibition suppresses Kenyon cell activity to maintain sparse, decorrelated odor coding and thus the odor-specificity of memories. PMID:24561998
User's Manual for PCSMS (Parallel Complex Sparse Matrix Solver). Version 1.
NASA Technical Reports Server (NTRS)
Reddy, C. J.
2000-01-01
PCSMS (Parallel Complex Sparse Matrix Solver) is a computer code written to make use of the existing real sparse direct solvers to solve complex, sparse matrix linear equations. PCSMS converts complex matrices into real matrices and use real, sparse direct matrix solvers to factor and solve the real matrices. The solution vector is reconverted to complex numbers. Though, this utility is written for Silicon Graphics (SGI) real sparse matrix solution routines, it is general in nature and can be easily modified to work with any real sparse matrix solver. The User's Manual is written to make the user acquainted with the installation and operation of the code. Driver routines are given to aid the users to integrate PCSMS routines in their own codes.
Bagby, M.O.
1995-12-01
Vegetable oils serve various industrial applications such as plasticizers, emulsifiers, surfactants, plastics and resins. Research and development approaches may take advantage of natural properties of the oils. More often it is advantageous to modify those properties for specific applications. One example is the preparation of ink vehicles using vegetable oils in the absence of petroleum. They are cost competitive with petroleum-based inks with similar quality factors. Vegetable oils have potential as renewable sources of fuels for the diesel engine. However, several characteristics can restrict their use. These include poor cold-engine startup, misfire and for selected fuels, high pour point and cloud point temperatures. Other characteristics include incomplete combustion causing carbon buildup, lube oil dilution and degradation, and elevated NO{sub x} emissions. Precombustion and fuel quality data are presented as a tool for understanding and solving these operational and durability problems.
Feng, Huihui
2016-01-01
Climate and vegetation change are two dominating factors for soil moisture trend. However, their individual contributions remain unknown due to their complex interaction. Here, I separated their contributions through a trajectory-based method across the global, regional and local scales. Our results demonstrated that climate change accounted for 98.78% and 114.64% of the global drying and wetting trend. Vegetation change exhibited a relatively weak influence (contributing 1.22% and −14.64% of the global drying and wetting) because it occurred in a limited area on land. Regionally, the impact of vegetation change cannot be neglected, which contributed −40.21% of the soil moisture change in the wetting zone. Locally, the contributions strongly correlated to the local environmental characteristics. Vegetation negatively affected soil moisture trends in the dry and sparsely vegetated regions and positively in the wet and densely vegetated regions. I conclude that individual contributions of climate and vegetation change vary at the global, regional and local scales. Climate change dominates the soil moisture trends, while vegetation change acts as a regulator to drying or wetting the soil under the changing climate. PMID:27600157
Feng, Huihui
2016-01-01
Climate and vegetation change are two dominating factors for soil moisture trend. However, their individual contributions remain unknown due to their complex interaction. Here, I separated their contributions through a trajectory-based method across the global, regional and local scales. Our results demonstrated that climate change accounted for 98.78% and 114.64% of the global drying and wetting trend. Vegetation change exhibited a relatively weak influence (contributing 1.22% and -14.64% of the global drying and wetting) because it occurred in a limited area on land. Regionally, the impact of vegetation change cannot be neglected, which contributed -40.21% of the soil moisture change in the wetting zone. Locally, the contributions strongly correlated to the local environmental characteristics. Vegetation negatively affected soil moisture trends in the dry and sparsely vegetated regions and positively in the wet and densely vegetated regions. I conclude that individual contributions of climate and vegetation change vary at the global, regional and local scales. Climate change dominates the soil moisture trends, while vegetation change acts as a regulator to drying or wetting the soil under the changing climate. PMID:27600157
Feng, Huihui
2016-09-07
Climate and vegetation change are two dominating factors for soil moisture trend. However, their individual contributions remain unknown due to their complex interaction. Here, I separated their contributions through a trajectory-based method across the global, regional and local scales. Our results demonstrated that climate change accounted for 98.78% and 114.64% of the global drying and wetting trend. Vegetation change exhibited a relatively weak influence (contributing 1.22% and -14.64% of the global drying and wetting) because it occurred in a limited area on land. Regionally, the impact of vegetation change cannot be neglected, which contributed -40.21% of the soil moisture change in the wetting zone. Locally, the contributions strongly correlated to the local environmental characteristics. Vegetation negatively affected soil moisture trends in the dry and sparsely vegetated regions and positively in the wet and densely vegetated regions. I conclude that individual contributions of climate and vegetation change vary at the global, regional and local scales. Climate change dominates the soil moisture trends, while vegetation change acts as a regulator to drying or wetting the soil under the changing climate.
NASA Astrophysics Data System (ADS)
Masmoudi, Atef; Zouari, Sonia; Ghribi, Abdelaziz
2015-11-01
We propose a new adaptive block-wise lossless image compression algorithm, which is based on the so-called alphabet reduction scheme combined with an adaptive arithmetic coding (AC). This new encoding algorithm is particularly efficient for lossless compression of images with sparse and locally sparse histograms. AC is a very efficient technique for lossless data compression and produces a rate that is close to the entropy; however, a compression performance loss occurs when encoding images or blocks with a limited number of active symbols by comparison with the number of symbols in the nominal alphabet, which consists in the amplification of the zero frequency problem. Generally, most methods add one to the frequency count of each symbol from the nominal alphabet, which leads to a statistical model distortion, and therefore reduces the efficiency of the AC. The aim of this work is to overcome this drawback by assigning to each image block the smallest possible set including all the existing symbols called active symbols. This is an alternative of using the nominal alphabet when applying the conventional arithmetic encoders. We show experimentally that the proposed method outperforms several lossless image compression encoders and standards including the conventional arithmetic encoders, JPEG2000, and JPEG-LS.
Not Available
1980-11-01
A review is presented of various experiments undertaken over the past few years in the U.S. to test the performance of vegetable oils in diesel engines, mainly with a view to on-farm energy self-sufficiency. The USDA Northern Regional Research Center in Peoria, Illinois, is screening native U.S. plant species as potential fuel oil sources.
Fermented and Acidified Vegetables
Technology Transfer Automated Retrieval System (TEKTRAN)
Vegetables may be preserved by fermentation, direct acidification, or a combination of these along with pasteurization or refrigeration and selected additives to yield products with an extended shelf life and enhanced safety. Organic acids such as lactic, acetic, sorbic and benzoic acids along with ...
ERIC Educational Resources Information Center
Shepard, Mary; Shepard, Ray
Vegetable Soup is a new children's television series whose purpose is to counter the negative and destructive effects of racial isolation. This manual gives detailed instructions for discussion of activities that are presented during the television series such as: crafts, games, recipes, language activities, and children's questions. A list of…
Object-oriented algorithmic laboratory for ordering sparse matrices
Kumfert, G K
2000-05-01
We focus on two known NP-hard problems that have applications in sparse matrix computations: the envelope/wavefront reduction problem and the fill reduction problem. Envelope/wavefront reducing orderings have a wide range of applications including profile and frontal solvers, incomplete factorization preconditioning, graph reordering for cache performance, gene sequencing, and spatial databases. Fill reducing orderings are generally limited to--but an inextricable part of--sparse matrix factorization. Our major contribution to this field is the design of new and improved heuristics for these NP-hard problems and their efficient implementation in a robust, cross-platform, object-oriented software package. In this body of research, we (1) examine current ordering algorithms, analyze their asymptotic complexity, and characterize their behavior in model problems, (2) introduce new and improved algorithms that address deficiencies found in previous heuristics, (3) implement an object-oriented library of these algorithms in a robust, modular fashion without significant loss of efficiency, and (4) extend our algorithms and software to address both generalized and constrained problems. We stress that the major contribution is the algorithms and the implementation; the whole being greater than the sum of its parts. The initial motivation for implementing our algorithms in object-oriented software was to manage the inherent complexity. During our research came the realization that the object-oriented implementation enabled new possibilities augmented algorithms that would not have been as natural to generalize from a procedural implementation. Some extensions are constructed from a family of related algorithmic components, thereby creating a poly-algorithm that can adapt its strategy to the properties of the specific problem instance dynamically. Other algorithms are tailored for special constraints by aggregating algorithmic components and having them collaboratively
Improved image registration by sparse patch-based deformation estimation.
Kim, Minjeong; Wu, Guorong; Wang, Qian; Lee, Seong-Whan; Shen, Dinggang
2015-01-15
Despite intensive efforts for decades, deformable image registration is still a challenging problem due to the potential large anatomical differences across individual images, which limits the registration performance. Fortunately, this issue could be alleviated if a good initial deformation can be provided for the two images under registration, which are often termed as the moving subject and the fixed template, respectively. In this work, we present a novel patch-based initial deformation prediction framework for improving the performance of existing registration algorithms. Our main idea is to estimate the initial deformation between subject and template in a patch-wise fashion by using the sparse representation technique. We argue that two image patches should follow the same deformation toward the template image if their patch-wise appearance patterns are similar. To this end, our framework consists of two stages, i.e., the training stage and the application stage. In the training stage, we register all training images to the pre-selected template, such that the deformation of each training image with respect to the template is known. In the application stage, we apply the following four steps to efficiently calculate the initial deformation field for the new test subject: (1) We pick a small number of key points in the distinctive regions of the test subject; (2) for each key point, we extract a local patch and form a coupled appearance-deformation dictionary from training images where each dictionary atom consists of the image intensity patch as well as their respective local deformations; (3) a small set of training image patches in the coupled dictionary are selected to represent the image patch of each subject key point by sparse representation. Then, we can predict the initial deformation for each subject key point by propagating the pre-estimated deformations on the selected training patches with the same sparse representation coefficients; and (4) we
Sparse regularization techniques provide novel insights into outcome integration processes.
Mohr, Holger; Wolfensteller, Uta; Frimmel, Steffi; Ruge, Hannes
2015-01-01
By exploiting information that is contained in the spatial arrangement of neural activations, multivariate pattern analysis (MVPA) can detect distributed brain activations which are not accessible by standard univariate analysis. Recent methodological advances in MVPA regularization techniques have made it feasible to produce sparse discriminative whole-brain maps with highly specific patterns. Furthermore, the most recent refinement, the Graph Net, explicitly takes the 3D-structure of fMRI data into account. Here, these advanced classification methods were applied to a large fMRI sample (N=70) in order to gain novel insights into the functional localization of outcome integration processes. While the beneficial effect of differential outcomes is well-studied in trial-and-error learning, outcome integration in the context of instruction-based learning has remained largely unexplored. In order to examine neural processes associated with outcome integration in the context of instruction-based learning, two groups of subjects underwent functional imaging while being presented with either differential or ambiguous outcomes following the execution of varying stimulus-response instructions. While no significant univariate group differences were found in the resulting fMRI dataset, L1-regularized (sparse) classifiers performed significantly above chance and also clearly outperformed the standard L2-regularized (dense) Support Vector Machine on this whole-brain between-subject classification task. Moreover, additional L2-regularization via the Elastic Net and spatial regularization by the Graph Net improved interpretability of discriminative weight maps but were accompanied by reduced classification accuracies. Most importantly, classification based on sparse regularization facilitated the identification of highly specific regions differentially engaged under ambiguous and differential outcome conditions, comprising several prefrontal regions previously associated with
Reliable positioning in a sparse GPS network, eastern Ontario
NASA Astrophysics Data System (ADS)
Samadi Alinia, H.; Tiampo, K.; Atkinson, G. M.
2013-12-01
Canada hosts two regions that are prone to large earthquakes: western British Columbia, and the St. Lawrence River region in eastern Canada. Although eastern Ontario is not as seismically active as other areas of eastern Canada, such as the Charlevoix/Ottawa Valley seismic zone, it experiences ongoing moderate seismicity. In historic times, potentially damaging events have occurred in New York State (Attica, 1929, M=5.7; Plattsburg, 2002, M=5.0), north-central Ontario (Temiskaming, 1935, M=6.2; North Bay, 2000, M=5.0), eastern Ontario (Cornwall, 1944, M=5.8), Georgian Bay (2005, MN=4.3), and western Quebec (Val-Des-Bois,2010, M=5.0, MN=5.8). In eastern Canada, the analysis of detailed, high-precision measurements of surface deformation is a key component in our efforts to better characterize the associated seismic hazard. The data from precise, continuous GPS stations is necessary to adequately characterize surface velocities from which patterns and rates of stress accumulation on faults can be estimated (Mazzotti and Adams, 2005; Mazzotti et al., 2005). Monitoring of these displacements requires employing high accuracy GPS positioning techniques. Detailed strain measurements can determine whether the regional strain everywhere is commensurate with a large event occurring every few hundred years anywhere within this general area or whether large earthquakes are limited to specific areas (Adams and Halchuck, 2003; Mazzotti and Adams, 2005). In many parts of southeastern Ontario and western Québec, GPS stations are distributed quite sparsely, with spacings of approximately 100 km or more. The challenge is to provide accurate solutions for these sparse networks with an approach that is capable of achieving high-accuracy positioning. Here, various reduction techniques are applied to a sparse network installed with the Southern Ontario Seismic Network in eastern Ontario. Recent developments include the implementation of precise point positioning processing on acquired
Seasonal variation of carbon fluxes in a sparse savanna in semi arid Sudan
Ardö, Jonas; Mölder, Meelis; El-Tahir, Bashir Awad; Elkhidir, Hatim Abdalla Mohammed
2008-01-01
Background Large spatial, seasonal and annual variability of major drivers of the carbon cycle (precipitation, temperature, fire regime and nutrient availability) are common in the Sahel region. This causes large variability in net ecosystem exchange and in vegetation productivity, the subsistence basis for a major part of the rural population in Sahel. This study compares the 2005 dry and wet season fluxes of CO2 for a grass land/sparse savanna site in semi arid Sudan and relates these fluxes to water availability and incoming photosynthetic photon flux density (PPFD). Data from this site could complement the current sparse observation network in Africa, a continent where climatic change could significantly impact the future and which constitute a weak link in our understanding of the global carbon cycle. Results The dry season (represented by Julian day 35–46, February 2005) was characterized by low soil moisture availability, low evapotranspiration and a high vapor pressure deficit. The mean daily NEE (net ecosystem exchange, Eq. 1) was -14.7 mmol d-1 for the 12 day period (negative numbers denote sinks, i.e. flux from the atmosphere to the biosphere). The water use efficiency (WUE) was 1.6 mmol CO2 mol H2O-1 and the light use efficiency (LUE) was 0.95 mmol CO2 mol PPFD-1. Photosynthesis is a weak, but linear function of PPFD. The wet season (represented by Julian day 266–273, September 2005) was, compared to the dry season, characterized by slightly higher soil moisture availability, higher evapotranspiration and a slightly lower vapor pressure deficit. The mean daily NEE was -152 mmol d-1 for the 8 day period. The WUE was lower, 0.97 mmol CO2 mol H2O-1 and the LUE was higher, 7.2 μmol CO2 mmol PPFD-1 during the wet season compared to the dry season. During the wet season photosynthesis increases with PPFD to about 1600 μmol m-2s-1 and then levels off. Conclusion Based on data collected during two short periods, the studied ecosystem was a sink of carbon
When vegetation change alters ecosystem water availability.
Scott, Russell L; Huxman, Travis E; Barron-Gafford, Greg A; Darrel Jenerette, G; Young, Jessica M; Hamerlynck, Erik P
2014-07-01
The combined effects of vegetation and climate change on biosphere-atmosphere water vapor (H2 O) and carbon dioxide (CO2 ) exchanges are expected to vary depending, in part, on how biotic activity is controlled by and alters water availability. This is particularly important when a change in ecosystem composition alters the fractional covers of bare soil, grass, and woody plants so as to influence the accessibility of shallower vs. deeper soil water pools. To study this, we compared 5 years of eddy covariance measurements of H2 O and CO2 fluxes over a riparian grassland, shrubland, and woodland. In comparison with the surrounding upland region, groundwater access at the riparian sites increased net carbon uptake (NEP) and evapotranspiration (ET), which were sustained over more of the year. Among the sites, the grassland used less of the stable groundwater resource, and increasing woody plant density decoupled NEP and ET from incident precipitation (P), resulting in greater exchange rates that were less variable year to year. Despite similar gross patterns, how groundwater accessibility affected NEP was more complex than ET. The grassland had higher respiration (Reco ) costs. Thus, while it had similar ET and gross carbon uptake (GEP) to the shrubland, grassland NEP was substantially less. Also, grassland carbon fluxes were more variable due to occasional flooding at the site, which both stimulated and inhibited NEP depending upon phenology. Woodland NEP was large, but surprisingly similar to the less mature, sparse shrubland, even while having much greater GEP. Woodland Reco was greater than the shrubland and responded strongly and positively to P, which resulted in a surprising negative NEP response to P. This is likely due to the large accumulation of carbon aboveground and in the surface soil. These long-term observations support the strong role that water accessibility can play when determining the consequences of ecosystem vegetation change. PMID:24777485
NASA Astrophysics Data System (ADS)
Boulet, G.; Mougenot, B.; Lhomme, J.-P.; Fanise, P.; Lili-Chabaane, Z.; Olioso, A.; Bahir, M.; Rivalland, V.; Jarlan, L.; Merlin, O.; Coudert, B.; Er-Raki, S.; Lagouarde, J.-P.
2015-11-01
Evapotranspiration is an important component of the water cycle, especially in semi-arid lands. A way to quantify the spatial distribution of evapotranspiration and water stress from remote-sensing data is to exploit the available surface temperature as a signature of the surface energy balance. Remotely sensed energy balance models enable one to estimate stress levels and, in turn, the water status of continental surfaces. Dual-source models are particularly useful since they allow derivation of a rough estimate of the water stress of the vegetation instead of that of a soil-vegetation composite. They either assume that the soil and the vegetation interact almost independently with the atmosphere (patch approach corresponding to a parallel resistance scheme) or are tightly coupled (layer approach corresponding to a series resistance scheme). The water status of both sources is solved simultaneously from a single surface temperature observation based on a realistic underlying assumption which states that, in most cases, the vegetation is unstressed, and that if the vegetation is stressed, evaporation is negligible. In the latter case, if the vegetation stress is not properly accounted for, the resulting evaporation will decrease to unrealistic levels (negative fluxes) in order to maintain the same total surface temperature. This work assesses the retrieval performances of total and component evapotranspiration as well as surface and plant water stress levels by (1) proposing a new dual-source model named Soil Plant Atmosphere and Remote Sensing Evapotranspiration (SPARSE) in two versions (parallel and series resistance networks) based on the TSEB (Two-Source Energy Balance model, Norman et al., 1995) model rationale as well as state-of-the-art formulations of turbulent and radiative exchange, (2) challenging the limits of the underlying hypothesis for those two versions through a synthetic retrieval test and (3) testing the water stress retrievals (vegetation water
Sparse matrix transform for fast projection to reduced dimension
Theiler, James P; Cao, Guangzhi; Bouman, Charles A
2010-01-01
We investigate three algorithms that use the sparse matrix transform (SMT) to produce variance-maximizing linear projections to a lower-dimensional space. The SMT expresses the projection as a sequence of Givens rotations and this enables computationally efficient implementation of the projection operator. The baseline algorithm uses the SMT to directly approximate the optimal solution that is given by principal components analysis (PCA). A variant of the baseline begins with a standard SMT solution, but prunes the sequence of Givens rotations to only include those that contribute to the variance maximization. Finally, a simpler and faster third algorithm is introduced; this also estimates the projection operator with a sequence of Givens rotations, but in this case, the rotations are chosen to optimize a criterion that more directly expresses the dimension reduction criterion.
Measuring sparseness in the brain: Comment on Bowers (2009)
Quiroga, R. Quian; Kreiman, G.
2011-01-01
Bowers (2009) challenged the common view in favor of distributed representations in psychological modeling and the main arguments given against localist and grandmother cell coding schemes. He revisited the results of several single-cell studies arguing that they do not support distributed representations. We praise the contribution of Bowers for joining evidence from psychological modeling and neurophysiological recordings, but disagree with several of his claims. In this comment we argue that distinctions between distributed, localist and grandmother cell coding can be troublesome with real data. Moreover, these distinctions seem to be lying within the same continuum, and we argue that it may be sensible to characterize coding schemes using a sparseness measure. We further argue that there may not be a unique coding scheme implemented in all brain areas and for all possible functions. In particular, current evidence suggests that the brain may use distributed codes in primary sensory areas and sparser and invariant representations in higher areas. PMID:20063978
Maximizing sparse matrix vector product performance in MIMD computers
McLay, R.T.; Kohli, H.S.; Swift, S.L.; Carey, G.F.
1994-12-31
A considerable component of the computational effort involved in conjugate gradient solution of structured sparse matrix systems is expended during the Matrix-Vector Product (MVP), and hence it is the focus of most efforts at improving performance. Such efforts are hindered on MIMD machines due to constraints on memory, cache and speed of memory-cpu data transfer. This paper describes a strategy for maximizing the performance of the local computations associated with the MVP. The method focuses on single stride memory access, and the efficient use of cache by pre-loading it with data that is re-used while bypassing it for other data. The algorithm is designed to behave optimally for varying grid sizes and number of unknowns per gridpoint. Results from an assembly language implementation of the strategy on the iPSC/860 show a significant improvement over the performance using FORTRAN.
Inferring propagation paths for sparsely observed perturbations on complex networks
Massucci, Francesco Alessandro; Wheeler, Jonathan; Beltrán-Debón, Raúl; Joven, Jorge; Sales-Pardo, Marta; Guimerà, Roger
2016-01-01
In a complex system, perturbations propagate by following paths on the network of interactions among the system’s units. In contrast to what happens with the spreading of epidemics, observations of general perturbations are often very sparse in time (there is a single observation of the perturbed system) and in “space” (only a few perturbed and unperturbed units are observed). A major challenge in many areas, from biology to the social sciences, is to infer the propagation paths from observations of the effects of perturbation under these sparsity conditions. We address this problem and show that it is possible to go beyond the usual approach of using the shortest paths connecting the known perturbed nodes. Specifically, we show that a simple and general probabilistic model, which we solved using belief propagation, provides fast and accurate estimates of the probabilities of nodes being perturbed.
Constraints on Fluctuations in Sparsely Characterized Biological Systems
NASA Astrophysics Data System (ADS)
Hilfinger, Andreas; Norman, Thomas M.; Vinnicombe, Glenn; Paulsson, Johan
2016-02-01
Biochemical processes are inherently stochastic, creating molecular fluctuations in otherwise identical cells. Such "noise" is widespread but has proven difficult to analyze because most systems are sparsely characterized at the single cell level and because nonlinear stochastic models are analytically intractable. Here, we exactly relate average abundances, lifetimes, step sizes, and covariances for any pair of components in complex stochastic reaction systems even when the dynamics of other components are left unspecified. Using basic mathematical inequalities, we then establish bounds for whole classes of systems. These bounds highlight fundamental trade-offs that show how efficient assembly processes must invariably exhibit large fluctuations in subunit levels and how eliminating fluctuations in one cellular component requires creating heterogeneity in another.
Multi-frame image super resolution based on sparse coding.
Kato, Toshiyuki; Hino, Hideitsu; Murata, Noboru
2015-06-01
An image super-resolution method from multiple observation of low-resolution images is proposed. The method is based on sub-pixel accuracy block matching for estimating relative displacements of observed images, and sparse signal representation for estimating the corresponding high-resolution image, where correspondence between high- and low-resolution images are modeled by a certain degradation process. Relative displacements of small patches of observed low-resolution images are accurately estimated by a computationally efficient block matching method. The matching scores of the block matching are used to select a subset of low-resolution patches for reconstructing a high-resolution patch, that is, an adaptive selection of informative low-resolution images is realized. The proposed method is shown to perform comparable or superior to conventional super-resolution methods through experiments using various images.
Constraints on Fluctuations in Sparsely Characterized Biological Systems
Hilfinger, Andreas; Norman, Thomas M.; Vinnicombe, Glenn
2016-01-01
Biochemical processes are inherently stochastic, creating molecular fluctuations in otherwise identical cells. Such “noise” is widespread but has proven difficult to analyze because most systems are sparsely characterized at the single cell level and because nonlinear stochastic models are analytically intractable. Here, we exactly relate average abundances, lifetimes, step sizes, and covariances for any pair of components in complex stochastic reaction systems even when the dynamics of other components are left unspecified. Using basic mathematical inequalities, we then establish bounds for whole classes of systems. These bounds highlight fundamental trade-offs that show how efficient assembly processes must invariably exhibit large fluctuations in subunit levels and how eliminating fluctuations in one cellular component requires creating heterogeneity in another. PMID:26894735
A Spectral Algorithm for Envelope Reduction of Sparse Matrices
NASA Technical Reports Server (NTRS)
Barnard, Stephen T.; Pothen, Alex; Simon, Horst D.
1993-01-01
The problem of reordering a sparse symmetric matrix to reduce its envelope size is considered. A new spectral algorithm for computing an envelope-reducing reordering is obtained by associating a Laplacian matrix with the given matrix and then sorting the components of a specified eigenvector of the Laplacian. This Laplacian eigenvector solves a continuous relaxation of a discrete problem related to envelope minimization called the minimum 2-sum problem. The permutation vector computed by the spectral algorithm is a closest permutation vector to the specified Laplacian eigenvector. Numerical results show that the new reordering algorithm usually computes smaller envelope sizes than those obtained from the current standard algorithms such as Gibbs-Poole-Stockmeyer (GPS) or SPARSPAK reverse Cuthill-McKee (RCM), in some cases reducing the envelope by more than a factor of two.
Sparse Parallel MRI Based on Accelerated Operator Splitting Schemes
Xie, Weisi; Su, Zhenghang
2016-01-01
Recently, the sparsity which is implicit in MR images has been successfully exploited for fast MR imaging with incomplete acquisitions. In this paper, two novel algorithms are proposed to solve the sparse parallel MR imaging problem, which consists of l1 regularization and fidelity terms. The two algorithms combine forward-backward operator splitting and Barzilai-Borwein schemes. Theoretically, the presented algorithms overcome the nondifferentiable property in l1 regularization term. Meanwhile, they are able to treat a general matrix operator that may not be diagonalized by fast Fourier transform and to ensure that a well-conditioned optimization system of equations is simply solved. In addition, we build connections between the proposed algorithms and the state-of-the-art existing methods and prove their convergence with a constant stepsize in Appendix. Numerical results and comparisons with the advanced methods demonstrate the efficiency of proposed algorithms. PMID:27746824
Profile analysis: The effects of rove on sparse spectra
NASA Astrophysics Data System (ADS)
Lentz, Jennifer J.
2005-11-01
Profile-analysis thresholds were measured in the presence and absence of overall level variation at different stimulus levels to determine whether nonlinear changes in the shape of the edges of excitation pattern peaks contribute to poorer spectral-shape sensitivity observed under roving levels. Roving levels decreased sensitivity for stimuli having few components separated widely in frequency to a greater extent than for stimuli having more densely spaced components. The stimulus level did not influence sensitivity when overall level variation was absent, suggesting that listeners rely on peaks in the excitation patterns and not the edges of the peaks (as would have been predicted by the near miss to Weber's law). Because the edges of the peaks were not used in the absence of roving levels, it follows that the larger rove effects for sparse stimuli were not likely due to excitation pattern inconstancy.
A Population Shift between Sparsely Populated Folding Intermediates Determines Amyloidogenicity
2016-01-01
The balance between protein folding and misfolding is a crucial determinant of amyloid assembly. Transient intermediates that are sparsely populated during protein folding have been identified as key players in amyloid aggregation. However, due to their ephemeral nature, structural characterization of these species remains challenging. Here, using the power of nonuniformly sampled NMR methods we investigate the folding pathway of amyloidogenic and nonamyloidogenic variants of β2-microglobulin (β2m) in atomic detail. Despite folding via common intermediate states, we show that the decreased population of the aggregation-prone ITrans state and population of a less stable, more dynamic species ablate amyloid formation by increasing the energy barrier for amyloid assembly. The results show that subtle changes in conformational dynamics can have a dramatic effect in determining whether a protein is amyloidogenic, without perturbation of the mechanism of protein folding. PMID:27117876
Sparse activity of identified dentate granule cells during spatial exploration
Diamantaki, Maria; Frey, Markus; Berens, Philipp; Preston-Ferrer, Patricia; Burgalossi, Andrea
2016-01-01
In the dentate gyrus – a key component of spatial memory circuits – granule cells (GCs) are known to be morphologically diverse and to display heterogeneous activity profiles during behavior. To resolve structure–function relationships, we juxtacellularly recorded and labeled single GCs in freely moving rats. We found that the vast majority of neurons were silent during exploration. Most active GCs displayed a characteristic spike waveform, fired at low rates and showed spatial activity. Primary dendritic parameters were sufficient for classifying neurons as active or silent with high accuracy. Our data thus support a sparse coding scheme in the dentate gyrus and provide a possible link between structural and functional heterogeneity among the GC population. DOI: http://dx.doi.org/10.7554/eLife.20252.001 PMID:27692065
Solving sparse triangular linear systems on parallel computers
NASA Technical Reports Server (NTRS)
Anderson, Edward; Saad, Youcef
1989-01-01
This paper describes and compares three parallel algorithms for solving sparse triangular systems of equations. These methods involve some preprocessing overhead and are primarily of interest in solving many systems with the same coefficient matrix. The first approach is to use a fixed blocksize and form the inverse of the diagonal blocks. The second approach is to use a variable blocksize and reorder the unknowns so that the diagonal blocks are diagonal matrices. The latter technique is called level scheduling because of how it is represented in the adjacency graph, and both row-wise and jagged diagonal storage for the off-diagonal blocks are considered. These techniques are analyzed for general parallel computers and experiments are presented for the eight-processor Alliant FX/8.
Dense and Sparse Matrix Operations on the Cell Processor
Williams, Samuel W.; Shalf, John; Oliker, Leonid; Husbands,Parry; Yelick, Katherine
2005-05-01
The slowing pace of commodity microprocessor performance improvements combined with ever-increasing chip power demands has become of utmost concern to computational scientists. Therefore, the high performance computing community is examining alternative architectures that address the limitations of modern superscalar designs. In this work, we examine STI's forthcoming Cell processor: a novel, low-power architecture that combines a PowerPC core with eight independent SIMD processing units coupled with a software-controlled memory to offer high FLOP/s/Watt. Since neither Cell hardware nor cycle-accurate simulators are currently publicly available, we develop an analytic framework to predict Cell performance on dense and sparse matrix operations, using a variety of algorithmic approaches. Results demonstrate Cell's potential to deliver more than an order of magnitude better GFLOP/s per watt performance, when compared with the Intel Itanium2 and Cray X1 processors.
Sparse Zero-Sum Games as Stable Functional Feature Selection
Sokolovska, Nataliya; Teytaud, Olivier; Rizkalla, Salwa; Clément, Karine; Zucker, Jean-Daniel
2015-01-01
In large-scale systems biology applications, features are structured in hidden functional categories whose predictive power is identical. Feature selection, therefore, can lead not only to a problem with a reduced dimensionality, but also reveal some knowledge on functional classes of variables. In this contribution, we propose a framework based on a sparse zero-sum game which performs a stable functional feature selection. In particular, the approach is based on feature subsets ranking by a thresholding stochastic bandit. We provide a theoretical analysis of the introduced algorithm. We illustrate by experiments on both synthetic and real complex data that the proposed method is competitive from the predictive and stability viewpoints. PMID:26325268
Sparse Zero-Sum Games as Stable Functional Feature Selection.
Sokolovska, Nataliya; Teytaud, Olivier; Rizkalla, Salwa; Clément, Karine; Zucker, Jean-Daniel
2015-01-01
In large-scale systems biology applications, features are structured in hidden functional categories whose predictive power is identical. Feature selection, therefore, can lead not only to a problem with a reduced dimensionality, but also reveal some knowledge on functional classes of variables. In this contribution, we propose a framework based on a sparse zero-sum game which performs a stable functional feature selection. In particular, the approach is based on feature subsets ranking by a thresholding stochastic bandit. We provide a theoretical analysis of the introduced algorithm. We illustrate by experiments on both synthetic and real complex data that the proposed method is competitive from the predictive and stability viewpoints. PMID:26325268
Multiple peaks of species abundance distributions induced by sparse interactions.
Obuchi, Tomoyuki; Kabashima, Yoshiyuki; Tokita, Kei
2016-08-01
We investigate the replicator dynamics with "sparse" symmetric interactions which represent specialist-specialist interactions in ecological communities. By considering a large self-interaction u, we conduct a perturbative expansion which manifests that the nature of the interactions has a direct impact on the species abundance distribution. The central results are all species coexistence in a realistic range of the model parameters and that a certain discrete nature of the interactions induces multiple peaks in the species abundance distribution, providing the possibility of theoretically explaining multiple peaks observed in various field studies. To get more quantitative information, we also construct a non-perturbative theory which becomes exact on tree-like networks if all the species coexist, providing exact critical values of u below which extinct species emerge. Numerical simulations in various different situations are conducted and they clarify the robustness of the presented mechanism of all species coexistence and multiple peaks in the species abundance distributions. PMID:27627322
Sparse Bayesian Learning for DOA Estimation with Mutual Coupling
Dai, Jisheng; Hu, Nan; Xu, Weichao; Chang, Chunqi
2015-01-01
Sparse Bayesian learning (SBL) has given renewed interest to the problem of direction-of-arrival (DOA) estimation. It is generally assumed that the measurement matrix in SBL is precisely known. Unfortunately, this assumption may be invalid in practice due to the imperfect manifold caused by unknown or misspecified mutual coupling. This paper describes a modified SBL method for joint estimation of DOAs and mutual coupling coefficients with uniform linear arrays (ULAs). Unlike the existing method that only uses stationary priors, our new approach utilizes a hierarchical form of the Student t prior to enforce the sparsity of the unknown signal more heavily. We also provide a distinct Bayesian inference for the expectation-maximization (EM) algorithm, which can update the mutual coupling coefficients more efficiently. Another difference is that our method uses an additional singular value decomposition (SVD) to reduce the computational complexity of the signal reconstruction process and the sensitivity to the measurement noise. PMID:26501284
W-band sparse synthetic aperture for computational imaging.
Venkatesh, S; Viswanathan, N; Schurig, D
2016-04-18
We present a sparse synthetic-aperture, active imaging system at W-band (75 - 110 GHz), which uses sub-harmonic mixer modules. The system employs mechanical scanning of the receiver module position, and a fixed transmitter module. A vector network analyzer provides the back end detection. A full-wave forward model allows accurate construction of the image transfer matrix. We solve the inverse problem to reconstruct scenes using the least squares technique. We demonstrate far-field, diffraction limited imaging of 2D and 3D objects and achieve a cross-range resolution of 3 mm and a depth-range resolution of 4 mm, respectively. Furthermore, we develop an information-based metric to evaluate the performance of a given image transfer matrix for noise-limited, computational imaging systems. We use this metric to find the optimal gain of the radiating element for a given range, both theoretically and experimentally in our system. PMID:27137270
Automated identification of crystallographic ligands using sparse-density representations
Carolan, C. G.; Lamzin, V. S.
2014-07-01
A novel procedure for identifying ligands in macromolecular crystallographic electron-density maps is introduced. Density clusters in such maps can be rapidly attributed to one of 82 different ligands in an automated manner. A novel procedure for the automatic identification of ligands in macromolecular crystallographic electron-density maps is introduced. It is based on the sparse parameterization of density clusters and the matching of the pseudo-atomic grids thus created to conformationally variant ligands using mathematical descriptors of molecular shape, size and topology. In large-scale tests on experimental data derived from the Protein Data Bank, the procedure could quickly identify the deposited ligand within the top-ranked compounds from a database of candidates. This indicates the suitability of the method for the identification of binding entities in fragment-based drug screening and in model completion in macromolecular structure determination.
Mandala Networks: ultra-small-world and highly sparse graphs
Sampaio Filho, Cesar I. N.; Moreira, André A.; Andrade, Roberto F. S.; Herrmann, Hans J.; Andrade, José S.
2015-01-01
The increasing demands in security and reliability of infrastructures call for the optimal design of their embedded complex networks topologies. The following question then arises: what is the optimal layout to fulfill best all the demands? Here we present a general solution for this problem with scale-free networks, like the Internet and airline networks. Precisely, we disclose a way to systematically construct networks which are robust against random failures. Furthermore, as the size of the network increases, its shortest path becomes asymptotically invariant and the density of links goes to zero, making it ultra-small world and highly sparse, respectively. The first property is ideal for communication and navigation purposes, while the second is interesting economically. Finally, we show that some simple changes on the original network formulation can lead to an improved topology against malicious attacks. PMID:25765450
Mandala networks: ultra-small-world and highly sparse graphs.
Sampaio Filho, Cesar I N; Moreira, André A; Andrade, Roberto F S; Herrmann, Hans J; Andrade, José S
2015-03-13
The increasing demands in security and reliability of infrastructures call for the optimal design of their embedded complex networks topologies. The following question then arises: what is the optimal layout to fulfill best all the demands? Here we present a general solution for this problem with scale-free networks, like the Internet and airline networks. Precisely, we disclose a way to systematically construct networks which are robust against random failures. Furthermore, as the size of the network increases, its shortest path becomes asymptotically invariant and the density of links goes to zero, making it ultra-small world and highly sparse, respectively. The first property is ideal for communication and navigation purposes, while the second is interesting economically. Finally, we show that some simple changes on the original network formulation can lead to an improved topology against malicious attacks.
Mandala Networks: ultra-small-world and highly sparse graphs
NASA Astrophysics Data System (ADS)
Sampaio Filho, Cesar I. N.; Moreira, André A.; Andrade, Roberto F. S.; Herrmann, Hans J.; Andrade, José S.
2015-03-01
The increasing demands in security and reliability of infrastructures call for the optimal design of their embedded complex networks topologies. The following question then arises: what is the optimal layout to fulfill best all the demands? Here we present a general solution for this problem with scale-free networks, like the Internet and airline networks. Precisely, we disclose a way to systematically construct networks which are robust against random failures. Furthermore, as the size of the network increases, its shortest path becomes asymptotically invariant and the density of links goes to zero, making it ultra-small world and highly sparse, respectively. The first property is ideal for communication and navigation purposes, while the second is interesting economically. Finally, we show that some simple changes on the original network formulation can lead to an improved topology against malicious attacks.
An alternative design for a sparse distributed memory
NASA Technical Reports Server (NTRS)
Jaeckel, Louis A.
1989-01-01
A new design for a Sparse Distributed Memory, called the selected-coordinate design, is described. As in the original design, there are a large number of memory locations, each of which may be activated by many different addresses (binary vectors) in a very large address space. Each memory location is defined by specifying ten selected coordinates (bit positions in the address vectors) and a set of corresponding assigned values, consisting of one bit for each selected coordinate. A memory location is activated by an address if, for all ten of the locations's selected coordinates, the corresponding bits in the address vector match the respective assigned value bits, regardless of the other bits in the address vector. Some comparative memory capacity and signal-to-noise ratio estimates for the both the new and original designs are given. A few possible hardware embodiments of the new design are described.
Multiple peaks of species abundance distributions induced by sparse interactions
NASA Astrophysics Data System (ADS)
Obuchi, Tomoyuki; Kabashima, Yoshiyuki; Tokita, Kei
2016-08-01
We investigate the replicator dynamics with "sparse" symmetric interactions which represent specialist-specialist interactions in ecological communities. By considering a large self-interaction u , we conduct a perturbative expansion which manifests that the nature of the interactions has a direct impact on the species abundance distribution. The central results are all species coexistence in a realistic range of the model parameters and that a certain discrete nature of the interactions induces multiple peaks in the species abundance distribution, providing the possibility of theoretically explaining multiple peaks observed in various field studies. To get more quantitative information, we also construct a non-perturbative theory which becomes exact on tree-like networks if all the species coexist, providing exact critical values of u below which extinct species emerge. Numerical simulations in various different situations are conducted and they clarify the robustness of the presented mechanism of all species coexistence and multiple peaks in the species abundance distributions.
Sparse Bayesian learning for DOA estimation with mutual coupling.
Dai, Jisheng; Hu, Nan; Xu, Weichao; Chang, Chunqi
2015-01-01
Sparse Bayesian learning (SBL) has given renewed interest to the problem of direction-of-arrival (DOA) estimation. It is generally assumed that the measurement matrix in SBL is precisely known. Unfortunately, this assumption may be invalid in practice due to the imperfect manifold caused by unknown or misspecified mutual coupling. This paper describes a modified SBL method for joint estimation of DOAs and mutual coupling coefficients with uniform linear arrays (ULAs). Unlike the existing method that only uses stationary priors, our new approach utilizes a hierarchical form of the Student t prior to enforce the sparsity of the unknown signal more heavily. We also provide a distinct Bayesian inference for the expectation-maximization (EM) algorithm, which can update the mutual coupling coefficients more efficiently. Another difference is that our method uses an additional singular value decomposition (SVD) to reduce the computational complexity of the signal reconstruction process and the sensitivity to the measurement noise. PMID:26501284
Doubly Sparse Relevance Vector Machine for Continuous Facial Behavior Estimation.
Kaltwang, Sebastian; Todorovic, Sinisa; Pantic, Maja
2016-09-01
Certain inner feelings and physiological states like pain are subjective states that cannot be directly measured, but can be estimated from spontaneous facial expressions. Since they are typically characterized by subtle movements of facial parts, analysis of the facial details is required. To this end, we formulate a new regression method for continuous estimation of the intensity of facial behavior interpretation, called Doubly Sparse Relevance Vector Machine (DSRVM). DSRVM enforces double sparsity by jointly selecting the most relevant training examples (a.k.a. relevance vectors) and the most important kernels associated with facial parts relevant for interpretation of observed facial expressions. This advances prior work on multi-kernel learning, where sparsity of relevant kernels is typically ignored. Empirical evaluation on challenging Shoulder Pain videos, and the benchmark DISFA and SEMAINE datasets demonstrate that DSRVM outperforms competing approaches with a multi-fold reduction of running times in training and testing. PMID:26595911
CARS Spectral Fitting with Multiple Resonant Species using Sparse Libraries
NASA Technical Reports Server (NTRS)
Cutler, Andrew D.; Magnotti, Gaetano
2010-01-01
The dual pump CARS technique is often used in the study of turbulent flames. Fast and accurate algorithms are needed for fitting dual-pump CARS spectra for temperature and multiple chemical species. This paper describes the development of such an algorithm. The algorithm employs sparse libraries, whose size grows much more slowly with number of species than a conventional library. The method was demonstrated by fitting synthetic "experimental" spectra containing 4 resonant species (N2, O2, H2 and CO2), both with noise and without it, and by fitting experimental spectra from a H2-air flame produced by a Hencken burner. In both studies, weighted least squares fitting of signal, as opposed to least squares fitting signal or square-root signal, was shown to produce the least random error and minimize bias error in the fitted parameters.
A bit allocation method for sparse source coding.
Kaaniche, Mounir; Fraysse, Aurélia; Pesquet-Popescu, Béatrice; Pesquet, Jean-Christophe
2014-01-01
In this paper, we develop an efficient bit allocation strategy for subband-based image coding systems. More specifically, our objective is to design a new optimization algorithm based on a rate-distortion optimality criterion. To this end, we consider the uniform scalar quantization of a class of mixed distributed sources following a Bernoulli-generalized Gaussian distribution. This model appears to be particularly well-adapted for image data, which have a sparse representation in a wavelet basis. In this paper, we propose new approximations of the entropy and the distortion functions using piecewise affine and exponential forms, respectively. Because of these approximations, bit allocation is reformulated as a convex optimization problem. Solving the resulting problem allows us to derive the optimal quantization step for each subband. Experimental results show the benefits that can be drawn from the proposed bit allocation method in a typical transform-based coding application.
Sparse labeling of proteins: Structural characterization from long range constraints
NASA Astrophysics Data System (ADS)
Prestegard, James H.; Agard, David A.; Moremen, Kelley W.; Lavery, Laura A.; Morris, Laura C.; Pederson, Kari
2014-04-01
Structural characterization of biologically important proteins faces many challenges associated with degradation of resolution as molecular size increases and loss of resolution improving tools such as perdeuteration when non-bacterial hosts must be used for expression. In these cases, sparse isotopic labeling (single or small subsets of amino acids) combined with long range paramagnetic constraints and improved computational modeling offer an alternative. This perspective provides a brief overview of this approach and two discussions of potential applications; one involving a very large system (an Hsp90 homolog) in which perdeuteration is possible and methyl-TROSY sequences can potentially be used to improve resolution, and one involving ligand placement in a glycosylated protein where resolution is achieved by single amino acid labeling (the sialyltransferase, ST6Gal1). This is not intended as a comprehensive review, but as a discussion of future prospects that promise impact on important questions in the structural biology area.
SAMBA: Sparse Approximation of Moment-Based Arbitrary Polynomial Chaos
NASA Astrophysics Data System (ADS)
Ahlfeld, R.; Belkouchi, B.; Montomoli, F.
2016-09-01
A new arbitrary Polynomial Chaos (aPC) method is presented for moderately high-dimensional problems characterised by limited input data availability. The proposed methodology improves the algorithm of aPC and extends the method, that was previously only introduced as tensor product expansion, to moderately high-dimensional stochastic problems. The fundamental idea of aPC is to use the statistical moments of the input random variables to develop the polynomial chaos expansion. This approach provides the possibility to propagate continuous or discrete probability density functions and also histograms (data sets) as long as their moments exist, are finite and the determinant of the moment matrix is strictly positive. For cases with limited data availability, this approach avoids bias and fitting errors caused by wrong assumptions. In this work, an alternative way to calculate the aPC is suggested, which provides the optimal polynomials, Gaussian quadrature collocation points and weights from the moments using only a handful of matrix operations on the Hankel matrix of moments. It can therefore be implemented without requiring prior knowledge about statistical data analysis or a detailed understanding of the mathematics of polynomial chaos expansions. The extension to more input variables suggested in this work, is an anisotropic and adaptive version of Smolyak's algorithm that is solely based on the moments of the input probability distributions. It is referred to as SAMBA (PC), which is short for Sparse Approximation of Moment-Based Arbitrary Polynomial Chaos. It is illustrated that for moderately high-dimensional problems (up to 20 different input variables or histograms) SAMBA can significantly simplify the calculation of sparse Gaussian quadrature rules. SAMBA's efficiency for multivariate functions with regard to data availability is further demonstrated by analysing higher order convergence and accuracy for a set of nonlinear test functions with 2, 5 and 10
Auditory Sketches: Very Sparse Representations of Sounds Are Still Recognizable.
Isnard, Vincent; Taffou, Marine; Viaud-Delmon, Isabelle; Suied, Clara
2016-01-01
Sounds in our environment like voices, animal calls or musical instruments are easily recognized by human listeners. Understanding the key features underlying this robust sound recognition is an important question in auditory science. Here, we studied the recognition by human listeners of new classes of sounds: acoustic and auditory sketches, sounds that are severely impoverished but still recognizable. Starting from a time-frequency representation, a sketch is obtained by keeping only sparse elements of the original signal, here, by means of a simple peak-picking algorithm. Two time-frequency representations were compared: a biologically grounded one, the auditory spectrogram, which simulates peripheral auditory filtering, and a simple acoustic spectrogram, based on a Fourier transform. Three degrees of sparsity were also investigated. Listeners were asked to recognize the category to which a sketch sound belongs: singing voices, bird calls, musical instruments, and vehicle engine noises. Results showed that, with the exception of voice sounds, very sparse representations of sounds (10 features, or energy peaks, per second) could be recognized above chance. No clear differences could be observed between the acoustic and the auditory sketches. For the voice sounds, however, a completely different pattern of results emerged, with at-chance or even below-chance recognition performances, suggesting that the important features of the voice, whatever they are, were removed by the sketch process. Overall, these perceptual results were well correlated with a model of auditory distances, based on spectro-temporal excitation patterns (STEPs). This study confirms the potential of these new classes of sounds, acoustic and auditory sketches, to study sound recognition.
Robust ellipse fitting based on sparse combination of data points.
Liang, Junli; Zhang, Miaohua; Liu, Ding; Zeng, Xianju; Ojowu, Ode; Zhao, Kexin; Li, Zhan; Liu, Han
2013-06-01
Ellipse fitting is widely applied in the fields of computer vision and automatic industry control, in which the procedure of ellipse fitting often follows the preprocessing step of edge detection in the original image. Therefore, the ellipse fitting method also depends on the accuracy of edge detection besides their own performance, especially due to the introduced outliers and edge point errors from edge detection which will cause severe performance degradation. In this paper, we develop a robust ellipse fitting method to alleviate the influence of outliers. The proposed algorithm solves ellipse parameters by linearly combining a subset of ("more accurate") data points (formed from edge points) rather than all data points (which contain possible outliers). In addition, considering that squaring the fitting residuals can magnify the contributions of these extreme data points, our algorithm replaces it with the absolute residuals to reduce this influence. Moreover, the norm of data point errors is bounded, and the worst case performance optimization is formed to be robust against data point errors. The resulting mixed l1-l2 optimization problem is further derived as a second-order cone programming one and solved by the computationally efficient interior-point methods. Note that the fitting approach developed in this paper specifically deals with the overdetermined system, whereas the current sparse representation theory is only applied to underdetermined systems. Therefore, the proposed algorithm can be looked upon as an extended application and development of the sparse representation theory. Some simulated and experimental examples are presented to illustrate the effectiveness of the proposed ellipse fitting approach. PMID:23412616
3D ear identification based on sparse representation.
Zhang, Lin; Ding, Zhixuan; Li, Hongyu; Shen, Ying
2014-01-01
Biometrics based personal authentication is an effective way for automatically recognizing, with a high confidence, a person's identity. Recently, 3D ear shape has attracted tremendous interests in research field due to its richness of feature and ease of acquisition. However, the existing ICP (Iterative Closet Point)-based 3D ear matching methods prevalent in the literature are not quite efficient to cope with the one-to-many identification case. In this paper, we aim to fill this gap by proposing a novel effective fully automatic 3D ear identification system. We at first propose an accurate and efficient template-based ear detection method. By utilizing such a method, the extracted ear regions are represented in a common canonical coordinate system determined by the ear contour template, which facilitates much the following stages of feature extraction and classification. For each extracted 3D ear, a feature vector is generated as its representation by making use of a PCA-based local feature descriptor. At the stage of classification, we resort to the sparse representation based classification approach, which actually solves an l1-minimization problem. To the best of our knowledge, this is the first work introducing the sparse representation framework into the field of 3D ear identification. Extensive experiments conducted on a benchmark dataset corroborate the effectiveness and efficiency of the proposed approach. The associated Matlab source code and the evaluation results have been made publicly online available at http://sse.tongji.edu.cn/linzhang/ear/srcear/srcear.htm. PMID:24740247
Robust ellipse fitting based on sparse combination of data points.
Liang, Junli; Zhang, Miaohua; Liu, Ding; Zeng, Xianju; Ojowu, Ode; Zhao, Kexin; Li, Zhan; Liu, Han
2013-06-01
Ellipse fitting is widely applied in the fields of computer vision and automatic industry control, in which the procedure of ellipse fitting often follows the preprocessing step of edge detection in the original image. Therefore, the ellipse fitting method also depends on the accuracy of edge detection besides their own performance, especially due to the introduced outliers and edge point errors from edge detection which will cause severe performance degradation. In this paper, we develop a robust ellipse fitting method to alleviate the influence of outliers. The proposed algorithm solves ellipse parameters by linearly combining a subset of ("more accurate") data points (formed from edge points) rather than all data points (which contain possible outliers). In addition, considering that squaring the fitting residuals can magnify the contributions of these extreme data points, our algorithm replaces it with the absolute residuals to reduce this influence. Moreover, the norm of data point errors is bounded, and the worst case performance optimization is formed to be robust against data point errors. The resulting mixed l1-l2 optimization problem is further derived as a second-order cone programming one and solved by the computationally efficient interior-point methods. Note that the fitting approach developed in this paper specifically deals with the overdetermined system, whereas the current sparse representation theory is only applied to underdetermined systems. Therefore, the proposed algorithm can be looked upon as an extended application and development of the sparse representation theory. Some simulated and experimental examples are presented to illustrate the effectiveness of the proposed ellipse fitting approach.
Sinogram denoising via simultaneous sparse representation in learned dictionaries
NASA Astrophysics Data System (ADS)
Karimi, Davood; Ward, Rabab K.
2016-05-01
Reducing the radiation dose in computed tomography (CT) is highly desirable but it leads to excessive noise in the projection measurements. This can significantly reduce the diagnostic value of the reconstructed images. Removing the noise in the projection measurements is, therefore, essential for reconstructing high-quality images, especially in low-dose CT. In recent years, two new classes of patch-based denoising algorithms proved superior to other methods in various denoising applications. The first class is based on sparse representation of image patches in a learned dictionary. The second class is based on the non-local means method. Here, the image is searched for similar patches and the patches are processed together to find their denoised estimates. In this paper, we propose a novel denoising algorithm for cone-beam CT projections. The proposed method has similarities to both these algorithmic classes but is more effective and much faster. In order to exploit both the correlation between neighboring pixels within a projection and the correlation between pixels in neighboring projections, the proposed algorithm stacks noisy cone-beam projections together to form a 3D image and extracts small overlapping 3D blocks from this 3D image for processing. We propose a fast algorithm for clustering all extracted blocks. The central assumption in the proposed algorithm is that all blocks in a cluster have a joint-sparse representation in a well-designed dictionary. We describe algorithms for learning such a dictionary and for denoising a set of projections using this dictionary. We apply the proposed algorithm on simulated and real data and compare it with three other algorithms. Our results show that the proposed algorithm outperforms some of the best denoising algorithms, while also being much faster.
Auditory Sketches: Very Sparse Representations of Sounds Are Still Recognizable
Isnard, Vincent; Taffou, Marine; Viaud-Delmon, Isabelle; Suied, Clara
2016-01-01
Sounds in our environment like voices, animal calls or musical instruments are easily recognized by human listeners. Understanding the key features underlying this robust sound recognition is an important question in auditory science. Here, we studied the recognition by human listeners of new classes of sounds: acoustic and auditory sketches, sounds that are severely impoverished but still recognizable. Starting from a time-frequency representation, a sketch is obtained by keeping only sparse elements of the original signal, here, by means of a simple peak-picking algorithm. Two time-frequency representations were compared: a biologically grounded one, the auditory spectrogram, which simulates peripheral auditory filtering, and a simple acoustic spectrogram, based on a Fourier transform. Three degrees of sparsity were also investigated. Listeners were asked to recognize the category to which a sketch sound belongs: singing voices, bird calls, musical instruments, and vehicle engine noises. Results showed that, with the exception of voice sounds, very sparse representations of sounds (10 features, or energy peaks, per second) could be recognized above chance. No clear differences could be observed between the acoustic and the auditory sketches. For the voice sounds, however, a completely different pattern of results emerged, with at-chance or even below-chance recognition performances, suggesting that the important features of the voice, whatever they are, were removed by the sketch process. Overall, these perceptual results were well correlated with a model of auditory distances, based on spectro-temporal excitation patterns (STEPs). This study confirms the potential of these new classes of sounds, acoustic and auditory sketches, to study sound recognition. PMID:26950589
Sparse imaging of cortical electrical current densities via wavelet transforms
NASA Astrophysics Data System (ADS)
Liao, Ke; Zhu, Min; Ding, Lei; Valette, Sébastien; Zhang, Wenbo; Dickens, Deanna
2012-11-01
While the cerebral cortex in the human brain is of functional importance, functions defined on this structure are difficult to analyze spatially due to its highly convoluted irregular geometry. This study developed a novel L1-norm regularization method using a newly proposed multi-resolution face-based wavelet method to estimate cortical electrical activities in electroencephalography (EEG) and magnetoencephalography (MEG) inverse problems. The proposed wavelets were developed based on multi-resolution models built from irregular cortical surface meshes, which were realized in this study too. The multi-resolution wavelet analysis was used to seek sparse representation of cortical current densities in transformed domains, which was expected due to the compressibility of wavelets, and evaluated using Monte Carlo simulations. The EEG/MEG inverse problems were solved with the use of the novel L1-norm regularization method exploring the sparseness in the wavelet domain. The inverse solutions obtained from the new method using MEG data were evaluated by Monte Carlo simulations too. The present results indicated that cortical current densities could be efficiently compressed using the proposed face-based wavelet method, which exhibited better performance than the vertex-based wavelet method. In both simulations and auditory experimental data analysis, the proposed L1-norm regularization method showed better source detection accuracy and less estimation errors than other two classic methods, i.e. weighted minimum norm (wMNE) and cortical low-resolution electromagnetic tomography (cLORETA). This study suggests that the L1-norm regularization method with the use of face-based wavelets is a promising tool for studying functional activations of the human brain.
Adaptive sparse polynomial chaos expansion based on least angle regression
NASA Astrophysics Data System (ADS)
Blatman, Géraud; Sudret, Bruno
2011-03-01
Polynomial chaos (PC) expansions are used in stochastic finite element analysis to represent the random model response by a set of coefficients in a suitable (so-called polynomial chaos) basis. The number of terms to be computed grows dramatically with the size of the input random vector, which makes the computational cost of classical solution schemes (may it be intrusive (i.e. of Galerkin type) or non intrusive) unaffordable when the deterministic finite element model is expensive to evaluate. To address such problems, the paper describes a non intrusive method that builds a sparse PC expansion. First, an original strategy for truncating the PC expansions, based on hyperbolic index sets, is proposed. Then an adaptive algorithm based on least angle regression (LAR) is devised for automatically detecting the significant coefficients of the PC expansion. Beside the sparsity of the basis, the experimental design used at each step of the algorithm is systematically complemented in order to avoid the overfitting phenomenon. The accuracy of the PC metamodel is checked using an estimate inspired by statistical learning theory, namely the corrected leave-one-out error. As a consequence, a rather small number of PC terms are eventually retained ( sparse representation), which may be obtained at a reduced computational cost compared to the classical "full" PC approximation. The convergence of the algorithm is shown on an analytical function. Then the method is illustrated on three stochastic finite element problems. The first model features 10 input random variables, whereas the two others involve an input random field, which is discretized into 38 and 30 - 500 random variables, respectively.
Sparse Multivariate Autoregressive Modeling for Mild Cognitive Impairment Classification
Li, Yang; Wee, Chong-Yaw; Jie, Biao; Peng, Ziwen
2014-01-01
Brain connectivity network derived from functional magnetic resonance imaging (fMRI) is becoming increasingly prevalent in the researches related to cognitive and perceptual processes. The capability to detect causal or effective connectivity is highly desirable for understanding the cooperative nature of brain network, particularly when the ultimate goal is to obtain good performance of control-patient classification with biological meaningful interpretations. Understanding directed functional interactions between brain regions via brain connectivity network is a challenging task. Since many genetic and biomedical networks are intrinsically sparse, incorporating sparsity property into connectivity modeling can make the derived models more biologically plausible. Accordingly, we propose an effective connectivity modeling of resting-state fMRI data based on the multivariate autoregressive (MAR) modeling technique, which is widely used to characterize temporal information of dynamic systems. This MAR modeling technique allows for the identification of effective connectivity using the Granger causality concept and reducing the spurious causality connectivity in assessment of directed functional interaction from fMRI data. A forward orthogonal least squares (OLS) regression algorithm is further used to construct a sparse MAR model. By applying the proposed modeling to mild cognitive impairment (MCI) classification, we identify several most discriminative regions, including middle cingulate gyrus, posterior cingulate gyrus, lingual gyrus and caudate regions, in line with results reported in previous findings. A relatively high classification accuracy of 91.89 % is also achieved, with an increment of 5.4 % compared to the fully-connected, non-directional Pearson-correlation-based functional connectivity approach. PMID:24595922
3D Ear Identification Based on Sparse Representation
Zhang, Lin; Ding, Zhixuan; Li, Hongyu; Shen, Ying
2014-01-01
Biometrics based personal authentication is an effective way for automatically recognizing, with a high confidence, a person’s identity. Recently, 3D ear shape has attracted tremendous interests in research field due to its richness of feature and ease of acquisition. However, the existing ICP (Iterative Closet Point)-based 3D ear matching methods prevalent in the literature are not quite efficient to cope with the one-to-many identification case. In this paper, we aim to fill this gap by proposing a novel effective fully automatic 3D ear identification system. We at first propose an accurate and efficient template-based ear detection method. By utilizing such a method, the extracted ear regions are represented in a common canonical coordinate system determined by the ear contour template, which facilitates much the following stages of feature extraction and classification. For each extracted 3D ear, a feature vector is generated as its representation by making use of a PCA-based local feature descriptor. At the stage of classification, we resort to the sparse representation based classification approach, which actually solves an l1-minimization problem. To the best of our knowledge, this is the first work introducing the sparse representation framework into the field of 3D ear identification. Extensive experiments conducted on a benchmark dataset corroborate the effectiveness and efficiency of the proposed approach. The associated Matlab source code and the evaluation results have been made publicly online available at http://sse.tongji.edu.cn/linzhang/ear/srcear/srcear.htm. PMID:24740247
Cruciferous Vegetables and Cancer Prevention
... vegetables are part of the Brassica genus of plants. They include the following vegetables, among others: Arugula ... Jain MG, Hislop GT, Howe GR, Ghadirian P. Plant foods, antioxidants, and prostate cancer risk: findings from ...
Choose a variety of fruits and vegetables daily: understanding the complexities.
Krebs-Smith, S M; Kantor, L S
2001-02-01
The 2000 edition of Nutrition and Your Health: Dietary Guidelines for Americans is the first to include a recommendation aimed specifically at fruits and vegetables, apart from grains. This paper discusses these changes in the Dietary Guidelines, summarizes the methods of assessment pertaining to fruit and vegetable intakes and their related factors, and reviews the data available on current levels and trends over time. Recent methodological advances in the measurement of both the aggregate U.S. food supply and foods consumed by individuals have allowed for better estimates with which recommendations can be compared. The data on individual intakes suggest the following: Americans are consuming fruits and vegetables at a level near the minimum recommendations; to be in concordance with energy-based recommendations, they would have to consume approximately 2 more servings per day; and dark green and deep yellow vegetables are accounting for a disproportionately small share of the total. Fruit and vegetable consumption appears to be rising, but only slightly, and this increase might be only an artifact of shifts in the population demographics. A number of studies suggest that low income households in poor central cities and sparsely populated rural areas often have less access to food stores and face higher prices for food, including fruits and vegetables, compared with other households. At the aggregate level, supplying enough fruits and vegetables to meet dietary recommendations for all U.S. consumers would require adjustments in U.S. agricultural production, trade, marketing practices and prices of these commodities. PMID:11160580
Technology Transfer Automated Retrieval System (TEKTRAN)
A change detection experiment for an invasive species, saltcedar, near Lovelock, Nevada, was conducted with multi-date Compact Airborne Spectrographic Imager (CASI) hyperspectral datasets. Classification and NDVI differencing change detection methods were tested, In the classification strategy, a p...
Karaosmanoglu, F.
1999-04-01
Using vegetable oils as fuel alternatives has economic, environmental, and energy benefits for Turkey. The present work provides insight to the status of vegetable oil fuels in Turkey. A brief historical background of the issue, as well as an up to date review of the research carried out on vegetable oil fuels, is given and the future of their production and application is discussed.
Nonlinearities in vegetation functioning
NASA Astrophysics Data System (ADS)
Ceballos-Núñez, Verónika; Müller, Markus; Metzler, Holger; Sierra, Carlos
2016-04-01
Given the current drastic changes in climate and atmospheric CO2 concentrations, and the role of vegetation in the global carbon cycle, there is increasing attention to the carbon allocation component in biosphere terrestrial models. Improving the representation of C allocation in models could be the key to having better predictions of the fate of C once it enters the vegetation and is partitioned to C pools of different residence times. C allocation has often been modeled using systems of ordinary differential equations, and it has been hypothesized that most models can be generalized with a specific form of a linear dynamical system. However, several studies have highlighted discrepancies between empirical observations and model predictions, attributing these differences to problems with model structure. Although efforts have been made to compare different models, the outcome of these qualitative assessments has been a conceptual categorization of them. In this contribution, we introduce a new effort to identify the main properties of groups of models by studying their mathematical structure. For this purpose, we performed a literature research of the relevant models of carbon allocation in vegetation and developed a database with their representation in symbolic mathematics. We used the Python package SymPy for symbolic mathematics as a common language and manipulated the models to calculate their Jacobian matrix at fixed points and their eigenvalues, among other mathematical analyses. Our preliminary results show a tendency of inverse proportionality between model complexity and size of time/space scale; complex interactions between the variables controlling carbon allocation in vegetation tend to operate at shorter time/space scales, and vice-versa. Most importantly, we found that although the linear structure is common, other structures with non-linearities have been also proposed. We, therefore, propose a new General Model that can accommodate these
An algorithm for inverse synthetic aperture imaging lidar based on sparse signal representation
NASA Astrophysics Data System (ADS)
Ren, X. Z.; Sun, X. M.
2014-12-01
In actual applications of inverse synthetic aperture imaging lidar, the issue of sparse aperture data arises when continuous measurements are impossible or the collected data during some periods are not valid. Hence, the imaging results obtained by traditional methods are limited by high sidelobes. Considering the sparse structure of actual target space in high frequency radar application, a novel imaging method based on sparse signal representation is proposed in this paper. Firstly, the range image is acquired by traditional pulse compression of the optical heterodyne process. Then, the redundant dictionary is constructed through the sparse azimuth sampling positions and the signal form after the range compression. Finally, the imaging results are obtained by solving an ill-posed problem based on sparse regularization. Simulation results confirm the effectiveness of the proposed method.
Bayesian Learning in Sparse Graphical Factor Models via Variational Mean-Field Annealing
Yoshida, Ryo; West, Mike
2010-01-01
We describe a class of sparse latent factor models, called graphical factor models (GFMs), and relevant sparse learning algorithms for posterior mode estimation. Linear, Gaussian GFMs have sparse, orthogonal factor loadings matrices, that, in addition to sparsity of the implied covariance matrices, also induce conditional independence structures via zeros in the implied precision matrices. We describe the models and their use for robust estimation of sparse latent factor structure and data/signal reconstruction. We develop computational algorithms for model exploration and posterior mode search, addressing the hard combinatorial optimization involved in the search over a huge space of potential sparse configurations. A mean-field variational technique coupled with annealing is developed to successively generate “artificial” posterior distributions that, at the limiting temperature in the annealing schedule, define required posterior modes in the GFM parameter space. Several detailed empirical studies and comparisons to related approaches are discussed, including analyses of handwritten digit image and cancer gene expression data. PMID:20890391
NASA Astrophysics Data System (ADS)
Gao, Wei; Zhu, Linli; Wang, Kaiyun
2015-12-01
Ontology, a model of knowledge representation and storage, has had extensive applications in pharmaceutics, social science, chemistry and biology. In the age of “big data”, the constructed concepts are often represented as higher-dimensional data by scholars, and thus the sparse learning techniques are introduced into ontology algorithms. In this paper, based on the alternating direction augmented Lagrangian method, we present an ontology optimization algorithm for ontological sparse vector learning, and a fast version of such ontology technologies. The optimal sparse vector is obtained by an iterative procedure, and the ontology function is then obtained from the sparse vector. Four simulation experiments show that our ontological sparse vector learning model has a higher precision ratio on plant ontology, humanoid robotics ontology, biology ontology and physics education ontology data for similarity measuring and ontology mapping applications.
NASA Astrophysics Data System (ADS)
Fang, Jun; Zhang, Lizao; Duan, Huiping; Huang, Lei; Li, Hongbin
2016-05-01
The application of sparse representation to SAR/ISAR imaging has attracted much attention over the past few years. This new class of sparse representation based imaging methods present a number of unique advantages over conventional range-Doppler methods, the basic idea behind these works is to formulate SAR/ISAR imaging as a sparse signal recovery problem. In this paper, we propose a new two-dimensional pattern-coupled sparse Bayesian learning(SBL) method to capture the underlying cluster patterns of the ISAR target images. Based on this model, an expectation-maximization (EM) algorithm is developed to infer the maximum a posterior (MAP) estimate of the hyperparameters, along with the posterior distribution of the sparse signal. Experimental results demonstrate that the proposed method is able to achieve a substantial performance improvement over existing algorithms, including the conventional SBL method.
NASA Technical Reports Server (NTRS)
Rogers, David
1988-01-01
The advent of the Connection Machine profoundly changes the world of supercomputers. The highly nontraditional architecture makes possible the exploration of algorithms that were impractical for standard Von Neumann architectures. Sparse distributed memory (SDM) is an example of such an algorithm. Sparse distributed memory is a particularly simple and elegant formulation for an associative memory. The foundations for sparse distributed memory are described, and some simple examples of using the memory are presented. The relationship of sparse distributed memory to three important computational systems is shown: random-access memory, neural networks, and the cerebellum of the brain. Finally, the implementation of the algorithm for sparse distributed memory on the Connection Machine is discussed.
The red edge in arid region vegetation: 340-1060 nm spectra
NASA Technical Reports Server (NTRS)
Ray, Terrill W.; Murray, Bruce C.; Chehbouni, A.; Njoku, Eni
1993-01-01
The remote sensing study of vegetated regions of the world has typically been focused on the use of broad-band vegetation indices such as NDVI. Various modifications of these indices have been developed in attempts to minimize the effect of soil background, e.g., SAVI, or to reduce the effect of the atmosphere, e.g., ARVI. Most of these indices depend on the so-called 'red edge,' the sharp transition between the strong absorption of chlorophyll pigment in visible wavelengths and the strong scattering in the near-infrared from the cellular structure of leaves. These broadband indices tend to become highly inaccurate as the green canopy cover becomes sparse. The advent of high spectral resolution remote sensing instrument such as the Airborne Visible and Infrared Imaging Spectrometer (AVIRIS) has allowed the detection of narrow spectral features in vegetation and there are reports of detection of the red edge even for pixels with very low levels of green vegetation cover by Vane et al. and Elvidge et al., and to characterize algal biomass in coastal areas. Spectral mixing approaches similar to those of Smith et al. can be extended into the high spectral resolution domain allowing for the analysis of more endmembers, and potentially, discrimination between material with narrow spectral differences. Vegetation in arid regions tends to be sparse, often with small leaves such as the creosote bush. Many types of arid region vegetation spend much of the year with their leaves in a senescent state, i.e., yellow, with lowered chlorophyll pigmentation. The sparseness of the leaves of many arid region plants has the dual effect of lowering the green leaf area which can be observed and of allowing more of the sub-shrub soil to be visible which further complicates the spectrum of a region covered with arid region vegetation. Elvidge examined the spectral characteristics of dry plant materials showing significant differences in the region of the red edge and the diagnostic ligno
A MRI-CT prostate registration using sparse representation technique
NASA Astrophysics Data System (ADS)
Yang, Xiaofeng; Jani, Ashesh B.; Rossi, Peter J.; Mao, Hui; Curran, Walter J.; Liu, Tian
2016-03-01
Purpose: To develop a new MRI-CT prostate registration using patch-based deformation prediction framework to improve MRI-guided prostate radiotherapy by incorporating multiparametric MRI into planning CT images. Methods: The main contribution is to estimate the deformation between prostate MRI and CT images in a patch-wise fashion by using the sparse representation technique. We assume that two image patches should follow the same deformation if their patch-wise appearance patterns are similar. Specifically, there are two stages in our proposed framework, i.e., the training stage and the application stage. In the training stage, each prostate MR images are carefully registered to the corresponding CT images and all training MR and CT images are carefully registered to a selected CT template. Thus, we obtain the dense deformation field for each training MR and CT image. In the application stage, for registering a new subject MR image with the same subject CT image, we first select a small number of key points at the distinctive regions of this subject CT image. Then, for each key point in the subject CT image, we extract the image patch, centered at the underlying key point. Then, we adaptively construct the coupled dictionary for the underlying point where each atom in the dictionary consists of image patches and the respective deformations obtained from training pair-wise MRI-CT images. Next, the subject image patch can be sparsely represented by a linear combination of training image patches in the dictionary, where we apply the same sparse coefficients to the respective deformations in the dictionary to predict the deformation for the subject MR image patch. After we repeat the same procedure for each subject CT key point, we use B-splines to interpolate a dense deformation field, which is used as the initialization to allow the registration algorithm estimating the remaining small segment of deformations from MRI to CT image
Delinquency, Child Abuse, and Suggestions to Stomp Out Violence.
ERIC Educational Resources Information Center
Neely, Margery A.
Studies have traced a possible link between patterns of family interaction, faulty socialization, child abuse, and delinquent behavior. Counselors can respond to the needs of society to reduce violence by their access to the research on human development and in their work with families and children. Some types of discipline and parental attitudes…
Vegetation, plant biomass, and net primary productivity patterns in the Canadian Arctic
NASA Astrophysics Data System (ADS)
Gould, W. A.; Raynolds, M.; Walker, D. A.
2003-01-01
We have developed maps of dominant vegetation types, plant functional types, percent vegetation cover, aboveground plant biomass, and above and belowground annual net primary productivity for Canada north of the northern limit of trees. The area mapped covers 2.5 million km2 including glaciers. Ice-free land covers 2.3 million km2 and represents 42% of all ice-free land in the Circumpolar Arctic. The maps combine information on climate, soils, geology, hydrology, remotely sensed vegetation classifications, previous vegetation studies, and regional expertise to define polygons drawn using photo-interpretation of a 1:4,000,000 scale advanced very high resolution radiometer (AVHRR) color infrared image basemap. Polygons are linked to vegetation description, associated properties, and descriptive literature through a series of lookup tables in a graphic information systems (GIS) database developed as a component of the Circumpolar Arctic Vegetation Map (CAVM) project. Polygons are classified into 20 landcover types including 17 vegetation types. Half of the region is sparsely vegetated (<50% vegetation cover), primarily in the High Arctic (bioclimatic subzones A-C). Whereas most (86%) of the estimated aboveground plant biomass (1.5 × 1015 g) and 87% of the estimated above and belowground annual net primary productivity (2.28 × 1014 g yr-1) are concentrated in the Low Arctic (subzones D and E). The maps present more explicit spatial patterns of vegetation and ecosystem attributes than have been previously available, the GIS database is useful in summarizing ecosystem properties and can be easily updated and integrated into circumpolar mapping efforts, and the derived estimates fall within the range of current published estimates.
Decreased Surface Albedo Driven By Denser Vegetation On the Tibetan Plateau Yangjian Zhang*, Li Tian
NASA Astrophysics Data System (ADS)
Zhang, Y.; Tian, L.
2014-12-01
The Tibetan Plateau (TP) has fundamental ecological and environmental significance to China and Asia through its influence on regional and continental climates. In recent years, climate warming has caused unprecedented changes to land surface processes on the TP, which would unavoidably undermine the ecological and environmental functions of the TP. Among the numerous land surface processes potentially impacted by climate warming, the effect of vegetation greenness on surface energy balance is one of the most critical, but has been long ignored. In this study, we investigated the spatial and temporal patterns of land surface albedo (LSA) on the TP and evaluated the vegetation greenness in relation to patterns of LSA. We found that LSA has been decreasing in most of the vegetated grasslands on the TP from 2000 to 2013, as compared to a flat trend for desert area. The regions where LSA has been decreasing were spatially correlated to areas of increased vegetation greenness. Along rising altitude, LSA decreasing rate exhibited an overall decreasing trend. Across the TP, elevated vegetation greenness in grasslands acted as a primary factor pulling down LSA. The driving effects of vegetation greenness on LSA vary with grassland types, as revealed by a more significant relationship between vegetation greenness and LSA for the sparsely vegetated zone (i.e. steppe) than the more densely vegetated zone (i.e. meadow). Furthermore, the driving effect of vegetation greenness on LSA exhibited an obvious dependence on altitude as effects with rising altitude were relatively strong up to 3,000m, then weakened from 3,500m to 5,000m, and then the effects again increased from 5,000 to 6,000m. The growing season LSA trend revealed in this study emphasizes the need to give greater attention to the growing season LSA flux in future surface energy balance studies.
Decreased surface albedo driven by denser vegetation on the Tibetan Plateau
NASA Astrophysics Data System (ADS)
Tian, Li; Zhang, Yangjian; Zhu, Juntao
2014-10-01
The Tibetan Plateau (TP) has fundamental ecological and environmental significance to China and Asia through its influence on regional and continental climates. In recent years, climate warming has caused unprecedented changes to land surface processes on the TP, which would unavoidably undermine the ecological and environmental functions of the TP. Among the numerous land surface processes potentially impacted by climate warming, the effect of vegetation greenness on surface energy balance is one of the most critical, but has been long ignored. In this study, we investigated the spatial and temporal patterns of land surface albedo (LSA) on the TP and evaluated the vegetation greenness in relation to patterns of LSA. We found that LSA has been decreasing in most of the vegetated grasslands on the TP from 2000 to 2013, as compared to a flat trend for desert area. The regions where LSA has been decreasing were spatially correlated to areas of increased vegetation greenness. Along rising altitude, LSA decreasing rate exhibited an overall decreasing trend. Across the TP, elevated vegetation greenness in grasslands acted as a primary factor pulling down LSA. The driving effects of vegetation greenness on LSA vary with grassland types, as revealed by a more significant relationship between vegetation greenness and LSA for the sparsely vegetated zone (i.e. steppe) than the more densely vegetated zone (i.e. meadow). Furthermore, the driving effect of vegetation greenness on LSA exhibited an obvious dependence on altitude as effects with rising altitude were relatively strong up to 3000 m, then weakened from 3500 m to 5000 m, and then the effects again increased from 5000 to 6000 m. The growing season LSA trend revealed in this study emphasizes the need to give greater attention to the growing season LSA flux in future surface energy balance studies.
Simulation of the vegetation spatial heterogeneity over the semiarid grassland
NASA Astrophysics Data System (ADS)
Zeng, X.; Lian, R.
2013-12-01
Along the transition zones between large-scale grassland and desert in arid and semiarid regions, vegetation is usually sparse and forms various fine scale vegetation patches. This work attempts to better understand the underlying ecohydrology of these observations by relating the fine scale patterns to the large-scale ecosystem behaviors, including the following issues: (1) what is the variability (and diversity) of vegetation pattern formation? (2) how is the fine scale heterogeneity relevant to the mean ecosystem state variables in large scale? and (3) how does the fine scale heterogeneity influence the system behaviors such as transition and resilience? In the first study, fine-scale horizontal interaction terms are introduced to a single column model of large scale grassland to explicitly reproduce the vegetation patterns of spots and stripes. High diversities are found in two very different spatial scales. (1) In terms of spatial vegetation formation (microscale), the ecosystem undergoes the transition from uniform grassland to regular and irregular vegetation patterns, and then to pure desert, as the moisture index (i.e., the ratio of precipitation over potential evaporation) decreases. An infinite number of equilibrium states coexist in the transition zone. (2) In terms of spatially averaged biomass (macroscale), the transition zone from grassland to desert consists of a narrow region with erratic vegetation patterns and high diversity of averaged biomass, in contrast to just a single critical point in representing the transition in the single column model. In the second study, a spatial implicit model is developed by the introducing of different leaf/root allocation strategy. By allocating higher percentage of its photosynthetic product to root and creating a root area lager than the crown area (i.e., the area occupied by leaves), plant can absorb water from surrounding bare soil and hence take the advantage over plant with lower root area. The optimal
Improved bidirectional retrieval of sparse patterns stored by Hebbian learning.
Sommer, Friedrich T.; Palm, Gunther
1999-03-01
The Willshaw model is asymptotically the most efficient neural associative memory (NAM), but its finite version is hampered by high retrieval errors. Iterative retrieval has been proposed in a large number of different models to improve performance in auto-association tasks. In this paper, bidirectional retrieval for the hetero-associative memory task is considered: we define information efficiency as a general performance measure for bidirectional associative memory (BAM) and determine its asymptotic bound for the bidirectional Willshaw model. For the finite Willshaw model, an efficient new bidirectional retrieval strategy is proposed, the appropriate combinatorial model analysis is derived, and implications of the proposed sparse BAM for applications and brain theory are discussed. The distribution of the dendritic sum in the finite Willshaw model given by Buckingham and Willshaw [Buckingham, J., & Willshaw, D. (1992). Performance characteristics of associative nets. Network, 3, 407-414] allows no fast numerical evaluation. We derive a combinatorial formula with a highly reduced evaluation time that is used in the improved error analysis of the basic model and for estimation of the retrieval error in the naive model extension, where bidirectional retrieval is employed in the hetero-associative Willshaw model. The analysis rules out the naive BAM extension as a promising improvement. A new bidirectional retrieval algorithm - called crosswise bidirectional (CB) retrieval - is presented. The cross talk error is significantly reduced without employing more complex learning procedures or dummy augmentation in the pattern coding, as proposed in other refined BAM models [Wang, Y. F., Cruz, J. B., & Mulligan, J. H. (1990). Two coding strategies for bidirectional associative memory. IEEE Trans. Neural Networks, 1(1), 81-92; Leung, C.-S., Chan, L.-W., & Lai, E. (1995). Stability, capacity and statistical dynamics of second-order bidirectional associative memory. IEEE Trans
Sparse aperture 3D passive image sensing and recognition
NASA Astrophysics Data System (ADS)
Daneshpanah, Mehdi
The way we perceive, capture, store, communicate and visualize the world has greatly changed in the past century Novel three dimensional (3D) imaging and display systems are being pursued both in academic and industrial settings. In many cases, these systems have revolutionized traditional approaches and/or enabled new technologies in other disciplines including medical imaging and diagnostics, industrial metrology, entertainment, robotics as well as defense and security. In this dissertation, we focus on novel aspects of sparse aperture multi-view imaging systems and their application in quantum-limited object recognition in two separate parts. In the first part, two concepts are proposed. First a solution is presented that involves a generalized framework for 3D imaging using randomly distributed sparse apertures. Second, a method is suggested to extract the profile of objects in the scene through statistical properties of the reconstructed light field. In both cases, experimental results are presented that demonstrate the feasibility of the techniques. In the second part, the application of 3D imaging systems in sensing and recognition of objects is addressed. In particular, we focus on the scenario in which only 10s of photons reach the sensor from the object of interest, as opposed to hundreds of billions of photons in normal imaging conditions. At this level, the quantum limited behavior of light will dominate and traditional object recognition practices may fail. We suggest a likelihood based object recognition framework that incorporates the physics of sensing at quantum-limited conditions. Sensor dark noise has been modeled and taken into account. This framework is applied to 3D sensing of thermal objects using visible spectrum detectors. Thermal objects as cold as 250K are shown to provide enough signature photons to be sensed and recognized within background and dark noise with mature, visible band, image forming optics and detector arrays. The results
Sparse-view proton computed tomography using modulated proton beams
Lee, Jiseoc; Kim, Changhwan; Cho, Seungryong; Min, Byungjun; Kwak, Jungwon; Park, Seyjoon; Lee, Se Byeong; Park, Sungyong
2015-02-15
Purpose: Proton imaging that uses a modulated proton beam and an intensity detector allows a relatively fast image acquisition compared to the imaging approach based on a trajectory tracking detector. In addition, it requires a relatively simple implementation in a conventional proton therapy equipment. The model of geometric straight ray assumed in conventional computed tomography (CT) image reconstruction is however challenged by multiple-Coulomb scattering and energy straggling in the proton imaging. Radiation dose to the patient is another important issue that has to be taken care of for practical applications. In this work, the authors have investigated iterative image reconstructions after a deconvolution of the sparsely view-sampled data to address these issues in proton CT. Methods: Proton projection images were acquired using the modulated proton beams and the EBT2 film as an intensity detector. Four electron-density cylinders representing normal soft tissues and bone were used as imaged object and scanned at 40 views that are equally separated over 360°. Digitized film images were converted to water-equivalent thickness by use of an empirically derived conversion curve. For improving the image quality, a deconvolution-based image deblurring with an empirically acquired point spread function was employed. They have implemented iterative image reconstruction algorithms such as adaptive steepest descent-projection onto convex sets (ASD-POCS), superiorization method–projection onto convex sets (SM-POCS), superiorization method–expectation maximization (SM-EM), and expectation maximization-total variation minimization (EM-TV). Performance of the four image reconstruction algorithms was analyzed and compared quantitatively via contrast-to-noise ratio (CNR) and root-mean-square-error (RMSE). Results: Objects of higher electron density have been reconstructed more accurately than those of lower density objects. The bone, for example, has been reconstructed
Linearithmic time sparse and convex maximum margin clustering.
Zhang, Xiao-Lei; Wu, Ji
2012-12-01
Recently, a new clustering method called maximum margin clustering (MMC) was proposed and has shown promising performances. It was originally formulated as a difficult nonconvex integer problem. To make the MMC problem practical, the researchers either relaxed the original MMC problem to inefficient convex optimization problems or reformulated it to nonconvex optimization problems, which sacrifice the convexity for efficiency. However, no approaches can both hold the convexity and be efficient. In this paper, a new linearithmic time sparse and convex MMC algorithm, called support-vector-regression-based MMC (SVR-MMC), is proposed. Generally, it first uses the SVR as the core of the MMC. Then, it is relaxed as a convex optimization problem, which is iteratively solved by the cutting-plane algorithm. Each cutting-plane subproblem is further decomposed to a serial supervised SVR problem by a new global extended-level method (GELM). Finally, each supervised SVR problem is solved in a linear time complexity by a new sparse-kernel SVR (SKSVR) algorithm. We further extend the SVR-MMC algorithm to the multiple-kernel clustering (MKC) problem and the multiclass MMC (M3C) problem, which are denoted as SVR-MKC and SVR-M3C, respectively. One key point of the algorithms is the utilization of the SVR. It can prevent the MMC and its extensions meeting an integer matrix programming problem. Another key point is the new SKSVR. It provides a linear time interface to the nonlinear kernel scenarios, so that the SVR-MMC and its extensions can keep a linearthmic time complexity in nonlinear kernel scenarios. Our experimental results on various real-world data sets demonstrate the effectiveness and the efficiency of the SVR-MMC and its two extensions. Moreover, the unsupervised application of the SVR-MKC to the voice activity detection (VAD) shows that the SVR-MKC can achieve good performances that are close to its supervised counterpart, meet the real-time demand of the VAD, and need no
NASA Astrophysics Data System (ADS)
Yousefi Lalimi, F.; Silvestri, S.; Moore, L. J.; Marani, M.
2015-12-01
Vegetation plays a key role in barrier islands dynamics by moderating the effects of storms and sea level through sand stabilization and trapping. Vegetation establishment and growth are qualitatively known to depend on environmental factors that vary with elevation and distance from the shoreline, but we lack an adequate characterization of the spatial distribution of barrier island vegetation in relation to topographic determinants. In this study, we develop and apply remote sensing techniques to extend local field observations to larger scales and explore the relations between vegetation density and topography. We use hyperspectral and Lidar data from Hog Island (VA) to address the following questions: 1) Can we map dune vegetation species and density distributions using hyperspectral remote sensing? 2) Can we use LiDAR data to estimate canopy properties (Leaf Area Index, LAI)?, and 3) What are the relations between vegetation presence and topographic properties as they result from two-way ecogeomorphic interactions? Our results show that whereas mapping dune vegetation species using hyperspectral remote sensing is challenging, optical and LiDAR data can be successfully used to characterize overall vegetation density. A comparison between LAI and NDVI (Normalized Difference Vegetation Index), both estimated from remote sensing data, shows a coherent relationship and the spatial distributions of estimated LAI and NDVI agree with the fractional vegetation cover observed in the field. The results also suggest that cover density increases with distance from the shoreline. In the foredune area, vegetation is more sparse than other areas while it is densest on the back side of dunes. Analysis of NDVI and LAI from remote sensing observations can thus provide characterizations of topography and vegetation properties at the island scale, which will assist in shedding light on the interactions between vegetation and sediment dynamics that give rise to coastal dune systems.
Rank preserving sparse learning for Kinect based scene classification.
Tao, Dapeng; Jin, Lianwen; Yang, Zhao; Li, Xuelong
2013-10-01
With the rapid development of the RGB-D sensors and the promptly growing population of the low-cost Microsoft Kinect sensor, scene classification, which is a hard, yet important, problem in computer vision, has gained a resurgence of interest recently. That is because the depth of information provided by the Kinect sensor opens an effective and innovative way for scene classification. In this paper, we propose a new scheme for scene classification, which applies locality-constrained linear coding (LLC) to local SIFT features for representing the RGB-D samples and classifies scenes through the cooperation between a new rank preserving sparse learning (RPSL) based dimension reduction and a simple classification method. RPSL considers four aspects: 1) it preserves the rank order information of the within-class samples in a local patch; 2) it maximizes the margin between the between-class samples on the local patch; 3) the L1-norm penalty is introduced to obtain the parsimony property; and 4) it models the classification error minimization by utilizing the least-squares error minimization. Experiments are conducted on the NYU Depth V1 dataset and demonstrate the robustness and effectiveness of RPSL for scene classification. PMID:23846511
Sparse expression bases in cancer reveal tumor drivers.
Logsdon, Benjamin A; Gentles, Andrew J; Miller, Chris P; Blau, C Anthony; Becker, Pamela S; Lee, Su-In
2015-02-18
We define a new category of candidate tumor drivers in cancer genome evolution: 'selected expression regulators' (SERs)-genes driving dysregulated transcriptional programs in cancer evolution. The SERs are identified from genome-wide tumor expression data with a novel method, namely SPARROW ( SPAR: se selected exp R: essi O: n regulators identified W: ith penalized regression). SPARROW uncovers a previously unknown connection between cancer expression variation and driver events, by using a novel sparse regression technique. Our results indicate that SPARROW is a powerful complementary approach to identify candidate genes containing driver events that are hard to detect from sequence data, due to a large number of passenger mutations and lack of comprehensive sequence information from a sufficiently large number of samples. SERs identified by SPARROW reveal known driver mutations in multiple human cancers, along with known cancer-associated processes and survival-associated genes, better than popular methods for inferring gene expression networks. We demonstrate that when applied to acute myeloid leukemia expression data, SPARROW identifies an apoptotic biomarker (PYCARD) for an investigational drug obatoclax. The PYCARD and obatoclax association is validated in 30 AML patient samples.
A Comprehensive Validation Methodology for Sparse Experimental Data
NASA Technical Reports Server (NTRS)
Norman, Ryan B.; Blattnig, Steve R.
2010-01-01
A comprehensive program of verification and validation has been undertaken to assess the applicability of models to space radiation shielding applications and to track progress as models are developed over time. The models are placed under configuration control, and automated validation tests are used so that comparisons can readily be made as models are improved. Though direct comparisons between theoretical results and experimental data are desired for validation purposes, such comparisons are not always possible due to lack of data. In this work, two uncertainty metrics are introduced that are suitable for validating theoretical models against sparse experimental databases. The nuclear physics models, NUCFRG2 and QMSFRG, are compared to an experimental database consisting of over 3600 experimental cross sections to demonstrate the applicability of the metrics. A cumulative uncertainty metric is applied to the question of overall model accuracy, while a metric based on the median uncertainty is used to analyze the models from the perspective of model development by analyzing subsets of the model parameter space.
Stacked Predictive Sparse Decomposition for Classification of Histology Sections
Zhou, Yin; Borowsky, Alexander; Barner, Kenneth; Spellman, Paul
2016-01-01
Image-based classification of histology sections, in terms of distinct components (e.g., tumor, stroma, normal), provides a series of indices for histology composition (e.g., the percentage of each distinct components in histology sections), and enables the study of nuclear properties within each component. Furthermore, the study of these indices, constructed from each whole slide image in a large cohort, has the potential to provide predictive models of clinical outcome. For example, correlations can be established between the constructed indices and the patients’ survival information at cohort level, which is a fundamental step towards personalized medicine. However, performance of the existing techniques is hindered as a result of large technical variations (e.g., variations of color/textures in tissue images due to non-standard experimental protocols) and biological heterogeneities (e.g., cell type, cell state) that are always present in a large cohort. We propose a system that automatically learns a series of dictionary elements for representing the underlying spatial distribution using stacked predictive sparse decomposition. The learned representation is then fed into the spatial pyramid matching framework with a linear support vector machine classifier. The system has been evaluated for classification of distinct histological components for two cohorts of tumor types. Throughput has been increased by using of graphical processing unit (GPU), and evaluation indicates a superior performance results, compared with previous research.
Design and evaluation of sparse quantization index modulation watermarking schemes
NASA Astrophysics Data System (ADS)
Cornelis, Bruno; Barbarien, Joeri; Dooms, Ann; Munteanu, Adrian; Cornelis, Jan; Schelkens, Peter
2008-08-01
In the past decade the use of digital data has increased significantly. The advantages of digital data are, amongst others, easy editing, fast, cheap and cross-platform distribution and compact storage. The most crucial disadvantages are the unauthorized copying and copyright issues, by which authors and license holders can suffer considerable financial losses. Many inexpensive methods are readily available for editing digital data and, unlike analog information, the reproduction in the digital case is simple and robust. Hence, there is great interest in developing technology that helps to protect the integrity of a digital work and the copyrights of its owners. Watermarking, which is the embedding of a signal (known as the watermark) into the original digital data, is one method that has been proposed for the protection of digital media elements such as audio, video and images. In this article, we examine watermarking schemes for still images, based on selective quantization of the coefficients of a wavelet transformed image, i.e. sparse quantization-index modulation (QIM) watermarking. Different grouping schemes for the wavelet coefficients are evaluated and experimentally verified for robustness against several attacks. Wavelet tree-based grouping schemes yield a slightly improved performance over block-based grouping schemes. Additionally, the impact of the deployment of error correction codes on the most promising configurations is examined. The utilization of BCH-codes (Bose, Ray-Chaudhuri, Hocquenghem) results in an improved robustness as long as the capacity of the error codes is not exceeded (cliff-effect).
Analysis of the Proportional Hazards Model with Sparse Longitudinal Covariates
Cao, Hongyuan; Churpek, Mathew M.; Zeng, Donglin; Fine, Jason P.
2014-01-01
Regression analysis of censored failure observations via the proportional hazards model permits time-varying covariates which are observed at death times. In practice, such longitudinal covariates are typically sparse and only measured at infrequent and irregularly spaced follow-up times. Full likelihood analyses of joint models for longitudinal and survival data impose stringent modelling assumptions which are difficult to verify in practice and which are complicated both inferentially and computationally. In this article, a simple kernel weighted score function is proposed with minimal assumptions. Two scenarios are considered: half kernel estimation in which observation ceases at the time of the event and full kernel estimation for data where observation may continue after the event, as with recurrent events data. It is established that these estimators are consistent and asymptotically normal. However, they converge at rates which are slower than the parametric rates which may be achieved with fully observed covariates, with the full kernel method achieving an optimal convergence rate which is superior to that of the half kernel method. Simulation results demonstrate that the large sample approximations are adequate for practical use and may yield improved performance relative to last value carried forward approach and joint modelling method. The analysis of the data from a cardiac arrest study demonstrates the utility of the proposed methods. PMID:26576066
Precise RFID localization in impaired environment through sparse signal recovery
NASA Astrophysics Data System (ADS)
Subedi, Saurav; Zhang, Yimin D.; Amin, Moeness G.
2013-05-01
Radio frequency identification (RFID) is a rapidly developing wireless communication technology for electronically identifying, locating, and tracking products, assets, and personnel. RFID has become one of the most important means to construct real-time locating systems (RTLS) that track and identify the location of objects in real time using simple, inexpensive tags and readers. The applicability and usefulness of RTLS techniques depend on their achievable accuracy. In particular, when multilateration-based localization techniques are exploited, the achievable accuracy primarily relies on the precision of the range estimates between a reader and the tags. Such range information can be obtained by using the received signal strength indicator (RSSI) and/or the phase difference of arrival (PDOA). In both cases, however, the accuracy is significantly compromised when the operation environment is impaired. In particular, multipath propagation significantly affects the measurement accuracy of both RSSI and phase information. In addition, because RFID systems are typically operated in short distances, RSSI and phase measurements are also coupled with the reader and tag antenna patterns, making accurate RFID localization very complicated and challenging. In this paper, we develop new methods to localize RFID tags or readers by exploiting sparse signal recovery techniques. The proposed method allows the channel environment and antenna patterns to be taken into account and be properly compensated at a low computational cost. As such, the proposed technique yields superior performance in challenging operation environments with the above-mentioned impairments.
Simplified signal processing for impedance spectroscopy with spectrally sparse sequences
NASA Astrophysics Data System (ADS)
Annus, P.; Land, R.; Reidla, M.; Ojarand, J.; Mughal, Y.; Min, M.
2013-04-01
Classical method for measurement of the electrical bio-impedance involves excitation with sinusoidal waveform. Sinusoidal excitation at fixed frequency points enables wide variety of signal processing options, most general of them being Fourier transform. Multiplication with two quadrature waveforms at desired frequency could be easily accomplished both in analogue and in digital domains, even simplest quadrature square waves can be considered, which reduces signal processing task in analogue domain to synchronous switching followed by low pass filter, and in digital domain requires only additions. So called spectrally sparse excitation sequences (SSS), which have been recently introduced into bio-impedance measurement domain, are very reasonable choice when simultaneous multifrequency excitation is required. They have many good properties, such as ease of generation and good crest factor compared to similar multisinusoids. Typically, the usage of discrete or fast Fourier transform in signal processing step is considered so far. Usage of simplified methods nevertheless would reduce computational burden, and enable simpler, less costly and less energy hungry signal processing platforms. Accuracy of the measurement with SSS excitation when using different waveforms for quadrature demodulation will be compared in order to evaluate the feasibility of the simplified signal processing. Sigma delta modulated sinusoid (binary signal) is considered to be a good alternative for a synchronous demodulation.
Sparse representation utilizing tight frame for phase retrieval
NASA Astrophysics Data System (ADS)
Shi, Baoshun; Lian, Qiusheng; Chen, Shuzhen
2015-12-01
We treat the phase retrieval (PR) problem of reconstructing the interest signal from its Fourier magnitude. Since the Fourier phase information is lost, the problem is ill-posed. Several techniques have been used to address this problem by utilizing various priors such as non-negative, support, and Fourier magnitude constraints. Recent methods exploiting sparsity are developed to improve the reconstruction quality. However, the previous algorithms of utilizing sparsity prior suffer from either the low reconstruction quality at low oversampled factors or being sensitive to noise. To address these issues, we propose a framework that exploits sparsity of the signal in the translation invariant Haar pyramid (TIHP) tight frame. Based on this sparsity prior, we formulate the sparse representation regularization term and incorporate it into the PR optimization problem. We propose the alternating iterative algorithm for solving the corresponding non-convex problem by dividing the problem into several subproblems. We give the optimal solution to each subproblem, and experimental simulations under the noise-free and noisy scenario indicate that our proposed algorithm can obtain a better reconstruction quality compared to the conventional alternative projection methods, even outperform the recent sparsity-based algorithms in terms of reconstruction quality.
Perfectly correlated phase screen realization using sparse spectrum harmonic augmentation.
Naeh, Itay; Katzir, Abraham
2014-09-20
The split-step Fourier method is commonly used to simulate the propagation of radiation in a turbulent atmosphere using two-dimensional phase screens that have the desired spatial spectral content given by the atmospheric power spectrum. Using existing methodologies, isotropy of the structure function can never be achieved, mainly along the axis of propagation, for several reasons. In this paper, we introduce the sparse spectrum harmonic augmentation method that will address the lack of isotropy along the propagation axis, the limited achievable frequencies, and the limited time development possible using known approaches. Following the methodology described will produce phase screens that are transversely endless, perfectly correlated along the propagation axis, and contain the desired spectral content, including the low frequencies that even though they contain most of the energy, are usually neglected. The methodology presented can be used for many aspects of wave propagation in random media, such as atmospheric propagation, underwater acoustics, radio wave propagation in the ionosphere, and more. PMID:25322093
Guaranteed Blind Sparse Spikes Deconvolution via Lifting and Convex Optimization
NASA Astrophysics Data System (ADS)
Chi, Yuejie
2016-06-01
Neural recordings, returns from radars and sonars, images in astronomy and single-molecule microscopy can be modeled as a linear superposition of a small number of scaled and delayed copies of a band-limited or diffraction-limited point spread function, which is either determined by the nature or designed by the users; in other words, we observe the convolution between a point spread function and a sparse spike signal with unknown amplitudes and delays. While it is of great interest to accurately resolve the spike signal from as few samples as possible, however, when the point spread function is not known a priori, this problem is terribly ill-posed. This paper proposes a convex optimization framework to simultaneously estimate the point spread function as well as the spike signal, by mildly constraining the point spread function to lie in a known low-dimensional subspace. By applying the lifting trick, we obtain an underdetermined linear system of an ensemble of signals with joint spectral sparsity, to which atomic norm minimization is applied. Under mild randomness assumptions of the low-dimensional subspace as well as a separation condition of the spike signal, we prove the proposed algorithm, dubbed as AtomicLift, is guaranteed to recover the spike signal up to a scaling factor as soon as the number of samples is large enough. The extension of AtomicLift to handle noisy measurements is also discussed. Numerical examples are provided to validate the effectiveness of the proposed approaches.
A Bayesian semiparametric model for bivariate sparse longitudinal data.
Das, Kiranmoy; Li, Runze; Sengupta, Subhajit; Wu, Rongling
2013-09-30
Mixed-effects models have recently become popular for analyzing sparse longitudinal data that arise naturally in biological, agricultural and biomedical studies. Traditional approaches assume independent residuals over time and explain the longitudinal dependence by random effects. However, when bivariate or multivariate traits are measured longitudinally, this fundamental assumption is likely to be violated because of intertrait dependence over time. We provide a more general framework where the dependence of the observations from the same subject over time is not assumed to be explained completely by the random effects of the model. We propose a novel, mixed model-based approach and estimate the error-covariance structure nonparametrically under a generalized linear model framework. We use penalized splines to model the general effect of time, and we consider a Dirichlet process mixture of normal prior for the random-effects distribution. We analyze blood pressure data from the Framingham Heart Study where body mass index, gender and time are treated as covariates. We compare our method with traditional methods including parametric modeling of the random effects and independent residual errors over time. We conduct extensive simulation studies to investigate the practical usefulness of the proposed method. The current approach is very helpful in analyzing bivariate irregular longitudinal traits. PMID:23553747
Convergence and Rate Analysis of Neural Networks for Sparse Approximation
Balavoine, Aurèle; Romberg, Justin; Rozell, Christopher J.
2013-01-01
We present an analysis of the Locally Competitive Algotihm (LCA), which is a Hopfield-style neural network that efficiently solves sparse approximation problems (e.g., approximating a vector from a dictionary using just a few nonzero coefficients). This class of problems plays a significant role in both theories of neural coding and applications in signal processing. However, the LCA lacks analysis of its convergence properties, and previous results on neural networks for nonsmooth optimization do not apply to the specifics of the LCA architecture. We show that the LCA has desirable convergence properties, such as stability and global convergence to the optimum of the objective function when it is unique. Under some mild conditions, the support of the solution is also proven to be reached in finite time. Furthermore, some restrictions on the problem specifics allow us to characterize the convergence rate of the system by showing that the LCA converges exponentially fast with an analytically bounded convergence rate. We support our analysis with several illustrative simulations. PMID:24199030
Adaptive sparse grid expansions of the vibrational Hamiltonian.
Strobusch, D; Scheurer, Ch
2014-02-21
The vibrational Hamiltonian involves two high dimensional operators, the kinetic energy operator (KEO), and the potential energy surface (PES). Both must be approximated for systems involving more than a few atoms. Adaptive approximation schemes are not only superior to truncated Taylor or many-body expansions (MBE), they also allow for error estimates, and thus operators of predefined precision. To this end, modified sparse grids (SG) are developed that can be combined with adaptive MBEs. This MBE/SG hybrid approach yields a unified, fully adaptive representation of the KEO and the PES. Refinement criteria, based on the vibrational self-consistent field (VSCF) and vibrational configuration interaction (VCI) methods, are presented. The combination of the adaptive MBE/SG approach and the VSCF plus VCI methods yields a black box like procedure to compute accurate vibrational spectra. This is demonstrated on a test set of molecules, comprising water, formaldehyde, methanimine, and ethylene. The test set is first employed to prove convergence for semi-empirical PM3-PESs and subsequently to compute accurate vibrational spectra from CCSD(T)-PESs that agree well with experimental values.
Learning signaling network structures with sparsely distributed data.
Sachs, Karen; Itani, Solomon; Carlisle, Jennifer; Nolan, Garry P; Pe'er, Dana; Lauffenburger, Douglas A
2009-02-01
Flow cytometric measurement of signaling protein abundances has proved particularly useful for elucidation of signaling pathway structure. The single cell nature of the data ensures a very large dataset size, providing a statistically robust dataset for structure learning. Moreover, the approach is easily scaled to many conditions in high throughput. However, the technology suffers from a dimensionality constraint: at the cutting edge, only about 12 protein species can be measured per cell, far from sufficient for most signaling pathways. Because the structure learning algorithm (in practice) requires that all variables be measured together simultaneously, this restricts structure learning to the number of variables that constitute the flow cytometer's upper dimensionality limit. To address this problem, we present here an algorithm that enables structure learning for sparsely distributed data, allowing structure learning beyond the measurement technology's upper dimensionality limit for simultaneously measurable variables. The algorithm assesses pairwise (or n-wise) dependencies, constructs "Markov neighborhoods" for each variable based on these dependencies, measures each variable in the context of its neighborhood, and performs structure learning using a constrained search.
Wide field of view multifocal scanning microscopy with sparse sampling
NASA Astrophysics Data System (ADS)
Wang, Jie; Wu, Jigang
2016-02-01
We propose to use sparsely sampled line scans with a sparsity-based reconstruction method to obtain images in a wide field of view (WFOV) multifocal scanning microscope. In the WFOV microscope, we used a holographically generated irregular focus grid to scan the sample in one dimension and then reconstructed the sample image from line scans by measuring the transmission of the foci through the sample during scanning. The line scans were randomly spaced with average spacing larger than the Nyquist sampling requirement, and the image was recovered with sparsity-based reconstruction techniques. With this scheme, the acquisition data can be significantly reduced and the restriction for equally spaced foci positions can be removed, indicating simpler experimental requirement. We built a prototype system and demonstrated the effectiveness of the reconstruction by recovering microscopic images of a U.S. Air Force target and an onion skin cell microscope slide with 40, 60, and 80% missing data with respect to the Nyquist sampling requirement.
Convergence and rate analysis of neural networks for sparse approximation.
Balavoine, Aurèle; Romberg, Justin; Rozell, Christopher J
2012-09-01
We present an analysis of the Locally Competitive Algorithm (LCA), which is a Hopfield-style neural network that efficiently solves sparse approximation problems (e.g., approximating a vector from a dictionary using just a few nonzero coefficients). This class of problems plays a significant role in both theories of neural coding and applications in signal processing. However, the LCA lacks analysis of its convergence properties, and previous results on neural networks for nonsmooth optimization do not apply to the specifics of the LCA architecture. We show that the LCA has desirable convergence properties, such as stability and global convergence to the optimum of the objective function when it is unique. Under some mild conditions, the support of the solution is also proven to be reached in finite time. Furthermore, some restrictions on the problem specifics allow us to characterize the convergence rate of the system by showing that the LCA converges exponentially fast with an analytically bounded convergence rate. We support our analysis with several illustrative simulations.
Distributed dictionary learning for sparse representation in sensor networks.
Liang, Junli; Zhang, Miaohua; Zeng, Xianyu; Yu, Guoyang
2014-06-01
This paper develops a distributed dictionary learning algorithm for sparse representation of the data distributed across nodes of sensor networks, where the sensitive or private data are stored or there is no fusion center or there exists a big data application. The main contributions of this paper are: 1) we decouple the combined dictionary atom update and nonzero coefficient revision procedure into two-stage operations to facilitate distributed computations, first updating the dictionary atom in terms of the eigenvalue decomposition of the sum of the residual (correlation) matrices across the nodes then implementing a local projection operation to obtain the related representation coefficients for each node; 2) we cast the aforementioned atom update problem as a set of decentralized optimization subproblems with consensus constraints. Then, we simplify the multiplier update for the symmetry undirected graphs in sensor networks and minimize the separable subproblems to attain the consistent estimates iteratively; and 3) dictionary atoms are typically constrained to be of unit norm in order to avoid the scaling ambiguity. We efficiently solve the resultant hidden convex subproblems by determining the optimal Lagrange multiplier. Some experiments are given to show that the proposed algorithm is an alternative distributed dictionary learning approach, and is suitable for the sensor network environment. PMID:24733009
Sparse Nonnegative Matrix Factorization Strategy for Cochlear Implants.
Hu, Hongmei; Lutman, Mark E; Ewert, Stephan D; Li, Guoping; Bleeck, Stefan
2015-01-01
Current cochlear implant (CI) strategies carry speech information via the waveform envelope in frequency subbands. CIs require efficient speech processing to maximize information transfer to the brain, especially in background noise, where the speech envelope is not robust to noise interference. In such conditions, the envelope, after decomposition into frequency bands, may be enhanced by sparse transformations, such as nonnegative matrix factorization (NMF). Here, a novel CI processing algorithm is described, which works by applying NMF to the envelope matrix (envelopogram) of 22 frequency channels in order to improve performance in noisy environments. It is evaluated for speech in eight-talker babble noise. The critical sparsity constraint parameter was first tuned using objective measures and then evaluated with subjective speech perception experiments for both normal hearing and CI subjects. Results from vocoder simulations with 10 normal hearing subjects showed that the algorithm significantly enhances speech intelligibility with the selected sparsity constraints. Results from eight CI subjects showed no significant overall improvement compared with the standard advanced combination encoder algorithm, but a trend toward improvement of word identification of about 10 percentage points at +15 dB signal-to-noise ratio (SNR) was observed in the eight CI subjects. Additionally, a considerable reduction of the spread of speech perception performance from 40% to 93% for advanced combination encoder to 80% to 100% for the suggested NMF coding strategy was observed. PMID:26721919
Novel calibration system with sparse wires for CMB polarization receivers
Tajima, O.; Nguyen, H.; Bischoff, C.; Brizius, A.; Buder, I.; Kusaka, A. /Chicago U., KICP
2011-07-01
B-modes in the cosmic microwave background (CMB) polarization is a smoking gun signature of the inflationary universe. To achieve better sensitivity to this faint signal, CMB polarization experiments aim to maximize the number of detector elements, resulting in a large focal plane receiver. Detector calibration of the polarization response becomes essential. It is extremely useful to be able to calibrate 'simultaneously' all detectors on the large focal plane. We developed a novel calibration system that rotates a large 'sparse' grid of metal wires, in front of and fully covering the field of view of the focal plane receiver. Polarized radiation is created via the reflection of ambient temperature from the wire surface. Since the detector has a finite beam size, the observed signal is smeared according to the beam property. The resulting smeared polarized radiation has a reasonable intensity (a few Kelvin or less) compared to the sky temperature ({approx}10 K observing condition). The system played a successful role for receiver calibration of QUIET, a CMB polarization experiment located in the Atacama desert in Chile. The successful performance revealed that this system is applicable to other experiments based on different technologies, e.g. TES bolometers.
Denoising Sparse Images from GRAPPA using the Nullspace Method (DESIGN)
Weller, Daniel S.; Polimeni, Jonathan R.; Grady, Leo; Wald, Lawrence L.; Adalsteinsson, Elfar; Goyal, Vivek K
2011-01-01
To accelerate magnetic resonance imaging using uniformly undersampled (nonrandom) parallel imaging beyond what is achievable with GRAPPA alone, the Denoising of Sparse Images from GRAPPA using the Nullspace method (DESIGN) is developed. The trade-off between denoising and smoothing the GRAPPA solution is studied for different levels of acceleration. Several brain images reconstructed from uniformly undersampled k-space data using DESIGN are compared against reconstructions using existing methods in terms of difference images (a qualitative measure), PSNR, and noise amplification (g-factors) as measured using the pseudo-multiple replica method. Effects of smoothing, including contrast loss, are studied in synthetic phantom data. In the experiments presented, the contrast loss and spatial resolution are competitive with existing methods. Results for several brain images demonstrate significant improvements over GRAPPA at high acceleration factors in denoising performance with limited blurring or smoothing artifacts. In addition, the measured g-factors suggest that DESIGN mitigates noise amplification better than both GRAPPA and L1 SPIR-iT (the latter limited here by uniform undersampling). PMID:22213069
Sparse detector sensor: profiling experiments for broad-scale classification
NASA Astrophysics Data System (ADS)
Russomanno, D. J.; Yeasin, M.; Jacobs, E.; Smith, M.; Sorower, S.
2008-04-01
This paper presents a simple prototype sparse detector imaging sensor built using sixteen off-the-shelf, retro-reflective, infrared-sensing elements placed at five-inch intervals in a vertical configuration. Profiling experiments for broad-scale classification of objects, such as humans, humans wearing large backpacks, and humans wearing small backpacks were conducted from data acquired from the sensor. Empirical analysis on models developed using fusion of various classifiers based on a diversity measure shows over ninety percent (90.07%) accuracy (using 10-fold cross validation) in categorizing sensed data into specific classes of interest, such as, humans wearing a large backpack. The results demonstrate that shadow images of sufficient resolution can be captured by the sensor such that broad-scale classification of objects is feasible. The sensor appears to be a low-cost alternative to traditional, high-resolution imaging sensors, and, after industrial packaging, it may be a good candidate for deployment in vast geographic regions in which low-cost, unattended ground sensors with highly accurate classification algorithms are needed.
Sparse radar imaging using 2D compressed sensing
NASA Astrophysics Data System (ADS)
Hou, Qingkai; Liu, Yang; Chen, Zengping; Su, Shaoying
2014-10-01
Radar imaging is an ill-posed linear inverse problem and compressed sensing (CS) has been proved to have tremendous potential in this field. This paper surveys the theory of radar imaging and a conclusion is drawn that the processing of ISAR imaging can be denoted mathematically as a problem of 2D sparse decomposition. Based on CS, we propose a novel measuring strategy for ISAR imaging radar and utilize random sub-sampling in both range and azimuth dimensions, which will reduce the amount of sampling data tremendously. In order to handle 2D reconstructing problem, the ordinary solution is converting the 2D problem into 1D by Kronecker product, which will increase the size of dictionary and computational cost sharply. In this paper, we introduce the 2D-SL0 algorithm into the reconstruction of imaging. It is proved that 2D-SL0 can achieve equivalent result as other 1D reconstructing methods, but the computational complexity and memory usage is reduced significantly. Moreover, we will state the results of simulating experiments and prove the effectiveness and feasibility of our method.
Sparse Spatio-temporal Inference of Electromagnetic Brain Sources
NASA Astrophysics Data System (ADS)
Stahlhut, Carsten; Attias, Hagai T.; Wipf, David; Hansen, Lars K.; Nagarajan, Srikantan S.
The electromagnetic brain activity measured via MEG (or EEG) can be interpreted as arising from a collection of current dipoles or sources located throughout the cortex. Because the number of candidate locations for these sources is much larger than the number of sensors, source reconstruction involves solving an inverse problem that is severely underdetermined. Bayesian graphical models provide a powerful means of incorporating prior assumptions that narrow the solution space and lead to tractable posterior distributions over the unknown sources given the observed data. In particular, this paper develops a hierarchical, spatio-temporal Bayesian model that accommodates the principled computation of sparse spatial and smooth temporal M/EEG source reconstructions consistent with neurophysiological assumptions in a variety of event-related imaging paradigms. The underlying methodology relies on the notion of automatic relevance determination (ARD) to express the unknown sources via a small collection of spatio-temporal basis functions. Experiments with several data sets provide evidence that the proposed model leads to improved source estimates. The underlying methodology is also well-suited for estimation problems that arise from other brain imaging modalities such as functional or diffusion weighted MRI.
Gaussian kernel width optimization for sparse Bayesian learning.
Mohsenzadeh, Yalda; Sheikhzadeh, Hamid
2015-04-01
Sparse kernel methods have been widely used in regression and classification applications. The performance and the sparsity of these methods are dependent on the appropriate choice of the corresponding kernel functions and their parameters. Typically, the kernel parameters are selected using a cross-validation approach. In this paper, a learning method that is an extension of the relevance vector machine (RVM) is presented. The proposed method can find the optimal values of the kernel parameters during the training procedure. This algorithm uses an expectation-maximization approach for updating kernel parameters as well as other model parameters; therefore, the speed of convergence and computational complexity of the proposed method are the same as the standard RVM. To control the convergence of this fully parameterized model, the optimization with respect to the kernel parameters is performed using a constraint on these parameters. The proposed method is compared with the typical RVM and other competing methods to analyze the performance. The experimental results on the commonly used synthetic data, as well as benchmark data sets, demonstrate the effectiveness of the proposed method in reducing the performance dependency on the initial choice of the kernel parameters. PMID:25794377
Sparse deformable models with application to cardiac motion analysis.
Yu, Yang; Zhang, Shaoting; Huang, Junzhou; Metaxas, Dimitris; Axel, Leon
2013-01-01
Deformable models have been widely used with success in medical image analysis. They combine bottom-up information derived from image appearance cues, with top-down shape-based constraints within a physics-based formulation. However, in many real world problems the observations extracted from the image data often contain gross errors, which adversely affect the deformation accuracy. To alleviate this issue, we introduce a new family of deformable models that are inspired from compressed sensing, a technique for efficiently reconstructing a signal based on its sparseness in some domain. In this problem, we employ sparsity to represent the outliers or gross errors, and combine it seamlessly with deformable models. The proposed new formulation is applied to the analysis of cardiac motion, using tagged magnetic resonance imaging (tMRI), where the automated tagging line tracking results are very noisy due to the poor image quality. Our new deformable models track the heart motion robustly, and the resulting strains are consistent with those calculated from manual labels. PMID:24683970
Extreme learning machine and adaptive sparse representation for image classification.
Cao, Jiuwen; Zhang, Kai; Luo, Minxia; Yin, Chun; Lai, Xiaoping
2016-09-01
Recent research has shown the speed advantage of extreme learning machine (ELM) and the accuracy advantage of sparse representation classification (SRC) in the area of image classification. Those two methods, however, have their respective drawbacks, e.g., in general, ELM is known to be less robust to noise while SRC is known to be time-consuming. Consequently, ELM and SRC complement each other in computational complexity and classification accuracy. In order to unify such mutual complementarity and thus further enhance the classification performance, we propose an efficient hybrid classifier to exploit the advantages of ELM and SRC in this paper. More precisely, the proposed classifier consists of two stages: first, an ELM network is trained by supervised learning. Second, a discriminative criterion about the reliability of the obtained ELM output is adopted to decide whether the query image can be correctly classified or not. If the output is reliable, the classification will be performed by ELM; otherwise the query image will be fed to SRC. Meanwhile, in the stage of SRC, a sub-dictionary that is adaptive to the query image instead of the entire dictionary is extracted via the ELM output. The computational burden of SRC thus can be reduced. Extensive experiments on handwritten digit classification, landmark recognition and face recognition demonstrate that the proposed hybrid classifier outperforms ELM and SRC in classification accuracy with outstanding computational efficiency.
Sparse Nonnegative Matrix Factorization Strategy for Cochlear Implants
Lutman, Mark E.; Ewert, Stephan D.; Li, Guoping; Bleeck, Stefan
2015-01-01
Current cochlear implant (CI) strategies carry speech information via the waveform envelope in frequency subbands. CIs require efficient speech processing to maximize information transfer to the brain, especially in background noise, where the speech envelope is not robust to noise interference. In such conditions, the envelope, after decomposition into frequency bands, may be enhanced by sparse transformations, such as nonnegative matrix factorization (NMF). Here, a novel CI processing algorithm is described, which works by applying NMF to the envelope matrix (envelopogram) of 22 frequency channels in order to improve performance in noisy environments. It is evaluated for speech in eight-talker babble noise. The critical sparsity constraint parameter was first tuned using objective measures and then evaluated with subjective speech perception experiments for both normal hearing and CI subjects. Results from vocoder simulations with 10 normal hearing subjects showed that the algorithm significantly enhances speech intelligibility with the selected sparsity constraints. Results from eight CI subjects showed no significant overall improvement compared with the standard advanced combination encoder algorithm, but a trend toward improvement of word identification of about 10 percentage points at +15 dB signal-to-noise ratio (SNR) was observed in the eight CI subjects. Additionally, a considerable reduction of the spread of speech perception performance from 40% to 93% for advanced combination encoder to 80% to 100% for the suggested NMF coding strategy was observed. PMID:26721919
Distributed dictionary learning for sparse representation in sensor networks.
Liang, Junli; Zhang, Miaohua; Zeng, Xianyu; Yu, Guoyang
2014-06-01
This paper develops a distributed dictionary learning algorithm for sparse representation of the data distributed across nodes of sensor networks, where the sensitive or private data are stored or there is no fusion center or there exists a big data application. The main contributions of this paper are: 1) we decouple the combined dictionary atom update and nonzero coefficient revision procedure into two-stage operations to facilitate distributed computations, first updating the dictionary atom in terms of the eigenvalue decomposition of the sum of the residual (correlation) matrices across the nodes then implementing a local projection operation to obtain the related representation coefficients for each node; 2) we cast the aforementioned atom update problem as a set of decentralized optimization subproblems with consensus constraints. Then, we simplify the multiplier update for the symmetry undirected graphs in sensor networks and minimize the separable subproblems to attain the consistent estimates iteratively; and 3) dictionary atoms are typically constrained to be of unit norm in order to avoid the scaling ambiguity. We efficiently solve the resultant hidden convex subproblems by determining the optimal Lagrange multiplier. Some experiments are given to show that the proposed algorithm is an alternative distributed dictionary learning approach, and is suitable for the sensor network environment.
Sparse source configurations in radio tomography of asteroids
NASA Astrophysics Data System (ADS)
Pursiainen, S.; Kaasalainen, M.
2014-07-01
Our research targets at progress in non-invasive imaging of asteroids to support future planetary research and extra-terrestrial mining activities. This presentation concerns principally radio tomography in which the permittivity distribution inside an asteroid is to be recovered based on the radio frequency signal transmitted from the asteroid's surface and gathered by an orbiter. The focus will be on a sparse distribution (Pursiainen and Kaasalainen, 2013) of signal sources that can be necessary in the challenging in situ environment and within tight payload limits. The general goal in our recent research has been to approximate the minimal number of source positions needed for robust localization of anomalies caused, for example, by an internal void. Characteristic to the localization problem are the large relative changes in signal speed caused by the high permittivity of typical asteroid minerals (e.g. basalt), meaning that a signal path can include strong refractions and reflections. This presentation introduces results of a laboratory experiment in which real travel time data was inverted using a hierarchical Bayesian approach combined with the iterative alternating sequential (IAS) posterior exploration algorithm. Special interest was paid to robustness of the inverse results regarding changes of the prior model and source positioning. According to our results, strongly refractive anomalies can be detected with three or four sources independently of their positioning.