Lee, Anthony; Yau, Christopher; Giles, Michael B.; Doucet, Arnaud; Holmes, Christopher C.
2011-01-01
We present a case-study on the utility of graphics cards to perform massively parallel simulation of advanced Monte Carlo methods. Graphics cards, containing multiple Graphics Processing Units (GPUs), are self-contained parallel computational devices that can be housed in conventional desktop and laptop computers and can be thought of as prototypes of the next generation of many-core processors. For certain classes of population-based Monte Carlo algorithms they offer massively parallel simulation, with the added advantage over conventional distributed multi-core processors that they are cheap, easily accessible, easy to maintain, easy to code, dedicated local devices with low power consumption. On a canonical set of stochastic simulation examples including population-based Markov chain Monte Carlo methods and Sequential Monte Carlo methods, we nd speedups from 35 to 500 fold over conventional single-threaded computer code. Our findings suggest that GPUs have the potential to facilitate the growth of statistical modelling into complex data rich domains through the availability of cheap and accessible many-core computation. We believe the speedup we observe should motivate wider use of parallelizable simulation methods and greater methodological attention to their design. PMID:22003276
Human sense utilization method on real-time computer graphics
NASA Astrophysics Data System (ADS)
Maehara, Hideaki; Ohgashi, Hitoshi; Hirata, Takao
1997-06-01
We are developing an adjustment method of real-time computer graphics, to obtain effective ones which give audience various senses intended by producer, utilizing human sensibility technologically. Generally, production of real-time computer graphics needs much adjustment of various parameters, such as 3D object models/their motions/attributes/view angle/parallax etc., in order that the graphics gives audience superior effects as reality of materials, sense of experience and so on. And it is also known it costs much to adjust such various parameters by trial and error. A graphics producer often evaluates his graphics to improve it. For example, it may lack 'sense of speed' or be necessary to be given more 'sense of settle down,' to improve it. On the other hand, we can know how the parameters in computer graphics affect such senses by means of statistically analyzing several samples of computer graphics which provide different senses. We paid attention to these two facts, so that we designed an adjustment method of the parameters by inputting phases of sense into a computer. By the way of using this method, it becomes possible to adjust real-time computer graphics more effectively than by conventional way of trial and error.
ERIC Educational Resources Information Center
Ives, William; Rovet, Joanne
1979-01-01
Reports three experiments which investigate: whether familiar objects have standard graphic orientations (Experiment 1); the relationship between use of object orientations and more conventional methods in depicting familiar objects in motion (Experiment 2); and whether orientations are used differently in novel objects whose only defining feature…
Acquired Codes of Meaning in Data Visualization and Infographics: Beyond Perceptual Primitives.
Byrne, Lydia; Angus, Daniel; Wiles, Janet
2016-01-01
While information visualization frameworks and heuristics have traditionally been reluctant to include acquired codes of meaning, designers are making use of them in a wide variety of ways. Acquired codes leverage a user's experience to understand the meaning of a visualization. They range from figurative visualizations which rely on the reader's recognition of shapes, to conventional arrangements of graphic elements which represent particular subjects. In this study, we used content analysis to codify acquired meaning in visualization. We applied the content analysis to a set of infographics and data visualizations which are exemplars of innovative and effective design. 88% of the infographics and 71% of data visualizations in the sample contain at least one use of figurative visualization. Conventions on the arrangement of graphics are also widespread in the sample. In particular, a comparison of representations of time and other quantitative data showed that conventions can be specific to a subject. These results suggest that there is a need for information visualization research to expand its scope beyond perceptual channels, to include social and culturally constructed meaning. Our paper demonstrates a viable method for identifying figurative techniques and graphic conventions and integrating them into heuristics for visualization design.
NASA Astrophysics Data System (ADS)
Hagita, Norihiro; Sawaki, Minako
1995-03-01
Most conventional methods in character recognition extract geometrical features such as stroke direction, connectivity of strokes, etc., and compare them with reference patterns in a stored dictionary. Unfortunately, geometrical features are easily degraded by blurs, stains and the graphical background designs used in Japanese newspaper headlines. This noise must be removed before recognition commences, but no preprocessing method is completely accurate. This paper proposes a method for recognizing degraded characters and characters printed on graphical background designs. This method is based on the binary image feature method and uses binary images as features. A new similarity measure, called the complementary similarity measure, is used as a discriminant function. It compares the similarity and dissimilarity of binary patterns with reference dictionary patterns. Experiments are conducted using the standard character database ETL-2 which consists of machine-printed Kanji, Hiragana, Katakana, alphanumeric, an special characters. The results show that this method is much more robust against noise than the conventional geometrical feature method. It also achieves high recognition rates of over 92% for characters with textured foregrounds, over 98% for characters with textured backgrounds, over 98% for outline fonts, and over 99% for reverse contrast characters.
Kin, Taichi; Nakatomi, Hirofumi; Shojima, Masaaki; Tanaka, Minoru; Ino, Kenji; Mori, Harushi; Kunimatsu, Akira; Oyama, Hiroshi; Saito, Nobuhito
2012-07-01
In this study, the authors used preoperative simulation employing 3D computer graphics (interactive computer graphics) to fuse all imaging data for brainstem cavernous malformations. The authors evaluated whether interactive computer graphics or 2D imaging correlated better with the actual operative field, particularly in identifying a developmental venous anomaly (DVA). The study population consisted of 10 patients scheduled for surgical treatment of brainstem cavernous malformations. Data from preoperative imaging (MRI, CT, and 3D rotational angiography) were automatically fused using a normalized mutual information method, and then reconstructed by a hybrid method combining surface rendering and volume rendering methods. With surface rendering, multimodality and multithreshold techniques for 1 tissue were applied. The completed interactive computer graphics were used for simulation of surgical approaches and assumed surgical fields. Preoperative diagnostic rates for a DVA associated with brainstem cavernous malformation were compared between conventional 2D imaging and interactive computer graphics employing receiver operating characteristic (ROC) analysis. The time required for reconstruction of 3D images was 3-6 hours for interactive computer graphics. Observation in interactive mode required approximately 15 minutes. Detailed anatomical information for operative procedures, from the craniotomy to microsurgical operations, could be visualized and simulated three-dimensionally as 1 computer graphic using interactive computer graphics. Virtual surgical views were consistent with actual operative views. This technique was very useful for examining various surgical approaches. Mean (±SEM) area under the ROC curve for rate of DVA diagnosis was significantly better for interactive computer graphics (1.000±0.000) than for 2D imaging (0.766±0.091; p<0.001, Mann-Whitney U-test). The authors report a new method for automatic registration of preoperative imaging data from CT, MRI, and 3D rotational angiography for reconstruction into 1 computer graphic. The diagnostic rate of DVA associated with brainstem cavernous malformation was significantly better using interactive computer graphics than with 2D images. Interactive computer graphics was also useful in helping to plan the surgical access corridor.
Li, Xiangrui; Lu, Zhong-Lin
2012-02-29
Display systems based on conventional computer graphics cards are capable of generating images with 8-bit gray level resolution. However, most experiments in vision research require displays with more than 12 bits of luminance resolution. Several solutions are available. Bit++ (1) and DataPixx (2) use the Digital Visual Interface (DVI) output from graphics cards and high resolution (14 or 16-bit) digital-to-analog converters to drive analog display devices. The VideoSwitcher (3) described here combines analog video signals from the red and blue channels of graphics cards with different weights using a passive resister network (4) and an active circuit to deliver identical video signals to the three channels of color monitors. The method provides an inexpensive way to enable high-resolution monochromatic displays using conventional graphics cards and analog monitors. It can also provide trigger signals that can be used to mark stimulus onsets, making it easy to synchronize visual displays with physiological recordings or response time measurements. Although computer keyboards and mice are frequently used in measuring response times (RT), the accuracy of these measurements is quite low. The RTbox is a specialized hardware and software solution for accurate RT measurements. Connected to the host computer through a USB connection, the driver of the RTbox is compatible with all conventional operating systems. It uses a microprocessor and high-resolution clock to record the identities and timing of button events, which are buffered until the host computer retrieves them. The recorded button events are not affected by potential timing uncertainties or biases associated with data transmission and processing in the host computer. The asynchronous storage greatly simplifies the design of user programs. Several methods are available to synchronize the clocks of the RTbox and the host computer. The RTbox can also receive external triggers and be used to measure RT with respect to external events. Both VideoSwitcher and RTbox are available for users to purchase. The relevant information and many demonstration programs can be found at http://lobes.usc.edu/.
Lessons from a doctoral thesis.
Peiris, A N; Mueller, R A; Sheridan, D P
1990-01-01
The production of a doctoral thesis is a time-consuming affair that until recently was done in conjunction with professional publishing services. Advances in computer technology have made many sophisticated desktop publishing techniques available to the microcomputer user. We describe the computer method used, the problems encountered, and the solutions improvised in the production of a doctoral thesis by computer. The Apple Macintosh was selected for its ease of use and intrinsic graphics capabilities. A scanner was used to incorporate text from published papers into a word processing program. The body of the text was updated and supplemented with new sections. Scanned graphics from the published papers were less suitable for publication, and the original data were replotted and modified with a graphics-drawing program. Graphics were imported and incorporated in the text. Final hard copy was produced by a laser printer and bound with both conventional and rapid new binding techniques. Microcomputer-based desktop processing methods provide a rapid and cost-effective means of communicating the written word. We anticipate that this evolving technology will have increased use by physicians in both the private and academic sectors.
Publishing on the WWW. Part 1 - Static graphics
Grech, V
2000-01-01
An on-line journal's ability to publish graphics at no additional cost is a major advantage over conventional printed journals. This article outlines technical, copyright and other issues related to graphic publishing on the world-wide-web. PMID:22368588
Using Visualization and Computation in the Analysis of Separation Processes
ERIC Educational Resources Information Center
Joo, Yong Lak; Choudhary, Devashish
2006-01-01
For decades, every chemical engineer has been asked to have a background in separations. The required separations course can, however, be uninspiring and superficial because understanding many separation processes involves conventional graphical methods and commercial process simulators. We utilize simple, user-friendly mathematical software,…
Absolute Coefficients and the Graphical Representation of Airfoil Characteristics
NASA Technical Reports Server (NTRS)
Munk, Max
1921-01-01
It is argued that there should be an agreement as to what conventions to use in determining absolute coefficients used in aeronautics and in how to plot those coefficients. Of particular importance are the absolute coefficients of lift and drag. The author argues for the use of the German method over the kind in common use in the United States and England, and for the Continental over the usual American and British method of graphically representing the characteristics of an airfoil. The author notes that, on the whole, it appears that the use of natural absolute coefficients in a polar diagram is the logical method for presentation of airfoil characteristics, and that serious consideration should be given to the advisability of adopting this method in all countries, in order to advance uniformity and accuracy in the science of aeronautics.
Shade matching assisted by digital photography and computer software.
Schropp, Lars
2009-04-01
To evaluate the efficacy of digital photographs and graphic computer software for color matching compared to conventional visual matching. The shade of a tab from a shade guide (Vita 3D-Master Guide) placed in a phantom head was matched to a second guide of the same type by nine observers. This was done for twelve selected shade tabs (tests). The shade-matching procedure was performed visually in a simulated clinic environment and with digital photographs, and the time spent for both procedures was recorded. An alternative arrangement of the shade tabs was used in the digital photographs. In addition, a graphic software program was used for color analysis. Hue, chroma, and lightness values of the test tab and all tabs of the second guide were derived from the digital photographs. According to the CIE L*C*h* color system, the color differences between the test tab and tabs of the second guide were calculated. The shade guide tab that deviated least from the test tab was determined to be the match. Shade matching performance by means of graphic software was compared with the two visual methods and tested by Chi-square tests (alpha= 0.05). Eight of twelve test tabs (67%) were matched correctly by the computer software method. This was significantly better (p < 0.02) than the performance of the visual shade matching methods conducted in the simulated clinic (32% correct match) and with photographs (28% correct match). No correlation between time consumption for the visual shade matching methods and frequency of correct match was observed. Shade matching assisted by digital photographs and computer software was significantly more reliable than by conventional visual methods.
"Bhimayana": Caste, Ambedkar, Art, and Pedagogy
ERIC Educational Resources Information Center
Anand, S.; Vellanki, Vivek
2015-01-01
"Bhimayana" is a graphic novel that narrates Dr. B.R. Ambedkar's experiences of being discriminated against. Using a graphic form inspired by Pardhan Gond art, "Bhimayana" breaks popular conventions of graphic narratives published in the West. The narrative of "Bhimayana" is interlaced with contemporary events and…
A Correlational Study of Graphic Organizers and Science Achievement of English Language Learners
NASA Astrophysics Data System (ADS)
Clarke, William Gordon
English language learners (ELLs) demonstrate lower academic performance and have lower graduation and higher dropout rates than their non-ELL peers. The primary purpose of this correlational quantitative study was to investigate the relationship between the use of graphic organizer-infused science instruction and science learning of high school ELLs. Another objective was to determine if the method of instruction, socioeconomic status (SES), gender, and English language proficiency (ELP) were predictors of academic achievement of high school ELLs. Data were gathered from a New York City (NYC) high school fall 2012-2013 archival records of 145 ninth-grade ELLs who had received biology instruction in freestanding English as a second language (ESL) classes, followed by a test of their learning of the material. Fifty-four (37.2%) of these records were of students who had learned science by the conventional textbook method, and 91 (62.8%) by using graphic organizers. Data analysis employed the Statistical Package for the Social Sciences (SPSS) software for multiple regression analysis, which found graphic organizer use to be a significant predictor of New York State Regents Living Environment (NYSRLE) test scores (p < .01). One significant regression model was returned whereby, when combined, the four predictor variables (method of instruction, SES, gender, and ELP) explained 36% of the variance of the NYSRLE score. Implications of the study findings noted graphic organizer use as advantageous for ELL science achievement. Recommendations made for practice were for (a) the adoption of graphic organizer infused-instruction, (b) establishment of a protocol for the implementation of graphic organizer-infused instruction, and (c) increased length of graphic organizer instructional time. Recommendations made for future research were (a) a replication quantitative correlational study in two or more high schools, (b) a quantitative quasi-experimental quantitative study to determine the influence of graphic organizer instructional intervention and ELL science achievement, (c) a quantitative quasi-experimental study to determine the effect of teacher-based factors on graphic organizer-infused instruction, and (c) a causal comparative study to determine the efficacy of graphic organizer use in testing modifications for high school ELL science.
ERIC Educational Resources Information Center
Boerman-Cornell, William
2012-01-01
Recent studies of graphic novels (book-length fiction or non-fiction narratives that employ the conventions of comic books to convey meaning) and multimodality have hinted that graphic novels (GNs) might offer a great deal of meaning-making potential to readers. Some studies have argued that graphic novels could be useful for English Language…
NASA Technical Reports Server (NTRS)
Elliott, R. D.; Werner, N. M.; Baker, W. M.
1975-01-01
The Aerodynamic Data Analysis and Integration System (ADAIS), developed as a highly interactive computer graphics program capable of manipulating large quantities of data such that addressable elements of a data base can be called up for graphic display, compared, curve fit, stored, retrieved, differenced, etc., was described. The general nature of the system is evidenced by the fact that limited usage has already occurred with data bases consisting of thermodynamic, basic loads, and flight dynamics data. Productivity using ADAIS of five times that for conventional manual methods of wind tunnel data analysis is routinely achieved. In wind tunnel data analysis, data from one or more runs of a particular test may be called up and displayed along with data from one or more runs of a different test. Curves may be faired through the data points by any of four methods, including cubic spline and least squares polynomial fit up to seventh order.
Hexographic Method of Complex Town-Planning Terrain Estimate
NASA Astrophysics Data System (ADS)
Khudyakov, A. Ju
2017-11-01
The article deals with the vital problem of a complex town-planning analysis based on the “hexographic” graphic analytic method, makes a comparison with conventional terrain estimate methods and contains the method application examples. It discloses a procedure of the author’s estimate of restrictions and building of a mathematical model which reflects not only conventional town-planning restrictions, but also social and aesthetic aspects of the analyzed territory. The method allows one to quickly get an idea of the territory potential. It is possible to use an unlimited number of estimated factors. The method can be used for the integrated assessment of urban areas. In addition, it is possible to use the methods of preliminary evaluation of the territory commercial attractiveness in the preparation of investment projects. The technique application results in simple informative graphics. Graphical interpretation is straightforward from the experts. A definite advantage is the free perception of the subject results as they are not prepared professionally. Thus, it is possible to build a dialogue between professionals and the public on a new level allowing to take into account the interests of various parties. At the moment, the method is used as a tool for the preparation of integrated urban development projects at the Department of Architecture in Federal State Autonomous Educational Institution of Higher Education “South Ural State University (National Research University)”, FSAEIHE SUSU (NRU). The methodology is included in a course of lectures as the material on architectural and urban design for architecture students. The same methodology was successfully tested in the preparation of business strategies for the development of some territories in the Chelyabinsk region. This publication is the first in a series of planned activities developing and describing the methodology of hexographical analysis in urban and architectural practice. It is also planned to create a software product that allows one to automate the process of site assessment on the basis of the methodology.
A human performance evaluation of graphic symbol-design features.
Samet, M G; Geiselman, R E; Landee, B M
1982-06-01
16 subjects learned each of two tactical display symbol sets (conventional symbols and iconic symbols) in turn and were then shown a series of graphic displays containing various symbol configurations. For each display, the subject was asked questions corresponding to different behavioral processes relating to symbol use (identification, search, comparison, pattern recognition). The results indicated that: (a) conventional symbols yielded faster pattern-recognition performance than iconic symbols, and iconic symbols did not yield faster identification than conventional symbols, and (b) the portrayal of additional feature information (through the use of perimeter density or vector projection coding) slowed processing of the core symbol information in four tasks, but certain symbol-design features created less perceptual interference and had greater correspondence with the portrayal of specific tactical concepts than others. The results were discussed in terms of the complexities involved in the selection of symbol design features for use in graphic tactical displays.
HMI conventions for process control graphics.
Pikaar, Ruud N
2012-01-01
Process operators supervise and control complex processes. To enable the operator to do an adequate job, instrumentation and process control engineers need to address several related topics, such as console design, information design, navigation, and alarm management. In process control upgrade projects, usually a 1:1 conversion of existing graphics is proposed. This paper suggests another approach, efficiently leading to a reduced number of new powerful process graphics, supported by a permanent process overview displays. In addition a road map for structuring content (process information) and conventions for the presentation of objects, symbols, and so on, has been developed. The impact of the human factors engineering approach on process control upgrade projects is illustrated by several cases.
ERIC Educational Resources Information Center
Simonson, Michael R., Ed.; And Others
1996-01-01
This proceedings volume contains 77 papers. Subjects addressed include: image processing; new faculty research methods; preinstructional activities for preservice teacher education; computer "window" presentation styles; interface design; stress management instruction; cooperative learning; graphical user interfaces; student attitudes,…
Statistical, Graphical, and Learning Methods for Sensing, Surveillance, and Navigation Systems
2016-06-28
harsh propagation environments. Conventional filtering techniques fail to provide satisfactory performance in many important nonlinear or non...Gaussian scenarios. In addition, there is a lack of a unified methodology for the design and analysis of different filtering techniques. To address...these problems, we have proposed a new filtering methodology called belief condensation (BC) DISTRIBUTION A: Distribution approved for public release
LUMIS Interactive graphics operating instructions and system specifications
NASA Technical Reports Server (NTRS)
Bryant, N. A.; Yu, T. C.; Landini, A. J.
1976-01-01
The LUMIS program has designed an integrated geographic information system to assist program managers and planning groups in metropolitan regions. Described is the system designed to interactively interrogate a data base, display graphically a portion of the region enclosed in the data base, and perform cross tabulations of variables within each city block, block group, or census tract. The system is designed to interface with U. S. Census DIME file technology, but can accept alternative districting conventions. The system is described on three levels: (1) introduction to the systems's concept and potential applications; (2) the method of operating the system on an interactive terminal; and (3) a detailed system specification for computer facility personnel.
Applications of computer-graphics animation for motion-perception research
NASA Technical Reports Server (NTRS)
Proffitt, D. R.; Kaiser, M. K.
1986-01-01
The advantages and limitations of using computer animated stimuli in studying motion perception are presented and discussed. Most current programs of motion perception research could not be pursued without the use of computer graphics animation. Computer generated displays afford latitudes of freedom and control that are almost impossible to attain through conventional methods. There are, however, limitations to this presentational medium. At present, computer generated displays present simplified approximations of the dynamics in natural events. Very little is known about how the differences between natural events and computer simulations influence perceptual processing. In practice, the differences are assumed to be irrelevant to the questions under study, and that findings with computer generated stimuli will generalize to natural events.
Applications guide for waste heat recovery
NASA Technical Reports Server (NTRS)
Moynihan, P. I.
1983-01-01
The state-of-the-art of commercially available organic Rankine cycle (ORC) hardware from a literature search and industry survey is assessed. Engineering criteria for applying ORC technology are established, and a set of nomograms to enable the rapid sizing of the equipment is presented. A comparison of an ORC system with conventional heat recovery techniques can be made with a nomogram developed for a recuperative heat exchanger. A graphical technique for evaluating the economic aspects of an ORC system and conventional heat recovery method is discussed: also included is a description of anticipated future trends in organic Rankine cycle R&D.
Stiltner, G.J.
1990-01-01
In 1987, the Water Resources Division of the U.S. Geological Survey undertook three pilot projects to evaluate electronic report processing systems as a means to improve the quality and timeliness of reports pertaining to water resources investigations. The three projects selected for study included the use of the following configuration of software and hardware: Ventura Publisher software on an IBM model AT personal computer, PageMaker software on a Macintosh computer, and FrameMaker software on a Sun Microsystems workstation. The following assessment criteria were to be addressed in the pilot studies: The combined use of text, tables, and graphics; analysis of time; ease of learning; compatibility with the existing minicomputer system; and technical limitations. It was considered essential that the camera-ready copy produced be in a format suitable for publication. Visual improvement alone was not a consideration. This report consolidates and summarizes the findings of the electronic report processing pilot projects. Text and table files originating on the existing minicomputer system were successfully transformed to the electronic report processing systems in American Standard Code for Information Interchange (ASCII) format. Graphics prepared using a proprietary graphics software package were transferred to all the electronic report processing software through the use of Computer Graphic Metafiles. Graphics from other sources were entered into the systems by scanning paper images. Comparative analysis of time needed to process text and tables by the electronic report processing systems and by conventional methods indicated that, although more time is invested in creating the original page composition for an electronically processed report , substantial time is saved in producing subsequent reports because the format can be stored and re-used by electronic means as a template. Because of the more compact page layouts, costs of printing the reports were 15% to 25% less than costs of printing the reports prepared by conventional methods. Because the largest report workload in the offices conducting water resources investigations is preparation of Water-Resources Investigations Reports, Open-File Reports, and annual State Data Reports, the pilot studies only involved these projects. (USGS)
Real-time lens distortion correction: speed, accuracy and efficiency
NASA Astrophysics Data System (ADS)
Bax, Michael R.; Shahidi, Ramin
2014-11-01
Optical lens systems suffer from nonlinear geometrical distortion. Optical imaging applications such as image-enhanced endoscopy and image-based bronchoscope tracking require correction of this distortion for accurate localization, tracking, registration, and measurement of image features. Real-time capability is desirable for interactive systems and live video. The use of a texture-mapping graphics accelerator, which is standard hardware on current motherboard chipsets and add-in video graphics cards, to perform distortion correction is proposed. Mesh generation for image tessellation, an error analysis, and performance results are presented. It is shown that distortion correction using commodity graphics hardware is substantially faster than using the main processor and can be performed at video frame rates (faster than 30 frames per second), and that the polar-based method of mesh generation proposed here is more accurate than a conventional grid-based approach. Using graphics hardware to perform distortion correction is not only fast and accurate but also efficient as it frees the main processor for other tasks, which is an important issue in some real-time applications.
NASA Astrophysics Data System (ADS)
Wiana, W.
2018-02-01
This research is related to the effort to design a more representative learning system to improve the learning result of digital fashion design, through the development of interactive multimedia based on motion graphic. This research is aimed to know the effect of interactive multimedia application based on motion graphic to increase the mastery of the concept and skill of the students to making fashion designing in digital format. The research method used is quasi experiment with research design of Non-equivalent Control Group Design. The lectures are conducted in two different classes, namely class A as the Experimental Class and class B as the Control Class. From the calculation result after interpreted using Normalize Gain, there is an increase of higher learning result in student with interactive learning based on motion graphic, compared with student achievement on conventional learning. In this research, interactive multimedia learning based on motion graphic is effective toward the improvement of student learning in concept mastering indicator and on the aspect of making fashion design in digital format.
Using a graphical programming language to write CAMAC/GPIB instrument drivers
NASA Technical Reports Server (NTRS)
Zambrana, Horacio; Johanson, William
1991-01-01
To reduce the complexities of conventional programming, graphical software was used in the development of instrumentation drivers. The graphical software provides a standard set of tools (graphical subroutines) which are sufficient to program the most sophisticated CAMAC/GPIB drivers. These tools were used and instrumentation drivers were successfully developed for operating CAMAC/GPIB hardware from two different manufacturers: LeCroy and DSP. The use of these tools is presented for programming a LeCroy A/D Waveform Analyzer.
Blood-Pressure Measuring System Gives Accurate Graphic Output
NASA Technical Reports Server (NTRS)
1965-01-01
The problem: To develop an instrument that will provide an external (indirect) measurement of arterial blood pressure in the form of an easily interpreted graphic trace that can be correlated with standard clinical blood-pressure measurements. From sphygmograms produced by conventional sphygmographs, it is very difficult to differentiate the systolic and diastolic blood-pressure pulses and to correlate these indices with the standard clinical values. It is nearly impossible to determine these indices when the subject is under physical or emotional stress. The solution: An electronic blood-pressure system, basically similar to conventional ausculatory sphygmomanometers, employing a standard occluding cuff, a gas-pressure source, and a gas-pressure regulator and valve. An electrical output transducer senses cuff pressure, and a microphone positioned on the brachial artery under the occluding cuff monitors the Korotkoff sounds from this artery. The output signals present the conventional systolic and diastolic indices in a clear, graphical display. The complete system also includes an electronic timer and cycle-control circuit.
Microrheology with optical tweezers: measuring the relative viscosity of solutions 'at a glance'.
Tassieri, Manlio; Del Giudice, Francesco; Robertson, Emma J; Jain, Neena; Fries, Bettina; Wilson, Rab; Glidle, Andrew; Greco, Francesco; Netti, Paolo Antonio; Maffettone, Pier Luca; Bicanic, Tihana; Cooper, Jonathan M
2015-03-06
We present a straightforward method for measuring the relative viscosity of fluids via a simple graphical analysis of the normalised position autocorrelation function of an optically trapped bead, without the need of embarking on laborious calculations. The advantages of the proposed microrheology method are evident when it is adopted for measurements of materials whose availability is limited, such as those involved in biological studies. The method has been validated by direct comparison with conventional bulk rheology methods, and has been applied both to characterise synthetic linear polyelectrolytes solutions and to study biomedical samples.
Microrheology with Optical Tweezers: Measuring the relative viscosity of solutions ‘at a glance'
Tassieri, Manlio; Giudice, Francesco Del; Robertson, Emma J.; Jain, Neena; Fries, Bettina; Wilson, Rab; Glidle, Andrew; Greco, Francesco; Netti, Paolo Antonio; Maffettone, Pier Luca; Bicanic, Tihana; Cooper, Jonathan M.
2015-01-01
We present a straightforward method for measuring the relative viscosity of fluids via a simple graphical analysis of the normalised position autocorrelation function of an optically trapped bead, without the need of embarking on laborious calculations. The advantages of the proposed microrheology method are evident when it is adopted for measurements of materials whose availability is limited, such as those involved in biological studies. The method has been validated by direct comparison with conventional bulk rheology methods, and has been applied both to characterise synthetic linear polyelectrolytes solutions and to study biomedical samples. PMID:25743468
A GPU-based calculation using the three-dimensional FDTD method for electromagnetic field analysis.
Nagaoka, Tomoaki; Watanabe, Soichi
2010-01-01
Numerical simulations with the numerical human model using the finite-difference time domain (FDTD) method have recently been performed frequently in a number of fields in biomedical engineering. However, the FDTD calculation runs too slowly. We focus, therefore, on general purpose programming on the graphics processing unit (GPGPU). The three-dimensional FDTD method was implemented on the GPU using Compute Unified Device Architecture (CUDA). In this study, we used the NVIDIA Tesla C1060 as a GPGPU board. The performance of the GPU is evaluated in comparison with the performance of a conventional CPU and a vector supercomputer. The results indicate that three-dimensional FDTD calculations using a GPU can significantly reduce run time in comparison with that using a conventional CPU, even a native GPU implementation of the three-dimensional FDTD method, while the GPU/CPU speed ratio varies with the calculation domain and thread block size.
Medical workstation design: enhancing graphical interface with 3D anatomical atlas
NASA Astrophysics Data System (ADS)
Hoo, Kent S., Jr.; Wong, Stephen T. C.; Grant, Ellen
1997-05-01
The huge data archive of the UCSF Hospital Integrated Picture Archiving and Communication System gives healthcare providers access to diverse kinds of images and text for diagnosis and patient management. Given the mass of information accessible, however, conventional graphical user interface (GUI) approach overwhelms the user with forms, menus, fields, lists, and other widgets and causes 'information overloading.' This article describes a new approach that complements the conventional GUI with 3D anatomical atlases and presents the usefulness of this approach with a clinical neuroimaging application.
A visual analysis of multi-attribute data using pixel matrix displays
NASA Astrophysics Data System (ADS)
Hao, Ming C.; Dayal, Umeshwar; Keim, Daniel; Schreck, Tobias
2007-01-01
Charts and tables are commonly used to visually analyze data. These graphics are simple and easy to understand, but charts show only highly aggregated data and present only a limited number of data values while tables often show too many data values. As a consequence, these graphics may either lose or obscure important information, so different techniques are required to monitor complex datasets. Users need more powerful visualization techniques to digest and compare detailed multi-attribute data to analyze the health of their business. This paper proposes an innovative solution based on the use of pixel-matrix displays to represent transaction-level information. With pixelmatrices, users can visualize areas of importance at a glance, a capability not provided by common charting techniques. We present our solutions to use colored pixel-matrices in (1) charts for visualizing data patterns and discovering exceptions, (2) tables for visualizing correlations and finding root-causes, and (3) time series for visualizing the evolution of long-running transactions. The solutions have been applied with success to product sales, Internet network performance analysis, and service contract applications demonstrating the benefits of our method over conventional graphics. The method is especially useful when detailed information is a key part of the analysis.
Chemistry for Kids: What Do Chemists Do? A Program for Grades 1-3.
ERIC Educational Resources Information Center
Umland, Jean B.
1985-01-01
Proposes a simple solution to the problem of expressing graphically absolute configuration, racemic character, and relative configuration in molecules. As demonstrated by several examples, the proposed convention brings added clarity and simplicity to structural drawings and renders graphic displays of chiral molecules as explicit and information…
Wargaming and interactive color graphics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bly, S.; Buzzell, C.; Smith, G.
1980-08-04
JANUS is a two-sided interactive color graphic simulation in which human commanders can direct their forces, each trying to accomplish their mission. This competitive synthetic battlefield is used to explore the range of human ingenuity under conditions of incomplete information about enemy strength and deployment. Each player can react to new situations by planning new unit movements, using conventional and nuclear weapons, or modifying unit objectives. Conventional direct fire among tanks, infantry fighting vehicles, helicopters, and other units is automated subject to constraints of target acquisition, reload rate, range, suppression, etc. Artillery and missile indirect fire systems deliver conventional munitions,more » smoke, and nuclear weapons. Players use reconnaissance units, helicopters, or fixed wing aircraft to search for enemy unit locations. Counter-battery radars acquire enemy artillery. The JANUS simulation at LLL has demonstrated the value of the computer as a sophisticated blackboard. A small dedicated minicomputer is adequate for detailed calculations, and may be preferable to sharing a more powerful machine. Real-time color interactive graphics are essential to allow realistic command decision inputs. Competitive human-versus-human synthetic experiences are intense and well-remembered. 2 figures.« less
ERIC Educational Resources Information Center
Boerman-Cornell, Bill
2013-01-01
Graphic novels (book-length fiction or nonfiction narratives told using the conventions of a comic book) bring together text and image in a way that seems to capture students' imaginations. Right now, there is little more than anecdotal research about how graphic novels can be used within specific middle school and high school disciplines. As…
Energy shadowing correction of ultrasonic pulse-echo records by digital signal processing
NASA Technical Reports Server (NTRS)
Kishonio, D.; Heyman, J. S.
1985-01-01
A numerical algorithm is described that enables the correction of energy shadowing during the ultrasonic testing of bulk materials. In the conventional method, an ultrasonic transducer transmits sound waves into a material that is immersed in water so that discontinuities such as defects can be revealed when the waves are reflected and then detected and displayed graphically. Since a defect that lies behind another defect is shadowed in that it receives less energy, the conventional method has a major drawback. The algorithm normalizes the energy of the incoming wave by measuring the energy of the waves reflected off the water/air interface. The algorithm is fast and simple enough to be adopted for real time applications in industry. Images of material defects with the shadowing corrections permit more quantitative interpretation of the material state.
Deep learning-based depth estimation from a synthetic endoscopy image training set
NASA Astrophysics Data System (ADS)
Mahmood, Faisal; Durr, Nicholas J.
2018-03-01
Colorectal cancer is the fourth leading cause of cancer deaths worldwide. The detection and removal of premalignant lesions through an endoscopic colonoscopy is the most effective way to reduce colorectal cancer mortality. Unfortunately, conventional colonoscopy has an almost 25% polyp miss rate, in part due to the lack of depth information and contrast of the surface of the colon. Estimating depth using conventional hardware and software methods is challenging in endoscopy due to limited endoscope size and deformable mucosa. In this work, we use a joint deep learning and graphical model-based framework for depth estimation from endoscopy images. Since depth is an inherently continuous property of an object, it can easily be posed as a continuous graphical learning problem. Unlike previous approaches, this method does not require hand-crafted features. Large amounts of augmented data are required to train such a framework. Since there is limited availability of colonoscopy images with ground-truth depth maps and colon texture is highly patient-specific, we generated training images using a synthetic, texture-free colon phantom to train our models. Initial results show that our system can estimate depths for phantom test data with a relative error of 0.164. The resulting depth maps could prove valuable for 3D reconstruction and automated Computer Aided Detection (CAD) to assist in identifying lesions.
Interactive computer aided technology, evolution in the design/manufacturing process
NASA Technical Reports Server (NTRS)
English, C. H.
1975-01-01
A powerful computer-operated three dimensional graphic system and associated auxiliary computer equipment used in advanced design, production design, and manufacturing was described. This system has made these activities more productive than when using older and more conventional methods to design and build aerospace vehicles. With the use of this graphic system, designers are now able to define parts using a wide variety of geometric entities, define parts as fully surface 3-dimensional models as well as "wire-frame" models. Once geometrically defined, the designer is able to take section cuts of the surfaced model and automatically determine all of the section properties of the planar cut, lightpen detect all of the surface patches and automatically determine the volume and weight of the part. Further, his designs are defined mathematically at a degree of accuracy never before achievable.
So long as they grow out of it: comics, the discourse of developmental normalcy, and disability.
Squier, Susan M
2008-06-01
This essay draws on two emerging fields--the study of comics or graphic fiction, and disability studies--to demonstrate how graphic fictions articulate the embodied, ethical, and sociopolitical experiences of impairment and disability. Examining David B's Epileptic and Paul Karasik and Judy Karasik's The Ride Together, I argue that these graphic novels unsettle conventional notions of normalcy and disability. In so doing, they also challenge our assumed dimensions and possibilities of the comics genre and medium, demonstrating the great potential comics hold for disability studies.
Drawing a Crowd: Graphic Novel Events Are Great Ways to Generate Excitement
ERIC Educational Resources Information Center
MacDonald, Heidi
2004-01-01
As graphic novels grow in popularity, with teen readers, libraries are finding the world of comics, manga, and anime a fertile field for inspiring events, as well as a great way to promote libraries to teens in general. These events can range from simple--a reading/signing--to elaborate--workshops, or even a mini convention. All provide a unique…
Quantitative graphical analysis of simultaneous dynamic PET/MRI for assessment of prostate cancer.
Rosenkrantz, Andrew B; Koesters, Thomas; Vahle, Anne-Kristin; Friedman, Kent; Bartlett, Rachel M; Taneja, Samir S; Ding, Yu-Shin; Logan, Jean
2015-04-01
Dynamic FDG imaging for prostate cancer characterization is limited by generally small size and low uptake in prostate tumors. Our aim in this pilot study was to explore feasibility of simultaneous PET/MRI to guide localization of prostate lesions for dynamic FDG analysis using a graphical approach. Three patients with biopsy-proven prostate cancer underwent simultaneous FDG PET/MRI, incorporating dynamic prostate imaging. Histology and multiparametric MRI findings were used to localize tumors, which in turn guided identification of tumors on FDG images. Regions of interest were manually placed on tumor and benign prostate tissue. Blood activity was extracted from a region of interest placed on the femoral artery on PET images. FDG data were analyzed by graphical analysis using the influx constant Ki (Patlak analysis) when FDG binding seemed irreversible and distribution volume VT (reversible graphical analysis) when FDG binding seemed reversible given the presence of washout. Given inherent coregistration, simultaneous acquisition facilitated use of MRI data to localize small lesions on PET and subsequent graphical analysis in all cases. In 2 cases with irreversible binding, tumor had higher Ki than benign using Patlak analysis (0.023 vs 0.006 and 0.019 vs 0.008 mL/cm3 per minute). In 1 case appearing reversible, tumor had higher VT than benign using reversible graphical analysis (0.68 vs 0.52 mL/cm3). Simultaneous PET/MRI allows localization of small prostate tumors for dynamic PET analysis. By taking advantage of inclusion of the femoral arteries in the FOV, we applied advanced PET data analysis methods beyond conventional static measures and without blood sampling.
NASA Astrophysics Data System (ADS)
Russkova, Tatiana V.
2017-11-01
One tool to improve the performance of Monte Carlo methods for numerical simulation of light transport in the Earth's atmosphere is the parallel technology. A new algorithm oriented to parallel execution on the CUDA-enabled NVIDIA graphics processor is discussed. The efficiency of parallelization is analyzed on the basis of calculating the upward and downward fluxes of solar radiation in both a vertically homogeneous and inhomogeneous models of the atmosphere. The results of testing the new code under various atmospheric conditions including continuous singlelayered and multilayered clouds, and selective molecular absorption are presented. The results of testing the code using video cards with different compute capability are analyzed. It is shown that the changeover of computing from conventional PCs to the architecture of graphics processors gives more than a hundredfold increase in performance and fully reveals the capabilities of the technology used.
Translating Scientific Conclusions about Risk for Public Audiences
NASA Astrophysics Data System (ADS)
Bowman, T. E.
2009-12-01
Climate change has been aptly described as a problem of risk management, yet the scientific community has not been successful in helping the public engage in risk management effectively. Behavioral science studies demonstrate that, while the public generally accepts the reality of anthropogenic climate change today, the immanence of impacts and scale of risk and opportunities for effective mitigation are poorly understood. Helping the public overcome these misperceptions and engage in decision-making about climate risks is, perhaps, the climate communication community’s most urgent priority. Scientific writing and graphic conventions are poorly suited for communicating with non-scientists. Using examples from the IPCC 4th Assessment, this session will demonstrate how specific conventions in science writing and graph making have obscured critical information about climate risks. The session will further demonstrate how reformatting the graphical information can create an exceptionally clear picture of where humanity stands and the implications of various emissions pathways for the future. Attendees will appreciate how presentations of science results can be tailored to answer the public’s questions more effectively by highlighting useful information in accurate, yet accessible ways. Decision-makers and the public urgently need information about climate impact risks and the consequences of various emissions pathways. Yet written and graphic descriptions from the IPCC and other assessment agencies burden non-scientists with multiple temperature baselines (e.g., pre-Industrial, mid-20th century, late 20th century, today), two confusingly similar measures for the key human contribution to atmospheric composition (CO2 and CO2-eq), and multiple ways of describing probability and certainty. The public is further confounded by inconsistent graphic conventions in scientific figures, including inconsistent color-coding, labeling, axis orientation, and treatment of uncertainty. Scientific figures tend to either include too many different messages or over-generalize, and neither approach helps non-scientists identify useful metrics and apply them to decision-making. This session will briefly illustrate each of these problematic scientific conventions and present a more effective translation of key IPCC figures in a new graphic format that help non-scientists appreciate our situation and opportunities. This translation has received positive reviews from informal learning institutions and will be useful to the broader science communication community.
Knowledge Glyphs: Visualization Theory Development to Support C2 Practice
2006-03-01
interface’s graphic structure (Calder and Linton, 2003). "• ’Glyphs’ as components of a typographical set (Microsoft Typography Standards). "* ’DataGlyphs...MOOTW) factors MIL STD 2525’s symbology set was designed for application in the context of geospatial representations - i.e., geographical maps. It is...the visual elements used to portray discrete entities. In a conventional windowing environment, such entities are likely to be graphically portrayed
Accelerating Molecular Dynamic Simulation on Graphics Processing Units
Friedrichs, Mark S.; Eastman, Peter; Vaidyanathan, Vishal; Houston, Mike; Legrand, Scott; Beberg, Adam L.; Ensign, Daniel L.; Bruns, Christopher M.; Pande, Vijay S.
2009-01-01
We describe a complete implementation of all-atom protein molecular dynamics running entirely on a graphics processing unit (GPU), including all standard force field terms, integration, constraints, and implicit solvent. We discuss the design of our algorithms and important optimizations needed to fully take advantage of a GPU. We evaluate its performance, and show that it can be more than 700 times faster than a conventional implementation running on a single CPU core. PMID:19191337
Bentzley, Brandon S.; Fender, Kimberly M.; Aston-Jones, Gary
2012-01-01
Rationale Behavioral-economic demand curve analysis offers several useful measures of drug self-administration. Although generation of demand curves previously required multiple days, recent within-session procedures allow curve construction from a single 110-min cocaine self-administration session, making behavioral-economic analyses available to a broad range of self-administration experiments. However, a mathematical approach of curve fitting has not been reported for the within-session threshold procedure. Objectives We review demand curve analysis in drug self-administration experiments and provide a quantitative method for fitting curves to single-session data that incorporates relative stability of brain drug concentration. Methods Sprague-Dawley rats were trained to self-administer cocaine, and then tested with the threshold procedure in which the cocaine dose was sequentially decreased on a fixed ratio-1 schedule. Price points (responses/mg cocaine) outside of relatively stable brain cocaine concentrations were removed before curves were fit. Curve-fit accuracy was determined by the degree of correlation between graphical and calculated parameters for cocaine consumption at low price (Q0) and the price at which maximal responding occurred (Pmax). Results Removing price points that occurred at relatively unstable brain cocaine concentrations generated precise estimates of Q0 and resulted in Pmax values with significantly closer agreement with graphical Pmax than conventional methods. Conclusion The exponential demand equation can be fit to single-session data using the threshold procedure for cocaine self-administration. Removing data points that occur during relatively unstable brain cocaine concentrations resulted in more accurate estimates of demand curve slope than graphical methods, permitting a more comprehensive analysis of drug self-administration via a behavioral-economic framework. PMID:23086021
Simulation of cryolipolysis as a novel method for noninvasive fat layer reduction.
Majdabadi, Abbas; Abazari, Mohammad
2016-12-20
Regarding previous problems in conventional liposuction methods, the need for development of new fat removal operations was appreciated. In this study we are going to simulate one of the novel methods, cryolipolysis, aimed to tackle those drawbacks. We think that simulation of clinical procedures contributes considerably in efficacious performance of the operations. To do this we have attempted to simulate temperature distribution in a sample fat of the human body. Using Abaqus software we have presented the graphical display of temperature-time variations within the medium. Findings of our simulation indicate that tissue temperature decreases after cold exposure of about 30 min. It can be seen that the minimum temperature of tissue occurs in shallow layers of the sample and the temperature in deeper layers of the sample remains nearly unchanged. It is clear that cold exposure time of more than the specific time (t > 30 min) does not result in considerable changes. Numerous clinical studies have proved the efficacy of cryolipolysis. This noninvasive technique has eliminated some of drawbacks of conventional methods. Findings of our simulation clearly prove the efficiency of this method, especially for superficial fat layers.
Startup of the Anammox Process in a Membrane Bioreactor (AnMBR) from Conventional Activated Sludge.
Gutwiński, P; Cema, G; Ziembińska-Buczyńska, A; Surmacz-Górska, J; Osadnik, M
2016-12-01
In this study, a laboratory-scale anammox process in a membrane bioreactor (AnMBR) was used to startup the anaerobic ammonium oxidation (anammox) process from conventional activated sludge. Stable operation was achieved after 125 days. From that time, nitrogen load was gradually increased. After six months, the average nitrogen removal efficiency exceeded 80%. The highest obtained special anammox activity (SAA) achieved was 0.17 g (
Zhang, Lei; Zeng, Zhi; Ji, Qiang
2011-09-01
Chain graph (CG) is a hybrid probabilistic graphical model (PGM) capable of modeling heterogeneous relationships among random variables. So far, however, its application in image and video analysis is very limited due to lack of principled learning and inference methods for a CG of general topology. To overcome this limitation, we introduce methods to extend the conventional chain-like CG model to CG model with more general topology and the associated methods for learning and inference in such a general CG model. Specifically, we propose techniques to systematically construct a generally structured CG, to parameterize this model, to derive its joint probability distribution, to perform joint parameter learning, and to perform probabilistic inference in this model. To demonstrate the utility of such an extended CG, we apply it to two challenging image and video analysis problems: human activity recognition and image segmentation. The experimental results show improved performance of the extended CG model over the conventional directed or undirected PGMs. This study demonstrates the promise of the extended CG for effective modeling and inference of complex real-world problems.
Phantom-based interactive simulation system for dental treatment training.
Sae-Kee, Bundit; Riener, Robert; Frey, Martin; Pröll, Thomas; Burgkart, Rainer
2004-01-01
In this paper, we propose a new interactive simulation system for dental treatment training. The system comprises a virtual reality environment and a force-torque measuring device to enhance the capabilities of a passive phantom of tooth anatomy in dental treatment training processes. The measuring device is connected to the phantom, and provides essential input data for generating the graphic animations of physical behaviors such as drilling and bleeding. The animation methods of those physical behaviors are also presented. This system is not only able to enhance interactivity and accessibility of the training system compared to conventional methods but it also provides possibilities of recording, evaluating, and verifying the training results.
Mesoscale and severe storms (Mass) data management and analysis system
NASA Technical Reports Server (NTRS)
Hickey, J. S.; Karitani, S.; Dickerson, M.
1984-01-01
Progress on the Mesoscale and Severe Storms (MASS) data management and analysis system is described. An interactive atmospheric data base management software package to convert four types of data (Sounding, Single Level, Grid, Image) into standard random access formats is implemented and integrated with the MASS AVE80 Series general purpose plotting and graphics display data analysis software package. An interactive analysis and display graphics software package (AVE80) to analyze large volumes of conventional and satellite derived meteorological data is enhanced to provide imaging/color graphics display utilizing color video hardware integrated into the MASS computer system. Local and remote smart-terminal capability is provided by installing APPLE III computer systems within individual scientist offices and integrated with the MASS system, thus providing color video display, graphics, and characters display of the four data types.
Ito, Hiroshi; Yokoi, Takashi; Ikoma, Yoko; Shidahara, Miho; Seki, Chie; Naganawa, Mika; Takahashi, Hidehiko; Takano, Harumasa; Kimura, Yuichi; Ichise, Masanori; Suhara, Tetsuya
2010-01-01
In positron emission tomography (PET) studies with radioligands for neuroreceptors, tracer kinetics have been described by the standard two-tissue compartment model that includes the compartments of nondisplaceable binding and specific binding to receptors. In the present study, we have developed a new graphic plot analysis to determine the total distribution volume (V(T)) and nondisplaceable distribution volume (V(ND)) independently, and therefore the binding potential (BP(ND)). In this plot, Y(t) is the ratio of brain tissue activity to time-integrated arterial input function, and X(t) is the ratio of time-integrated brain tissue activity to time-integrated arterial input function. The x-intercept of linear regression of the plots for early phase represents V(ND), and the x-intercept of linear regression of the plots for delayed phase after the equilibrium time represents V(T). BP(ND) can be calculated by BP(ND)=V(T)/V(ND)-1. Dynamic PET scanning with measurement of arterial input function was performed on six healthy men after intravenous rapid bolus injection of [(11)C]FLB457. The plot yielded a curve in regions with specific binding while it yielded a straight line through all plot data in regions with no specific binding. V(ND), V(T), and BP(ND) values calculated by the present method were in good agreement with those by conventional non-linear least-squares fitting procedure. This method can be used to distinguish graphically whether the radioligand binding includes specific binding or not.
Accelerated Adaptive MGS Phase Retrieval
NASA Technical Reports Server (NTRS)
Lam, Raymond K.; Ohara, Catherine M.; Green, Joseph J.; Bikkannavar, Siddarayappa A.; Basinger, Scott A.; Redding, David C.; Shi, Fang
2011-01-01
The Modified Gerchberg-Saxton (MGS) algorithm is an image-based wavefront-sensing method that can turn any science instrument focal plane into a wavefront sensor. MGS characterizes optical systems by estimating the wavefront errors in the exit pupil using only intensity images of a star or other point source of light. This innovative implementation of MGS significantly accelerates the MGS phase retrieval algorithm by using stream-processing hardware on conventional graphics cards. Stream processing is a relatively new, yet powerful, paradigm to allow parallel processing of certain applications that apply single instructions to multiple data (SIMD). These stream processors are designed specifically to support large-scale parallel computing on a single graphics chip. Computationally intensive algorithms, such as the Fast Fourier Transform (FFT), are particularly well suited for this computing environment. This high-speed version of MGS exploits commercially available hardware to accomplish the same objective in a fraction of the original time. The exploit involves performing matrix calculations in nVidia graphic cards. The graphical processor unit (GPU) is hardware that is specialized for computationally intensive, highly parallel computation. From the software perspective, a parallel programming model is used, called CUDA, to transparently scale multicore parallelism in hardware. This technology gives computationally intensive applications access to the processing power of the nVidia GPUs through a C/C++ programming interface. The AAMGS (Accelerated Adaptive MGS) software takes advantage of these advanced technologies, to accelerate the optical phase error characterization. With a single PC that contains four nVidia GTX-280 graphic cards, the new implementation can process four images simultaneously to produce a JWST (James Webb Space Telescope) wavefront measurement 60 times faster than the previous code.
Maurer, S A; Kussmann, J; Ochsenfeld, C
2014-08-07
We present a low-prefactor, cubically scaling scaled-opposite-spin second-order Møller-Plesset perturbation theory (SOS-MP2) method which is highly suitable for massively parallel architectures like graphics processing units (GPU). The scaling is reduced from O(N⁵) to O(N³) by a reformulation of the MP2-expression in the atomic orbital basis via Laplace transformation and the resolution-of-the-identity (RI) approximation of the integrals in combination with efficient sparse algebra for the 3-center integral transformation. In contrast to previous works that employ GPUs for post Hartree-Fock calculations, we do not simply employ GPU-based linear algebra libraries to accelerate the conventional algorithm. Instead, our reformulation allows to replace the rate-determining contraction step with a modified J-engine algorithm, that has been proven to be highly efficient on GPUs. Thus, our SOS-MP2 scheme enables us to treat large molecular systems in an accurate and efficient manner on a single GPU-server.
Acceleration of FDTD mode solver by high-performance computing techniques.
Han, Lin; Xi, Yanping; Huang, Wei-Ping
2010-06-21
A two-dimensional (2D) compact finite-difference time-domain (FDTD) mode solver is developed based on wave equation formalism in combination with the matrix pencil method (MPM). The method is validated for calculation of both real guided and complex leaky modes of typical optical waveguides against the bench-mark finite-difference (FD) eigen mode solver. By taking advantage of the inherent parallel nature of the FDTD algorithm, the mode solver is implemented on graphics processing units (GPUs) using the compute unified device architecture (CUDA). It is demonstrated that the high-performance computing technique leads to significant acceleration of the FDTD mode solver with more than 30 times improvement in computational efficiency in comparison with the conventional FDTD mode solver running on CPU of a standard desktop computer. The computational efficiency of the accelerated FDTD method is in the same order of magnitude of the standard finite-difference eigen mode solver and yet require much less memory (e.g., less than 10%). Therefore, the new method may serve as an efficient, accurate and robust tool for mode calculation of optical waveguides even when the conventional eigen value mode solvers are no longer applicable due to memory limitation.
Thin-plate spline (TPS) graphical analysis of the mandible on cephalometric radiographs.
Chang, H P; Liu, P H; Chang, H F; Chang, C H
2002-03-01
We describe two cases of Class III malocclusion with and without orthodontic treatment. A thin-plate spline (TPS) analysis of lateral cephalometric radiographs was used to visualize transformations of the mandible. The actual sites of mandibular skeletal change are not detectable with conventional cephalometric analysis. These case analyses indicate that specific patterns of mandibular transformation are associated with Class III malocclusion with or without orthopaedic therapy, and visualization of these deformations is feasible using TPS graphical analysis.
Molecular Monte Carlo Simulations Using Graphics Processing Units: To Waste Recycle or Not?
Kim, Jihan; Rodgers, Jocelyn M; Athènes, Manuel; Smit, Berend
2011-10-11
In the waste recycling Monte Carlo (WRMC) algorithm, (1) multiple trial states may be simultaneously generated and utilized during Monte Carlo moves to improve the statistical accuracy of the simulations, suggesting that such an algorithm may be well posed for implementation in parallel on graphics processing units (GPUs). In this paper, we implement two waste recycling Monte Carlo algorithms in CUDA (Compute Unified Device Architecture) using uniformly distributed random trial states and trial states based on displacement random-walk steps, and we test the methods on a methane-zeolite MFI framework system to evaluate their utility. We discuss the specific implementation details of the waste recycling GPU algorithm and compare the methods to other parallel algorithms optimized for the framework system. We analyze the relationship between the statistical accuracy of our simulations and the CUDA block size to determine the efficient allocation of the GPU hardware resources. We make comparisons between the GPU and the serial CPU Monte Carlo implementations to assess speedup over conventional microprocessors. Finally, we apply our optimized GPU algorithms to the important problem of determining free energy landscapes, in this case for molecular motion through the zeolite LTA.
Wachter, S. Blake; Johnson, Ken; Albert, Robert; Syroid, Noah; Drews, Frank; Westenskow, Dwayne
2006-01-01
Objective Authors developed a picture-graphics display for pulmonary function to present typical respiratory data used in perioperative and intensive care environments. The display utilizes color, shape and emergent alerting to highlight abnormal pulmonary physiology. The display serves as an adjunct to traditional operating room displays and monitors. Design To evaluate the prototype, nineteen clinician volunteers each managed four adverse respiratory events and one normal event using a high-resolution patient simulator which included the new displays (intervention subjects) and traditional displays (control subjects). Between-group comparisons included (i) time to diagnosis and treatment for each adverse respiratory event; (ii) the number of unnecessary treatments during the normal scenario; and (iii) self-reported workload estimates while managing study events. Measurements Two expert anesthesiologists reviewed video-taped transcriptions of the volunteers to determine time to treat and time to diagnosis. Time values were then compared between groups using a Mann-Whitney-U Test. Estimated workload for both groups was assessed using the NASA-TLX and compared between groups using an ANOVA. P-values < 0.05 were considered significant. Results Clinician volunteers detected and treated obstructed endotracheal tubes and intrinsic PEEP problems faster with graphical rather than conventional displays (p < 0.05). During the normal scenario simulation, 3 clinicians using the graphical display, and 5 clinicians using the conventional display gave unnecessary treatments. Clinician-volunteers reported significantly lower subjective workloads using the graphical display for the obstructed endotracheal tube scenario (p < 0.001) and the intrinsic PEEP scenario (p < 0.03). Conclusion Authors conclude that the graphical pulmonary display may serve as a useful adjunct to traditional displays in identifying adverse respiratory events. PMID:16929038
Conversion of KEGG metabolic pathways to SBGN maps including automatic layout
2013-01-01
Background Biologists make frequent use of databases containing large and complex biological networks. One popular database is the Kyoto Encyclopedia of Genes and Genomes (KEGG) which uses its own graphical representation and manual layout for pathways. While some general drawing conventions exist for biological networks, arbitrary graphical representations are very common. Recently, a new standard has been established for displaying biological processes, the Systems Biology Graphical Notation (SBGN), which aims to unify the look of such maps. Ideally, online repositories such as KEGG would automatically provide networks in a variety of notations including SBGN. Unfortunately, this is non‐trivial, since converting between notations may add, remove or otherwise alter map elements so that the existing layout cannot be simply reused. Results Here we describe a methodology for automatic translation of KEGG metabolic pathways into the SBGN format. We infer important properties of the KEGG layout and treat these as layout constraints that are maintained during the conversion to SBGN maps. Conclusions This allows for the drawing and layout conventions of SBGN to be followed while creating maps that are still recognizably the original KEGG pathways. This article details the steps in this process and provides examples of the final result. PMID:23953132
Real-time traffic sign recognition based on a general purpose GPU and deep-learning.
Lim, Kwangyong; Hong, Yongwon; Choi, Yeongwoo; Byun, Hyeran
2017-01-01
We present a General Purpose Graphics Processing Unit (GPGPU) based real-time traffic sign detection and recognition method that is robust against illumination changes. There have been many approaches to traffic sign recognition in various research fields; however, previous approaches faced several limitations when under low illumination or wide variance of light conditions. To overcome these drawbacks and improve processing speeds, we propose a method that 1) is robust against illumination changes, 2) uses GPGPU-based real-time traffic sign detection, and 3) performs region detecting and recognition using a hierarchical model. This method produces stable results in low illumination environments. Both detection and hierarchical recognition are performed in real-time, and the proposed method achieves 0.97 F1-score on our collective dataset, which uses the Vienna convention traffic rules (Germany and South Korea).
NASA Astrophysics Data System (ADS)
Osada, Masakazu; Kaise, Mitsuru; Ozeki, Takeshi; Tsunakawa, Hirofumi; Tsunakawa, Kiyoshi; Takayanagi, Takashi; Suzuki, Nobuaki; Miwa, Jun; Ohta, Yasuhiko; Kanai, Koichi
1999-07-01
We have proposed a new user interface with workflow customization, implemented and evaluted in Endoscopy Department Mini-PACS that has been introduced and routinely used for two years at Toshiba General Hospital. We have set some task at endoscopy image acquisition units during examinations for two different types of user interfaces and compared performance. One is a command-button based operation using a remote control, and another is that with eight graphic buttons which are displayed on a CRT monitor and can be customized. Results of the two-year study show that mean number of input diagnosis codes per examination with graphic and customized interface is significantly greater than that with conventional interface. Also, mean time to complete one upper gastric endoscopy examination with new user interface is about 17 percent less than that with conventional interface. These result suggest that systems with the visualized and customized operation and feedback encourages physicians to use more functions and to compete tasks more efficiently than systems with conventional command-button based user interfaces.
Braille and Tactile Graphics: Youths with Visual Impairments Share Their Experiences
ERIC Educational Resources Information Center
Rosenblum, L. Penny; Herzberg, Tina S.
2015-01-01
Introduction: Data were collected from youths with visual impairment about their experiences with tactile graphics and braille materials used in mathematics and science classes. Methods: Youths answered questions and explored four tactile graphics made using different production methods. They located specific information on each graphic and shared…
Building Regression Models: The Importance of Graphics.
ERIC Educational Resources Information Center
Dunn, Richard
1989-01-01
Points out reasons for using graphical methods to teach simple and multiple regression analysis. Argues that a graphically oriented approach has considerable pedagogic advantages in the exposition of simple and multiple regression. Shows that graphical methods may play a central role in the process of building regression models. (Author/LS)
SU-E-T-07: 4DCT Robust Optimization for Esophageal Cancer Using Intensity Modulated Proton Therapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liao, L; Department of Industrial Engineering, University of Houston, Houston, TX; Yu, J
2015-06-15
Purpose: To develop a 4DCT robust optimization method to reduce the dosimetric impact from respiratory motion in intensity modulated proton therapy (IMPT) for esophageal cancer. Methods: Four esophageal cancer patients were selected for this study. The different phases of CT from a set of 4DCT were incorporated into the worst-case dose distribution robust optimization algorithm. 4DCT robust treatment plans were designed and compared with the conventional non-robust plans. Result doses were calculated on the average and maximum inhale/exhale phases of 4DCT. Dose volume histogram (DVH) band graphic and ΔD95%, ΔD98%, ΔD5%, ΔD2% of CTV between different phases were used tomore » evaluate the robustness of the plans. Results: Compare to the IMPT plans optimized using conventional methods, the 4DCT robust IMPT plans can achieve the same quality in nominal cases, while yield a better robustness to breathing motion. The mean ΔD95%, ΔD98%, ΔD5% and ΔD2% of CTV are 6%, 3.2%, 0.9% and 1% for the robustly optimized plans vs. 16.2%, 11.8%, 1.6% and 3.3% from the conventional non-robust plans. Conclusion: A 4DCT robust optimization method was proposed for esophageal cancer using IMPT. We demonstrate that the 4DCT robust optimization can mitigate the dose deviation caused by the diaphragm motion.« less
Data structures and organisation: Special problems in scientific applications
NASA Astrophysics Data System (ADS)
Read, Brian J.
1989-12-01
In this paper we discuss and offer answers to the following questions: What, really, are the benifits of databases in physics? Are scientific databases essentially different from conventional ones? What are the drawbacks of a commercial database management system for use with scientific data? Do they outweigh the advantages? Do databases systems have adequate graphics facilities, or is a separate graphics package necessary? SQL as a standard language has deficiencies, but what are they for scientific data in particular? Indeed, is the relational model appropriate anyway? Or, should we turn to object oriented databases?
Faraji, Hakim; Helalizadeh, Masoumeh; Kordi, Mohammad Reza
2018-01-01
A rapid, simple, and sensitive approach to the analysis of trihalomethanes (THMs) in swimming pool water samples has been developed. The main goal of this study was to overcome or to improve the shortcomings of conventional dispersive liquid-liquid microextraction (DLLME) and to maximize the realization of green analytical chemistry principles. The method involves a simple vortex-assisted microextraction step, in the absence of the dispersive solvent, followed by salting-out effect for the elimination of the centrifugation step. A bell-shaped device and a solidifiable solvent were used to simplify the extraction solvent collection after phase separation. Optimization of the independent variables was performed by using chemometric methods in three steps. The method was statistically validated based on authentic guidance documents. The completion time for extraction was less than 8 min, and the limits of detection were in the range between 4 and 72 ng L -1 . Using this method, good linearity and precision were achieved. The results of THMs determination in different real samples showed that in some cases the concentration of total THMs was more than threshold values of THMs determined by accredited healthcare organizations. This method indicated satisfactory analytical figures of merit. Graphical Abstract A novel green microextraction technique for overcoming the challenges of conventional DLLME. The proposed procedure complies with the principles of green/sustainable analytical chemistry, comprising decreasing the sample size, making easy automation of the process, reducing organic waste, diminishing energy consumption, replacing toxic reagents with safer reagents, and enhancing operator safety.
Baker, Arthur W; Haridy, Salah; Salem, Joseph; Ilieş, Iulian; Ergai, Awatef O; Samareh, Aven; Andrianas, Nicholas; Benneyan, James C; Sexton, Daniel J; Anderson, Deverick J
2017-11-24
Traditional strategies for surveillance of surgical site infections (SSI) have multiple limitations, including delayed and incomplete outbreak detection. Statistical process control (SPC) methods address these deficiencies by combining longitudinal analysis with graphical presentation of data. We performed a pilot study within a large network of community hospitals to evaluate performance of SPC methods for detecting SSI outbreaks. We applied conventional Shewhart and exponentially weighted moving average (EWMA) SPC charts to 10 previously investigated SSI outbreaks that occurred from 2003 to 2013. We compared the results of SPC surveillance to the results of traditional SSI surveillance methods. Then, we analysed the performance of modified SPC charts constructed with different outbreak detection rules, EWMA smoothing factors and baseline SSI rate calculations. Conventional Shewhart and EWMA SPC charts both detected 8 of the 10 SSI outbreaks analysed, in each case prior to the date of traditional detection. Among detected outbreaks, conventional Shewhart chart detection occurred a median of 12 months prior to outbreak onset and 22 months prior to traditional detection. Conventional EWMA chart detection occurred a median of 7 months prior to outbreak onset and 14 months prior to traditional detection. Modified Shewhart and EWMA charts additionally detected several outbreaks earlier than conventional SPC charts. Shewhart and SPC charts had low false-positive rates when used to analyse separate control hospital SSI data. Our findings illustrate the potential usefulness and feasibility of real-time SPC surveillance of SSI to rapidly identify outbreaks and improve patient safety. Further study is needed to optimise SPC chart selection and calculation, statistical outbreak detection rules and the process for reacting to signals of potential outbreaks. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
ERIC Educational Resources Information Center
Association for the Development of Computer-based Instructional Systems.
The second of three volumes of papers presented at the 1979 ADCIS convention, this collection includes 37 papers presented to four special interest groups--computer based training, deaf education, elementary/secondary education/junior colleges, and health education. The eight papers on computer based training describe computer graphics, computer…
Conci, Ricardo Augusto; Tomazi, Flavio Henrique Silveira; Noritomi, Pedro Yoshito; da Silva, Jorge Vicente Lopes; Fritscher, Guilherme Genehr; Heitz, Claiton
2015-07-01
To compare the mechanical stress on the mandibular condyle after the reduction and fixation of mandibular condylar fractures using the neck screw and 2 other conventional techniques according to 3-dimensional finite element analysis. A 3-dimensional finite element model of a mandible was created and graphically simulated on a computer screen. The model was fixed with 3 different techniques: a 2.0-mm plate with 4 screws, 2 plates (1 1.5-mm plate and 1 2.0-mm plate) with 4 screws, and a neck screw. Loads were applied that simulated muscular action, with restrictions of the upper movements of the mandible, differentiation of the cortical and medullary bone, and the virtual "folds" of the plates and screws so that they could adjust to the condylar surface. Afterward, the data were exported for graphic visualization of the results and quantitative analysis was performed. The 2-plate technique exhibited better stability in regard to displacement of fractures, deformity of the synthesis materials, and minimum and maximum tension values. The results with the neck screw were satisfactory and were similar to those found when a miniplate was used. Although the study shows that 2 isolated plates yielded better results compared with the other groups using other fixation systems and methods, the neck screw could be an option for condylar fracture reduction. Copyright © 2015 American Association of Oral and Maxillofacial Surgeons. Published by Elsevier Inc. All rights reserved.
Physics of the zero-
NASA Astrophysics Data System (ADS)
Zhou, Lei; Song, Zhengyong; Huang, Xueqin; Chan, C. T.
2012-12-01
A short overview is presented on the research works related to the zero-
Real-time traffic sign recognition based on a general purpose GPU and deep-learning
Hong, Yongwon; Choi, Yeongwoo; Byun, Hyeran
2017-01-01
We present a General Purpose Graphics Processing Unit (GPGPU) based real-time traffic sign detection and recognition method that is robust against illumination changes. There have been many approaches to traffic sign recognition in various research fields; however, previous approaches faced several limitations when under low illumination or wide variance of light conditions. To overcome these drawbacks and improve processing speeds, we propose a method that 1) is robust against illumination changes, 2) uses GPGPU-based real-time traffic sign detection, and 3) performs region detecting and recognition using a hierarchical model. This method produces stable results in low illumination environments. Both detection and hierarchical recognition are performed in real-time, and the proposed method achieves 0.97 F1-score on our collective dataset, which uses the Vienna convention traffic rules (Germany and South Korea). PMID:28264011
NASA Astrophysics Data System (ADS)
Ehricke, Hans-Heino; Daiber, Gerhard; Sonntag, Ralf; Strasser, Wolfgang; Lochner, Mathias; Rudi, Lothar S.; Lorenz, Walter J.
1992-09-01
In stereotactic treatment planning the spatial relationships between a variety of objects has to be taken into account in order to avoid destruction of vital brain structures and rupture of vasculature. The visualization of these highly complex relations may be supported by 3-D computer graphics methods. In this context the three-dimensional display of the intracranial vascular tree and additional objects, such as neuroanatomy, pathology, stereotactic devices, or isodose surfaces, is of high clinical value. We report an advanced rendering method for a depth-enhanced maximum intensity projection from magnetic resonance angiography (MRA) and a walk-through approach to the analysis of MRA volume data. Furthermore, various methods for a multiple-object 3-D rendering in stereotaxy are discussed. The development of advanced applications in medical imaging can hardly be successful if image acquisition problems are disregarded. We put particular emphasis on the use of conventional MRI and MRA for stereotactic guidance. The problem of MR distortion is discussed and a novel three- dimensional approach to the quantification and correction of the distortion patterns is presented. Our results suggest that the sole use of MR for stereotactic guidance is highly practical. The true three-dimensionality of the acquired datasets opens up new perspectives to stereotactic treatment planning. For the first time it is possible now to integrate all the necessary information into 3-D scenes, thus enabling an interactive 3-D planning.
NASA Astrophysics Data System (ADS)
Vucinic, Dean; Deen, Danny; Oanta, Emil; Batarilo, Zvonimir; Lacor, Chris
This paper focuses on visualization and manipulation of graphical content in distributed network environments. The developed graphical middleware and 3D desktop prototypes were specialized for situational awareness. This research was done in the LArge Scale COllaborative decision support Technology (LASCOT) project, which explored and combined software technologies to support human-centred decision support system for crisis management (earthquake, tsunami, flooding, airplane or oil-tanker incidents, chemical, radio-active or other pollutants spreading, etc.). The performed state-of-the-art review did not identify any publicly available large scale distributed application of this kind. Existing proprietary solutions rely on the conventional technologies and 2D representations. Our challenge was to apply the "latest" available technologies, such Java3D, X3D and SOAP, compatible with average computer graphics hardware. The selected technologies are integrated and we demonstrate: the flow of data, which originates from heterogeneous data sources; interoperability across different operating systems and 3D visual representations to enhance the end-users interactions.
Von Dreele, Robert
2017-08-29
One of the goals in developing GSAS-II was to expand from the capabilities of the original General Structure Analysis System (GSAS) which largely encompassed just structure refinement and post refinement analysis. GSAS-II has been written almost entirely in Python loaded with graphics, GUI and mathematical packages (matplotlib, pyOpenGL, wxpython, numpy and scipy). Thus, GSAS-II has a fully developed modern GUI as well as extensive graphical display of data and results. However, the structure and operation of Python has required new approaches to many of the algorithms used in crystal structure analysis. The extensions beyond GSAS include image calibration/integration as wellmore » as peak fitting and unit cell indexing for powder data which are precursors for structure solution. Structure solution within GSAS-II begins with either Pawley or LeBail extracted structure factors from powder data or those measured in a single crystal experiment. Both charge flipping and Monte Carlo-Simulated Annealing techniques are available; the former can be applied to (3+1) incommensurate structures as well as conventional 3D structures.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Maurer, S. A.; Kussmann, J.; Ochsenfeld, C., E-mail: Christian.Ochsenfeld@cup.uni-muenchen.de
2014-08-07
We present a low-prefactor, cubically scaling scaled-opposite-spin second-order Møller-Plesset perturbation theory (SOS-MP2) method which is highly suitable for massively parallel architectures like graphics processing units (GPU). The scaling is reduced from O(N{sup 5}) to O(N{sup 3}) by a reformulation of the MP2-expression in the atomic orbital basis via Laplace transformation and the resolution-of-the-identity (RI) approximation of the integrals in combination with efficient sparse algebra for the 3-center integral transformation. In contrast to previous works that employ GPUs for post Hartree-Fock calculations, we do not simply employ GPU-based linear algebra libraries to accelerate the conventional algorithm. Instead, our reformulation allows tomore » replace the rate-determining contraction step with a modified J-engine algorithm, that has been proven to be highly efficient on GPUs. Thus, our SOS-MP2 scheme enables us to treat large molecular systems in an accurate and efficient manner on a single GPU-server.« less
Ngounou, Guy Merlin; Kom, Martin
2014-12-01
In this paper we present an instrumentation amplifier with discrete elements and optimized noise for the amplification of very low signals. In amplifying signals of very weak amplitude, the noise can completely absorb these signals if the used amplifier does not present the optimal guarantee to minimize the noise. Based on related research and re-viewing of recent patents Journal of Medical Systems, 30:205-209, 2006, we suggest an approach of noise reduction in amplification much more thoroughly than re-viewing of recent patents and we deduce from it the general criteria necessary and essential to achieve this optimization. The comparison of these criteria with the provisions adopted in practice leads to the inadequacy of conventional amplifiers for effective noise reduction. The amplifier we propose is an instrumentation amplifier with active negative feedback and optimized noise for the amplification of signals with very low amplitude. The application of this method in the case of electro cardio graphic signals (ECG) provides simulation results fully in line with forecasts.
Geometrical and Graphical Solutions of Quadratic Equations.
ERIC Educational Resources Information Center
Hornsby, E. John, Jr.
1990-01-01
Presented are several geometrical and graphical methods of solving quadratic equations. Discussed are Greek origins, Carlyle's method, von Staudt's method, fixed graph methods and imaginary solutions. (CW)
Kim, Jane Paik; Roberts, Laura Weiss
Empirical ethics inquiry works from the notion that stakeholder perspectives are necessary for gauging the ethical acceptability of human studies and assuring that research aligns with societal expectations. Although common, studies involving different populations often entail comparisons of trends that problematize the interpretation of results. Using graphical model selection - a technique aimed at transcending limitations of conventional methods - this report presents data on the ethics of clinical research with two objectives: (1) to display the patterns of views held by ill and healthy individuals in clinical research as a test of the study's original hypothesis and (2) to introduce graphical model selection as a key analytic tool for ethics research. In this IRB-approved, NIH-funded project, data were collected from 60 mentally ill and 43 physically ill clinical research protocol volunteers, 47 healthy protocol-consented participants, and 29 healthy individuals without research protocol experience. Respondents were queried on the ethical acceptability of research involving people with mental and physical illness (i.e., cancer, HIV, depression, schizophrenia, and post-traumatic stress disorder) and non-illness related sources of vulnerability (e.g., age, class, gender, ethnicity). Using a statistical algorithm, we selected graphical models to display interrelationships among responses to questions. Both mentally and physically ill protocol volunteers revealed a high degree of connectivity among ethically-salient perspectives. Healthy participants, irrespective of research protocol experience, revealed patterns of views that were not highly connected. Between ill and healthy protocol participants, the pattern of views is vastly different. Experience with illness was tied to dense connectivity, whereas healthy individuals expressed views with sparse connections. In offering a nuanced perspective on the interrelation of ethically relevant responses, graphical model selection has the potential to bring new insights to the field of ethics.
Antinomies of Semiotics in Graphic Design
ERIC Educational Resources Information Center
Storkerson, Peter
2010-01-01
The following paper assesses the roles played by semiotics in graphic design and in graphic design education, which both reflects and shapes practice. It identifies a series of factors; graphic design education methods and culture; semiotic theories themselves and their application to graphic design; the two wings of Peircian semiotics and…
2010-01-01
Background The pathology report of radical prostatectomy specimens plays an important role in clinical decisions and the prognostic evaluation in Prostate Cancer (PCa). The anatomical schema is a helpful tool to document PCa extension for clinical and research purposes. To achieve electronic documentation and analysis, an appropriate documentation model for anatomical schemas is needed. For this purpose we developed cMDX. Methods The document architecture of cMDX was designed according to Open Packaging Conventions by separating the whole data into template data and patient data. Analogue custom XML elements were considered to harmonize the graphical representation (e.g. tumour extension) with the textual data (e.g. histological patterns). The graphical documentation was based on the four-layer visualization model that forms the interaction between different custom XML elements. Sensible personal data were encrypted with a 256-bit cryptographic algorithm to avoid misuse. In order to assess the clinical value, we retrospectively analysed the tumour extension in 255 patients after radical prostatectomy. Results The pathology report with cMDX can represent pathological findings of the prostate in schematic styles. Such reports can be integrated into the hospital information system. "cMDX" documents can be converted into different data formats like text, graphics and PDF. Supplementary tools like cMDX Editor and an analyser tool were implemented. The graphical analysis of 255 prostatectomy specimens showed that PCa were mostly localized in the peripheral zone (Mean: 73% ± 25). 54% of PCa showed a multifocal growth pattern. Conclusions cMDX can be used for routine histopathological reporting of radical prostatectomy specimens and provide data for scientific analysis. PMID:21078179
GEMPAK5. Part 2: GEMPLT programmer's guide, version 5.0
NASA Technical Reports Server (NTRS)
Desjardins, Mary L.; Brill, Keith F.; Schotz, Steven S.
1991-01-01
GEMPAK is a general meteorological software package used to analyze and display conventional meteorological data as well as satellite derived parameters. The GEMPAK Programmer's Guide describes the subroutines which can be used in the GEMPAK graphics and transformation subsystem, GEMPLT.
Spatial Visualization by Isometric View
ERIC Educational Resources Information Center
Yue, Jianping
2007-01-01
Spatial visualization is a fundamental skill in technical graphics and engineering designs. From conventional multiview drawing to modern solid modeling using computer-aided design, visualization skills have always been essential for representing three-dimensional objects and assemblies. Researchers have developed various types of tests to measure…
Automatic superposition of drug molecules based on their common receptor site
NASA Astrophysics Data System (ADS)
Kato, Yuichi; Inoue, Atsushi; Yamada, Miho; Tomioka, Nobuo; Itai, Akiko
1992-10-01
We have prevously developed a new rational method for superposing molecules in terms of submolecular physical and chemical properties, but not in terms of atom positions or chemical structures as has been done in the conventional methods. The program was originally developed for interactive use on a three-dimensional graphic display, providing goodness-of-fit indices on molecular shape, hydrogen bonds, electrostatic interactions and others. Here, we report a new unbiased searching method for the best superposition of molecules, covering all the superposing modes and conformational freedom, as an additional function of the program. The function is based on a novel least-squares method which superposes the expected positions and orientations of hydrogen bonding partners in the receptor that are deduced from both molecules. The method not only gives reliability and reproducibility to the result of the superposition, but also allows us to save labor and time. It is demonstrated that this method is very efficient for finding the correct superposing mode in such systems where hydrogen bonds play important roles.
Research principles and the construction of mnemonic diagrams
NASA Technical Reports Server (NTRS)
Venda, V. F.; Mitkin, A. A.
1973-01-01
Mnemonic diagrams are defined as a variety of information display devices, the essential element of which is conventional graphical presentation of technological or functional-operational links in a controlled system or object. Graphically displaying the operational structure of an object, the interd dependence between different parameters, and the interdependence between indicators and control organs, the mneomonic diagram reduces the load on the operator's memory and facilitates perception and reprocessing of information and decision making, while at the same time playing the role of visual support to the information activity of the operator. The types of mnemonic diagrams are listed.
A standard format and a graphical user interface for spin system specification.
Biternas, A G; Charnock, G T P; Kuprov, Ilya
2014-03-01
We introduce a simple and general XML format for spin system description that is the result of extensive consultations within Magnetic Resonance community and unifies under one roof all major existing spin interaction specification conventions. The format is human-readable, easy to edit and easy to parse using standard XML libraries. We also describe a graphical user interface that was designed to facilitate construction and visualization of complicated spin systems. The interface is capable of generating input files for several popular spin dynamics simulation packages. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liscom, W.L.
This book presents a complete graphic and statistical portrait of the dramatic shifts in global energy flows during the 1970s and the resultant transfer of economic and political power from the industrial nations to the oil-producing states. The information was extracted from government-source documents and compiled in a computer data base. Computer graphics were combined with the data base to produce over 400 full-color graphs. The energy commodities covered are oil, natural gas, coal, nuclear, and conventional electric-power generation. Also included are data on hydroelectric and geothermal power, oil shale, tar sands, and other alternative energy sources. 72 references.
Liang, Xiaoyun; Vaughan, David N; Connelly, Alan; Calamante, Fernando
2018-05-01
The conventional way to estimate functional networks is primarily based on Pearson correlation along with classic Fisher Z test. In general, networks are usually calculated at the individual-level and subsequently aggregated to obtain group-level networks. However, such estimated networks are inevitably affected by the inherent large inter-subject variability. A joint graphical model with Stability Selection (JGMSS) method was recently shown to effectively reduce inter-subject variability, mainly caused by confounding variations, by simultaneously estimating individual-level networks from a group. However, its benefits might be compromised when two groups are being compared, given that JGMSS is blinded to other groups when it is applied to estimate networks from a given group. We propose a novel method for robustly estimating networks from two groups by using group-fused multiple graphical-lasso combined with stability selection, named GMGLASS. Specifically, by simultaneously estimating similar within-group networks and between-group difference, it is possible to address inter-subject variability of estimated individual networks inherently related with existing methods such as Fisher Z test, and issues related to JGMSS ignoring between-group information in group comparisons. To evaluate the performance of GMGLASS in terms of a few key network metrics, as well as to compare with JGMSS and Fisher Z test, they are applied to both simulated and in vivo data. As a method aiming for group comparison studies, our study involves two groups for each case, i.e., normal control and patient groups; for in vivo data, we focus on a group of patients with right mesial temporal lobe epilepsy.
Chen, Ya-Chen; Hsiao, Tzu-Chien
2018-07-01
Respiratory inductance plethysmography (RIP) sensor is an inexpensive, non-invasive, easy-to-use transducer for collecting respiratory movement data. Studies have reported that the RIP signal's amplitude and frequency can be used to discriminate respiratory diseases. However, with the conventional approach of RIP data analysis, respiratory muscle effort cannot be estimated. In this paper, the estimation of the respiratory muscle effort through RIP signal was proposed. A complementary ensemble empirical mode decomposition method was used, to extract hidden signals from the RIP signals based on the frequency bands of the activities of different respiratory muscles. To validate the proposed method, an experiment to collect subjects' RIP signal under thoracic breathing (TB) and abdominal breathing (AB) was conducted. The experimental results for both the TB and AB indicate that the proposed method can be used to loosely estimate the activities of thoracic muscles, abdominal muscles, and diaphragm. Graphical abstract ᅟ.
Stone, B N; Griesinger, G L; Modelevsky, J L
1984-01-01
We describe an interactive computational tool, PLASMAP, which allows the user to electronically store, retrieve, and display circular restriction maps. PLASMAP permits users to construct libraries of plasmid restriction maps as a set of files which may be edited in the laboratory at any time. The display feature of PLASMAP quickly generates device-independent, artist-quality, full-color or monochrome, hard copies or CRT screens of complex, conventional circular restriction maps. PMID:6320096
Determining Normal-Distribution Tolerance Bounds Graphically
NASA Technical Reports Server (NTRS)
Mezzacappa, M. A.
1983-01-01
Graphical method requires calculations and table lookup. Distribution established from only three points: mean upper and lower confidence bounds and lower confidence bound of standard deviation. Method requires only few calculations with simple equations. Graphical procedure establishes best-fit line for measured data and bounds for selected confidence level and any distribution percentile.
DOT National Transportation Integrated Search
1981-01-01
This report describes a method for locating historic site information using a computer graphics program. If adopted for use by the Virginia Department of Highways and Transportation, this method should significantly reduce the time now required to de...
DOT National Transportation Integrated Search
2002-04-01
This document catalogs the symbols presented with the various interfaces used by Federal Aviation Administration Airway Facilities specialists. It includes a high-level overview of each system and the symbols and coding conventions used. These data w...
APEX/SPIN: a free test platform to measure speech intelligibility.
Francart, Tom; Hofmann, Michael; Vanthornhout, Jonas; Van Deun, Lieselot; van Wieringen, Astrid; Wouters, Jan
2017-02-01
Measuring speech intelligibility in quiet and noise is important in clinical practice and research. An easy-to-use free software platform for conducting speech tests is presented, called APEX/SPIN. The APEX/SPIN platform allows the use of any speech material in combination with any noise. A graphical user interface provides control over a large range of parameters, such as number of loudspeakers, signal-to-noise ratio and parameters of the procedure. An easy-to-use graphical interface is provided for calibration and storage of calibration values. To validate the platform, perception of words in quiet and sentences in noise were measured both with APEX/SPIN and with an audiometer and CD player, which is a conventional setup in current clinical practice. Five normal-hearing listeners participated in the experimental evaluation. Speech perception results were similar for the APEX/SPIN platform and conventional procedures. APEX/SPIN is a freely available and open source platform that allows the administration of all kinds of custom speech perception tests and procedures.
NASA Astrophysics Data System (ADS)
Shen, Yanfeng; Cesnik, Carlos E. S.
2016-04-01
This paper presents a parallelized modeling technique for the efficient simulation of nonlinear ultrasonics introduced by the wave interaction with fatigue cracks. The elastodynamic wave equations with contact effects are formulated using an explicit Local Interaction Simulation Approach (LISA). The LISA formulation is extended to capture the contact-impact phenomena during the wave damage interaction based on the penalty method. A Coulomb friction model is integrated into the computation procedure to capture the stick-slip contact shear motion. The LISA procedure is coded using the Compute Unified Device Architecture (CUDA), which enables the highly parallelized supercomputing on powerful graphic cards. Both the explicit contact formulation and the parallel feature facilitates LISA's superb computational efficiency over the conventional finite element method (FEM). The theoretical formulations based on the penalty method is introduced and a guideline for the proper choice of the contact stiffness is given. The convergence behavior of the solution under various contact stiffness values is examined. A numerical benchmark problem is used to investigate the new LISA formulation and results are compared with a conventional contact finite element solution. Various nonlinear ultrasonic phenomena are successfully captured using this contact LISA formulation, including the generation of nonlinear higher harmonic responses. Nonlinear mode conversion of guided waves at fatigue cracks is also studied.
Paalme, Toomas; Vilbaste, Allan; Kevvai, Kaspar; Nisamedtinov, Ildar; Hälvin-Tanilas, Kristel
2017-11-01
Standardized analytical methods, where each B vitamin is extracted from a given sample individually using separate procedures, typically ensure that the extraction conditions provide the maximum recovery of each vitamin. However, in the human gastrointestinal tract (GIT), the extraction conditions are the same for all vitamins. Here, we present an analytically feasible extraction protocol that simulates conditions in the GIT and provides a measure of the content of bioavailable vitamins using LC-MS stable isotope dilution assay. The results show that the activities of both human gastric and duodenal juices were insufficient to liberate absorbable vitamers (AV) from pure cofactors. The use of an intestinal brush border membrane (IBBM) fraction derived from the mucosal tissue of porcine small intestine ensured at least 70% AV recovery. The rate of AV liberation, however, was strongly dependent on the cofactor, e.g., in the case of NADH, it was magnitudes higher than in the case of thiamine diphosphate. For some vitamins in some food matrices, the use of the IBBM fraction assay resulted in lower values for the content of AV than conventional vitamin determination methods. Conventional methods likely overestimate the actual bioavailability of some vitamins in these cases. Graphical abstract Assessment of bioavailable B vitamin content in food.
Stereoscopic 3D graphics generation
NASA Astrophysics Data System (ADS)
Li, Zhi; Liu, Jianping; Zan, Y.
1997-05-01
Stereoscopic display technology is one of the key techniques of areas such as simulation, multimedia, entertainment, virtual reality, and so on. Moreover, stereoscopic 3D graphics generation is an important part of stereoscopic 3D display system. In this paper, at first, we describe the principle of stereoscopic display and summarize some methods to generate stereoscopic 3D graphics. Secondly, to overcome the problems which came from the methods of user defined models (such as inconvenience, long modifying period and so on), we put forward the vector graphics files defined method. Thus we can design more directly; modify the model simply and easily; generate more conveniently; furthermore, we can make full use of graphics accelerator card and so on. Finally, we discuss the problem of how to speed up the generation.
Message Design Guidelines For Screen-Based Programs.
ERIC Educational Resources Information Center
Rimar, G. I.
1996-01-01
Effective message design for screen-based computer or video instructional programs requires knowledge from many disciplines. Evaluates current conventions and suggests a new set of guidelines for screen-based designers. Discusses screen layout, highlighting and cueing, text font and style, text positioning, color, and graphical user interfaces for…
Communication modality sampling for a toddler with Angelman syndrome.
Hyppa Martin, Jolene; Reichle, Joe; Dimian, Adele; Chen, Mo
2013-10-01
Vocal, gestural, and graphic communication modes were implemented concurrently with a toddler with Angelman syndrome to identify the most efficiently learned communication mode to emphasize in an initial augmentative communication system. Symbols representing preferred objects were introduced in vocal, gestural, and graphic communication modes using an alternating treatment single-subject experimental design. Conventionally accepted prompting strategies were used to teach symbols in each communication mode. Because the learner did not vocally imitate, vocal mode intervention focused on increasing vocal frequency as an initial step. When graphic and gestural mode performances were compared, the learner most accurately produced requests in graphic mode (percentage of nonoverlapping data = 96). Given the lack of success in prompting vocal productions, a comparison between vocal and the other two communication modes was not made. A growing body of evidence suggests that concurrent modality sampling is a promising low-inference, data-driven procedure that can be used to inform selection of a communication mode(s) for initial emphasis with young children. Concurrent modality sampling can guide clinical decisions regarding the allocation of treatment resources to promote success in building an initial communicative repertoire.
Intelligent Visual Input: A Graphical Method for Rapid Entry of Patient-Specific Data
Bergeron, Bryan P.; Greenes, Robert A.
1987-01-01
Intelligent Visual Input (IVI) provides a rapid, graphical method of data entry for both expert system interaction and medical record keeping purposes. Key components of IVI include: a high-resolution graphic display; an interface supportive of rapid selection, i.e., one utilizing a mouse or light pen; algorithm simplification modules; and intelligent graphic algorithm expansion modules. A prototype IVI system, designed to facilitate entry of physical exam findings, is used to illustrates the potential advantages of this approach.
AVE-SESAME program for the REEDA System
NASA Technical Reports Server (NTRS)
Hickey, J. S.
1981-01-01
The REEDA system software was modified and improved to process the AVE-SESAME severe storm data. A random access file system for the AVE storm data was designed, tested, and implemented. The AVE/SESAME software was modified to incorporate the random access file input and to interface with new graphics hardware/software now available on the REEDA system. Software was developed to graphically display the AVE/SESAME data in the convention normally used by severe storm researchers. Software was converted to AVE/SESAME software systems and interfaced with existing graphics hardware/software available on the REEDA System. Software documentation was provided for existing AVE/SESAME programs underlining functional flow charts and interacting questions. All AVE/SESAME data sets in random access format was processed to allow developed software to access the entire AVE/SESAME data base. The existing software was modified to allow for processing of different AVE/SESAME data set types including satellite surface and radar data.
New Computer Simulations of Macular Neural Functioning
NASA Technical Reports Server (NTRS)
Ross, Muriel D.; Doshay, D.; Linton, S.; Parnas, B.; Montgomery, K.; Chimento, T.
1994-01-01
We use high performance graphics workstations and supercomputers to study the functional significance of the three-dimensional (3-D) organization of gravity sensors. These sensors have a prototypic architecture foreshadowing more complex systems. Scaled-down simulations run on a Silicon Graphics workstation and scaled-up, 3-D versions run on a Cray Y-MP supercomputer. A semi-automated method of reconstruction of neural tissue from serial sections studied in a transmission electron microscope has been developed to eliminate tedious conventional photography. The reconstructions use a mesh as a step in generating a neural surface for visualization. Two meshes are required to model calyx surfaces. The meshes are connected and the resulting prisms represent the cytoplasm and the bounding membranes. A finite volume analysis method is employed to simulate voltage changes along the calyx in response to synapse activation on the calyx or on calyceal processes. The finite volume method insures that charge is conserved at the calyx-process junction. These and other models indicate that efferent processes act as voltage followers, and that the morphology of some afferent processes affects their functioning. In a final application, morphological information is symbolically represented in three dimensions in a computer. The possible functioning of the connectivities is tested using mathematical interpretations of physiological parameters taken from the literature. Symbolic, 3-D simulations are in progress to probe the functional significance of the connectivities. This research is expected to advance computer-based studies of macular functioning and of synaptic plasticity.
Kim, Jae-Hong; Kim, Ki-Baek; Kim, Woong-Chul; Kim, Ji-Hwan; Kim, Hae-Young
2014-03-01
This study aimed to evaluate the accuracy and precision of polyurethane (PUT) dental arch models fabricated using a three-dimensional (3D) subtractive rapid prototyping (RP) method with an intraoral scanning technique by comparing linear measurements obtained from PUT models and conventional plaster models. Ten plaster models were duplicated using a selected standard master model and conventional impression, and 10 PUT models were duplicated using the 3D subtractive RP technique with an oral scanner. Six linear measurements were evaluated in terms of x, y, and z-axes using a non-contact white light scanner. Accuracy was assessed using mean differences between two measurements, and precision was examined using four quantitative methods and the Bland-Altman graphical method. Repeatability was evaluated in terms of intra-examiner variability, and reproducibility was assessed in terms of inter-examiner and inter-method variability. The mean difference between plaster models and PUT models ranged from 0.07 mm to 0.33 mm. Relative measurement errors ranged from 2.2% to 7.6% and intraclass correlation coefficients ranged from 0.93 to 0.96, when comparing plaster models and PUT models. The Bland-Altman plot showed good agreement. The accuracy and precision of PUT dental models for evaluating the performance of oral scanner and subtractive RP technology was acceptable. Because of the recent improvements in block material and computerized numeric control milling machines, the subtractive RP method may be a good choice for dental arch models.
Burns, J; Polus, S; Brereton, L; Chilcott, J; Ward, S E; Pfadenhauer, L M; Rehfuess, E A
2018-03-01
We describe a combination of methods for assessing the effectiveness of complex interventions, especially where substantial heterogeneity with regard to the population, intervention, comparison, outcomes, and study design of interest is expected. We applied these methods in a recent systematic review of the effectiveness of reinforced home-based palliative care (rHBPC) interventions, which included home-based care with an additional and explicit component of lay caregiver support. We first summarized the identified evidence, deemed inappropriate for statistical pooling, graphically by creating harvest plots. Although very useful as a tool for summary and presentation of overall effectiveness, such graphical summary approaches may obscure relevant differences between studies. Thus, we then used a gap analysis and conducted expert consultations to look beyond the aggregate level at how the identified evidence of effectiveness may be explained. The goal of these supplemental methods was to step outside of the conventional systematic review and explore this heterogeneity from a broader perspective, based on the experience of palliative care researchers and practitioners. The gap analysis and expert consultations provided valuable input into possible underlying explanations in the evidence, which could be helpful in the further adaptation and testing of existing rHBPC interventions or the development and evaluation of new ones. We feel that such a combination of methods could prove accessible, understandable, and useful in informing decisions and could thus help increase the relevance of systematic reviews to the decision-making process. Copyright © 2017 John Wiley & Sons, Ltd.
ERIC Educational Resources Information Center
Heldman, Bill
2010-01-01
With few exceptions, students interact with technology in one way or another every day. And yet, in most U.S. schools, the term "computer science" (CS) refers only to generic skills classes, such as keyboarding and computer applications. Even most Web programming classes usually teach students only how to use conventional graphical user…
AMPAC as an intelligent communication core for printing process
NASA Astrophysics Data System (ADS)
Mishina, Hiromichi; Yuasa, Tomonori
2000-12-01
The feature analysis of the conventional exchange format used in the field of the graphic arts is performed. It becomes clear from this consideration that most standard defining the transmission format have too strict limitations to adapt for the communication required in the creative and flexible work.
GPU acceleration of particle-in-cell methods
NASA Astrophysics Data System (ADS)
Cowan, Benjamin; Cary, John; Meiser, Dominic
2015-11-01
Graphics processing units (GPUs) have become key components in many supercomputing systems, as they can provide more computations relative to their cost and power consumption than conventional processors. However, to take full advantage of this capability, they require a strict programming model which involves single-instruction multiple-data execution as well as significant constraints on memory accesses. To bring the full power of GPUs to bear on plasma physics problems, we must adapt the computational methods to this new programming model. We have developed a GPU implementation of the particle-in-cell (PIC) method, one of the mainstays of plasma physics simulation. This framework is highly general and enables advanced PIC features such as high order particles and absorbing boundary conditions. The main elements of the PIC loop, including field interpolation and particle deposition, are designed to optimize memory access. We describe the performance of these algorithms and discuss some of the methods used. Work supported by DARPA contract W31P4Q-15-C-0061 (SBIR).
Suppressing multiples using an adaptive multichannel filter based on L1-norm
NASA Astrophysics Data System (ADS)
Shi, Ying; Jing, Hongliang; Zhang, Wenwu; Ning, Dezhi
2017-08-01
Adaptive subtraction is an important link for removing surface-related multiples in the wave equation-based method. In this paper, we propose an adaptive multichannel subtraction method based on the L1-norm. We achieve enhanced compensation for the mismatch between the input seismogram and the predicted multiples in terms of the amplitude, phase, frequency band, and travel time. Unlike the conventional L2-norm, the proposed method does not rely on the assumption that the primary and the multiples are orthogonal, and also takes advantage of the fact that the L1-norm is more robust when dealing with outliers. In addition, we propose a frequency band extension via modulation to reconstruct the high frequencies to compensate for the frequency misalignment. We present a parallel computing scheme to accelerate the subtraction algorithm on graphic processing units (GPUs), which significantly reduces the computational cost. The synthetic and field seismic data tests show that the proposed method effectively suppresses the multiples.
The use of interpractive graphic displays for interpretation of surface design parameters
NASA Technical Reports Server (NTRS)
Talcott, N. A., Jr.
1981-01-01
An interactive computer graphics technique known as the Graphic Display Data method has been developed to provide a convenient means for rapidly interpreting large amounts of surface design data. The display technique should prove valuable in such disciplines as aerodynamic analysis, structural analysis, and experimental data analysis. To demonstrate the system's features, an example is presented of the Graphic Data Display method used as an interpretive tool for radiation equilibrium temperature distributions over the surface of an aerodynamic vehicle. Color graphic displays were also examined as a logical extension of the technique to improve its clarity and to allow the presentation of greater detail in a single display.
Predicted Weather Display and Decision Support Interface for Flight Deck
NASA Technical Reports Server (NTRS)
Johnson, Walter W. (Inventor); Wong, Dominic G. (Inventor); Koteskey, Robert W. (Inventor); Wu, Shu-Chieh (Inventor)
2017-01-01
A system and method for providing visual depictions of a predictive weather forecast for in-route vehicle trajectory planning. The method includes displaying weather information on a graphical display, displaying vehicle position information on the graphical display, selecting a predictive interval, displaying predictive weather information for the predictive interval on the graphical display, and displaying predictive vehicle position information for the predictive interval on the graphical display, such that the predictive vehicle position information is displayed relative to the predictive weather information, for in-route trajectory planning.
GPU-based acceleration of computations in nonlinear finite element deformation analysis.
Mafi, Ramin; Sirouspour, Shahin
2014-03-01
The physics of deformation for biological soft-tissue is best described by nonlinear continuum mechanics-based models, which then can be discretized by the FEM for a numerical solution. However, computational complexity of such models have limited their use in applications requiring real-time or fast response. In this work, we propose a graphic processing unit-based implementation of the FEM using implicit time integration for dynamic nonlinear deformation analysis. This is the most general formulation of the deformation analysis. It is valid for large deformations and strains and can account for material nonlinearities. The data-parallel nature and the intense arithmetic computations of nonlinear FEM equations make it particularly suitable for implementation on a parallel computing platform such as graphic processing unit. In this work, we present and compare two different designs based on the matrix-free and conventional preconditioned conjugate gradients algorithms for solving the FEM equations arising in deformation analysis. The speedup achieved with the proposed parallel implementations of the algorithms will be instrumental in the development of advanced surgical simulators and medical image registration methods involving soft-tissue deformation. Copyright © 2013 John Wiley & Sons, Ltd.
Stable stress‐drop measurements and their variability: Implications for ground‐motion prediction
Hanks, Thomas C.; Baltay, Annemarie S.; Beroza, Gregory C.
2013-01-01
We estimate the arms‐stress drop, Graphic, (Hanks, 1979) using acceleration time records of 59 earthquakes from two earthquake sequences in eastern Honshu, Japan. These acceleration‐based static stress drops compare well to stress drops calculated for the same events by Baltay et al. (2011) using an empirical Green’s function (eGf) approach. This agreement supports the assumption that earthquake acceleration time histories in the bandwidth between the corner frequency and a maximum observed frequency can be considered white, Gaussian, noise. Although the Graphic is computationally simpler than the eGf‐based Graphic‐stress drop, and is used as the “stress parameter” to describe the earthquake source in ground‐motion prediction equations, we find that it only compares well to the Graphic at source‐station distances of ∼20 km or less because there is no consideration of whole‐path anelastic attenuation or scattering. In these circumstances, the correlation between the Graphic and Graphic is strong. Events with high and low stress drops obtained through the eGf method have similarly high and low Graphic. We find that the inter‐event standard deviation of stress drop, for the population of earthquakes considered, is similar for both methods, 0.40 for the Graphic method and 0.42 for the Graphic, in log10 units, provided we apply the ∼20 km distance restriction to Graphic. This indicates that the observed variability is inherent to the source, rather than attributable to uncertainties in stress‐drop estimates
The Development of a Concept Inventory for Engineering Graphics
ERIC Educational Resources Information Center
Nozaki, Steven Yasuto
2017-01-01
Graphics education has been a central component in the engineering course of study. As technology advances, so have the methods in which graphics are expressed and practiced. Though the of media of graphics education have evolved, the concepts of the field remain relatively consistent. Tools are needed to measure students' comprehension,…
Robust Gaussian Graphical Modeling via l1 Penalization
Sun, Hokeun; Li, Hongzhe
2012-01-01
Summary Gaussian graphical models have been widely used as an effective method for studying the conditional independency structure among genes and for constructing genetic networks. However, gene expression data typically have heavier tails or more outlying observations than the standard Gaussian distribution. Such outliers in gene expression data can lead to wrong inference on the dependency structure among the genes. We propose a l1 penalized estimation procedure for the sparse Gaussian graphical models that is robustified against possible outliers. The likelihood function is weighted according to how the observation is deviated, where the deviation of the observation is measured based on its own likelihood. An efficient computational algorithm based on the coordinate gradient descent method is developed to obtain the minimizer of the negative penalized robustified-likelihood, where nonzero elements of the concentration matrix represents the graphical links among the genes. After the graphical structure is obtained, we re-estimate the positive definite concentration matrix using an iterative proportional fitting algorithm. Through simulations, we demonstrate that the proposed robust method performs much better than the graphical Lasso for the Gaussian graphical models in terms of both graph structure selection and estimation when outliers are present. We apply the robust estimation procedure to an analysis of yeast gene expression data and show that the resulting graph has better biological interpretation than that obtained from the graphical Lasso. PMID:23020775
Modeling biochemical transformation processes and information processing with Narrator.
Mandel, Johannes J; Fuss, Hendrik; Palfreyman, Niall M; Dubitzky, Werner
2007-03-27
Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs), which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Narrator is a flexible and intuitive systems biology tool. It is specifically intended for users aiming to construct and simulate dynamic models of biology without recourse to extensive mathematical detail. Its design facilitates mappings to different formal languages and frameworks. The combined set of features makes Narrator unique among tools of its kind. Narrator is implemented as Java software program and available as open-source from http://www.narrator-tool.org.
Modeling biochemical transformation processes and information processing with Narrator
Mandel, Johannes J; Fuß, Hendrik; Palfreyman, Niall M; Dubitzky, Werner
2007-01-01
Background Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs), which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Results Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Conclusion Narrator is a flexible and intuitive systems biology tool. It is specifically intended for users aiming to construct and simulate dynamic models of biology without recourse to extensive mathematical detail. Its design facilitates mappings to different formal languages and frameworks. The combined set of features makes Narrator unique among tools of its kind. Narrator is implemented as Java software program and available as open-source from . PMID:17389034
NASA Astrophysics Data System (ADS)
Griffard, Phyllis Baudoin
1999-11-01
The main research question of this study was: What gaps in biochemical understanding are revealed by a range of university introductory biology students as they work through a critically acclaimed multimedia program on photosynthesis, and what are the corresponding implications for elaboration of the Ausubel-Novak-Gowin Learning Theory (ANG, now Human Constructivism)? Twelve students, mixed for ability, gender and ethnicity, were recruited from two sections of "Bio 101." Before and after instruction in photosynthesis, in-depth clinical interviews were conducted during which participants completed a range of cognitive tasks such as sorting, concept mapping, explaining and predicting. Some tasks involved interacting with a computer simulation of photosynthesis. This study primarily employed qualitative case study and verbal analysis methods. Verbal analysis of the clinical interviews revealed numerous gaps that were categorized into typologies. The two major categories were propositional gaps and processing gaps. Propositional gaps were evident in development of participants' concepts, links and constructs. Significant among these were conceptual distance gaps and continuity of matter gaps. Gaps such as convention gaps and relative significance gaps seem to be due to naivete in the discipline. Processing gaps included gaps in graphic decoding skills and relevant cognitive habits such as self-monitoring and consulting prior knowledge. Although the gaps were easier to detect and isolate with the above-average participants, all participants showed evidence of at least some of these gaps. Since some gaps are not unexpected at all but the highest literacy levels, not all the gaps identified are to be considered deficiencies. The gaps identified support the attention given by ANG theorists to the role of prior knowledge and metacognition as well as the value of graphic organizers in knowledge construction. In addition, this study revealed numerous gaps in graphic decoding, indicating that both direct experience and explicit instruction are needed if students are to "learn how to learn with graphics," especially those graphics central to understanding a computer simulation's representations of structures, inputs, processes and outputs. It is hypothesized that gaps similar to those revealed in this study may be at the root of some alternative conceptions documented in the literature.
Subsonic Wing Optimization for Handling Qualities Using ACSYNT
NASA Technical Reports Server (NTRS)
Soban, Danielle Suzanne
1996-01-01
The capability to accurately and rapidly predict aircraft stability derivatives using one comprehensive analysis tool has been created. The PREDAVOR tool has the following capabilities: rapid estimation of stability derivatives using a vortex lattice method, calculation of a longitudinal handling qualities metric, and inherent methodology to optimize a given aircraft configuration for longitudinal handling qualities, including an intuitive graphical interface. The PREDAVOR tool may be applied to both subsonic and supersonic designs, as well as conventional and unconventional, symmetric and asymmetric configurations. The workstation-based tool uses as its model a three-dimensional model of the configuration generated using a computer aided design (CAD) package. The PREDAVOR tool was applied to a Lear Jet Model 23 and the North American XB-70 Valkyrie.
Causal inference in biology networks with integrated belief propagation.
Chang, Rui; Karr, Jonathan R; Schadt, Eric E
2015-01-01
Inferring causal relationships among molecular and higher order phenotypes is a critical step in elucidating the complexity of living systems. Here we propose a novel method for inferring causality that is no longer constrained by the conditional dependency arguments that limit the ability of statistical causal inference methods to resolve causal relationships within sets of graphical models that are Markov equivalent. Our method utilizes Bayesian belief propagation to infer the responses of perturbation events on molecular traits given a hypothesized graph structure. A distance measure between the inferred response distribution and the observed data is defined to assess the 'fitness' of the hypothesized causal relationships. To test our algorithm, we infer causal relationships within equivalence classes of gene networks in which the form of the functional interactions that are possible are assumed to be nonlinear, given synthetic microarray and RNA sequencing data. We also apply our method to infer causality in real metabolic network with v-structure and feedback loop. We show that our method can recapitulate the causal structure and recover the feedback loop only from steady-state data which conventional method cannot.
Concerning the Video Drift Method to Measure Double Stars
NASA Astrophysics Data System (ADS)
Nugent, Richard L.; Iverson, Ernest W.
2015-05-01
Classical methods to measure position angles and separations of double stars rely on just a few measurements either from visual observations or photographic means. Visual and photographic CCD observations are subject to errors from the following sources: misalignments from eyepiece/camera/barlow lens/micrometer/focal reducers, systematic errors from uncorrected optical distortions, aberrations from the telescope system, camera tilt, magnitude and color effects. Conventional video methods rely on calibration doubles and graphically calculating the east-west direction plus careful choice of select video frames stacked for measurement. Atmospheric motion is one of the larger sources of error in any exposure/measurement method which is on the order of 0.5-1.5. Ideally, if a data set from a short video can be used to derive position angle and separation, with each data set self-calibrating independent of any calibration doubles or star catalogues, this would provide measurements of high systematic accuracy. These aims are achieved by the video drift method first proposed by the authors in 2011. This self calibrating video method automatically analyzes 1,000's of measurements from a short video clip.
Kim, Jae-Hong; Kim, Ki-Baek; Kim, Woong-Chul; Kim, Ji-Hwan
2014-01-01
Objective This study aimed to evaluate the accuracy and precision of polyurethane (PUT) dental arch models fabricated using a three-dimensional (3D) subtractive rapid prototyping (RP) method with an intraoral scanning technique by comparing linear measurements obtained from PUT models and conventional plaster models. Methods Ten plaster models were duplicated using a selected standard master model and conventional impression, and 10 PUT models were duplicated using the 3D subtractive RP technique with an oral scanner. Six linear measurements were evaluated in terms of x, y, and z-axes using a non-contact white light scanner. Accuracy was assessed using mean differences between two measurements, and precision was examined using four quantitative methods and the Bland-Altman graphical method. Repeatability was evaluated in terms of intra-examiner variability, and reproducibility was assessed in terms of inter-examiner and inter-method variability. Results The mean difference between plaster models and PUT models ranged from 0.07 mm to 0.33 mm. Relative measurement errors ranged from 2.2% to 7.6% and intraclass correlation coefficients ranged from 0.93 to 0.96, when comparing plaster models and PUT models. The Bland-Altman plot showed good agreement. Conclusions The accuracy and precision of PUT dental models for evaluating the performance of oral scanner and subtractive RP technology was acceptable. Because of the recent improvements in block material and computerized numeric control milling machines, the subtractive RP method may be a good choice for dental arch models. PMID:24696823
How do the outcomes of the DEKA Arm compare to conventional prostheses?
Resnik, Linda J; Borgia, Matthew L; Acluche, Frantzy; Cancio, Jill M; Latlief, Gail; Sasson, Nicole
2018-01-01
Objectives were to 1) compare self-reported function, dexterity, activity performance, quality of life and community integration of the DEKA Arm to conventional prostheses; and 2) examine differences in outcomes by conventional prosthesis type, terminal device type and by DEKA Arm configuration level. This was a two-part study; Part A consisted of in-laboratory training. Part B consisted of home use. Study participants were 23 prosthesis users (mean age = 45 ± 16; 87% male) who completed Part A, and 15 (mean age = 45 ± 18; 87% male) who completed Parts A and B. Outcomes including self-report and performance measures, were collected at Baseline using participants' personal prostheses and at the End of Parts A and B. Scores were compared using paired t-tests. Wilcoxon signed-rank tests were used to compare outcomes for the full sample, and for the sample stratified by device and terminal device type. Analysis of outcomes by configuration level was performed graphically. At the End of Part A activity performance using the DEKA Arm and conventional prosthesis was equivalent, but slower with the DEKA Arm. After Part B, performance using the DEKA Arm surpassed conventional prosthesis scores, and speed of activity completion was equivalent. Participants reported using the DEKA Arm to perform more activities, had less perceived disability, and less difficulty in activities at the End of A and B as compared to Baseline. No differences were observed in dexterity, prosthetic skill, spontaneity, pain, community integration or quality of life. Comparisons stratified by device type revealed similar patterns. Graphic comparisons revealed variations by configuration level. Participants using the DEKA Arm had less perceived disability and more engagement of the prosthesis in everyday tasks, although activity performance was slower. After home use experience, activity performance was improved and activity speed equivalent to using conventional prostheses.
Design curves for circular and annular duct silencers
NASA Technical Reports Server (NTRS)
Watson, Willie R.; Ramakrishnan, R.
1989-01-01
Conventional models of sound propagation between porous walls (Scott, 1946) are adapted in order to calculate design curves for the lined circular and annular-duct silencers used in HVAC systems. The derivation of the governing equations is outlined, and results for two typical cases are presented graphically. Good agreement with published experimental data is demonstrated.
A Matlab/Simulink-Based Interactive Module for Servo Systems Learning
ERIC Educational Resources Information Center
Aliane, N.
2010-01-01
This paper presents an interactive module for learning both the fundamental and practical issues of servo systems. This module, developed using Simulink in conjunction with the Matlab graphical user interface (Matlab-GUI) tool, is used to supplement conventional lectures in control engineering and robotics subjects. First, the paper introduces the…
Ardoino, Ilaria; Lanzoni, Monica; Marano, Giuseppe; Boracchi, Patrizia; Sagrini, Elisabetta; Gianstefani, Alice; Piscaglia, Fabio; Biganzoli, Elia M
2017-04-01
The interpretation of regression models results can often benefit from the generation of nomograms, 'user friendly' graphical devices especially useful for assisting the decision-making processes. However, in the case of multinomial regression models, whenever categorical responses with more than two classes are involved, nomograms cannot be drawn in the conventional way. Such a difficulty in managing and interpreting the outcome could often result in a limitation of the use of multinomial regression in decision-making support. In the present paper, we illustrate the derivation of a non-conventional nomogram for multinomial regression models, intended to overcome this issue. Although it may appear less straightforward at first sight, the proposed methodology allows an easy interpretation of the results of multinomial regression models and makes them more accessible for clinicians and general practitioners too. Development of prediction model based on multinomial logistic regression and of the pertinent graphical tool is illustrated by means of an example involving the prediction of the extent of liver fibrosis in hepatitis C patients by routinely available markers.
ERIC Educational Resources Information Center
Smith, Derrick W.; Smothers, Sinikka M.
2012-01-01
Introduction: The purpose of the study presented here was to determine how well tactile graphics (specifically data analysis graphs) in secondary mathematics and science braille textbooks correlated with the print graphics. Method: A content analysis was conducted on 598 separate data analysis graphics from 10 mathematics and science textbooks.…
ERIC Educational Resources Information Center
Wood, Mara
2015-01-01
Graphic novels use visual literacy and multimodal learning two methods of teaching. Graphic novels also have a history of being motivating to students. This study aims to quantify the degree of influence graphic novels have on secondary student comprehension and motivation. Students were recruited from two classrooms taught by one twelfth-grade…
Towards real-time image deconvolution: application to confocal and STED microscopy
Zanella, R.; Zanghirati, G.; Cavicchioli, R.; Zanni, L.; Boccacci, P.; Bertero, M.; Vicidomini, G.
2013-01-01
Although deconvolution can improve the quality of any type of microscope, the high computational time required has so far limited its massive spreading. Here we demonstrate the ability of the scaled-gradient-projection (SGP) method to provide accelerated versions of the most used algorithms in microscopy. To achieve further increases in efficiency, we also consider implementations on graphic processing units (GPUs). We test the proposed algorithms both on synthetic and real data of confocal and STED microscopy. Combining the SGP method with the GPU implementation we achieve a speed-up factor from about a factor 25 to 690 (with respect the conventional algorithm). The excellent results obtained on STED microscopy images demonstrate the synergy between super-resolution techniques and image-deconvolution. Further, the real-time processing allows conserving one of the most important property of STED microscopy, i.e the ability to provide fast sub-diffraction resolution recordings. PMID:23982127
PC graphics generation and management tool for real-time applications
NASA Technical Reports Server (NTRS)
Truong, Long V.
1992-01-01
A graphics tool was designed and developed for easy generation and management of personal computer graphics. It also provides methods and 'run-time' software for many common artificial intelligence (AI) or expert system (ES) applications.
Almeida, Jonas S.; Iriabho, Egiebade E.; Gorrepati, Vijaya L.; Wilkinson, Sean R.; Grüneberg, Alexander; Robbins, David E.; Hackney, James R.
2012-01-01
Background: Image bioinformatics infrastructure typically relies on a combination of server-side high-performance computing and client desktop applications tailored for graphic rendering. On the server side, matrix manipulation environments are often used as the back-end where deployment of specialized analytical workflows takes place. However, neither the server-side nor the client-side desktop solution, by themselves or combined, is conducive to the emergence of open, collaborative, computational ecosystems for image analysis that are both self-sustained and user driven. Materials and Methods: ImageJS was developed as a browser-based webApp, untethered from a server-side backend, by making use of recent advances in the modern web browser such as a very efficient compiler, high-end graphical rendering capabilities, and I/O tailored for code migration. Results: Multiple versioned code hosting services were used to develop distinct ImageJS modules to illustrate its amenability to collaborative deployment without compromise of reproducibility or provenance. The illustrative examples include modules for image segmentation, feature extraction, and filtering. The deployment of image analysis by code migration is in sharp contrast with the more conventional, heavier, and less safe reliance on data transfer. Accordingly, code and data are loaded into the browser by exactly the same script tag loading mechanism, which offers a number of interesting applications that would be hard to attain with more conventional platforms, such as NIH's popular ImageJ application. Conclusions: The modern web browser was found to be advantageous for image bioinformatics in both the research and clinical environments. This conclusion reflects advantages in deployment scalability and analysis reproducibility, as well as the critical ability to deliver advanced computational statistical procedures machines where access to sensitive data is controlled, that is, without local “download and installation”. PMID:22934238
NASA Astrophysics Data System (ADS)
Choi, Sunghoon; Lee, Seungwan; Lee, Haenghwa; Lee, Donghoon; Choi, Seungyeon; Shin, Jungwook; Seo, Chang-Woo; Kim, Hee-Joung
2017-03-01
Digital tomosynthesis offers the advantage of low radiation doses compared to conventional computed tomography (CT) by utilizing small numbers of projections ( 80) acquired over a limited angular range. It produces 3D volumetric data, although there are artifacts due to incomplete sampling. Based upon these characteristics, we developed a prototype digital tomosynthesis R/F system for applications in chest imaging. Our prototype chest digital tomosynthesis (CDT) R/F system contains an X-ray tube with high power R/F pulse generator, flat-panel detector, R/F table, electromechanical radiographic subsystems including a precise motor controller, and a reconstruction server. For image reconstruction, users select between analytic and iterative reconstruction methods. Our reconstructed images of Catphan700 and LUNGMAN phantoms clearly and rapidly described the internal structures of phantoms using graphics processing unit (GPU) programming. Contrast-to-noise ratio (CNR) values of the CTP682 module of Catphan700 were higher in images using a simultaneous algebraic reconstruction technique (SART) than in those using filtered back-projection (FBP) for all materials by factors of 2.60, 3.78, 5.50, 2.30, 3.70, and 2.52 for air, lung foam, low density polyethylene (LDPE), Delrin® (acetal homopolymer resin), bone 50% (hydroxyapatite), and Teflon, respectively. Total elapsed times for producing 3D volume were 2.92 s and 86.29 s on average for FBP and SART (20 iterations), respectively. The times required for reconstruction were clinically feasible. Moreover, the total radiation dose from our system (5.68 mGy) was lower than that of conventional chest CT scan. Consequently, our prototype tomosynthesis R/F system represents an important advance in digital tomosynthesis applications.
Optimization of ultrahigh-speed multiplex PCR for forensic analysis.
Gibson-Daw, Georgiana; Crenshaw, Karin; McCord, Bruce
2018-01-01
In this paper, we demonstrate the design and optimization of an ultrafast PCR amplification technique, used with a seven-locus multiplex that is compatible with conventional capillary electrophoresis systems as well as newer microfluidic chip devices. The procedure involves the use of a high-speed polymerase and a rapid cycling protocol to permit multiplex PCR amplification of forensic short tandem repeat loci in 6.5 min. We describe the selection and optimization of master mix reagents such as enzyme, buffer, MgCl 2 , and dNTPs, as well as primer ratios, total volume, and cycle conditions, in order to get the best profile in the shortest time possible. Sensitivity and reproducibility studies are also described. The amplification process utilizes a small high-speed thermocycler and compact laptop, making it portable and potentially useful for rapid, inexpensive on-site genotyping. The seven loci of the multiplex were taken from conventional STR genotyping kits and selected for their size and lack of overlap. Analysis was performed using conventional capillary electrophoresis and microfluidics with fluorescent detection. Overall, this technique provides a more rapid method for rapid sample screening of suspects and victims. Graphical abstract Rapid amplification of forensic DNA using high speed thermal cycling followed by capillary or microfluidic electrophoresis.
Bigman, Cabral A; Nagler, Rebekah H; Viswanath, K
2016-08-01
As countries implement Article 11 of the World Health Organization (WHO) Framework Convention on Tobacco Control, graphic warning labels that use images of people and their body parts to illustrate the consequences of smoking are being added to cigarette packs. According to exemplification theory, these case examples-exemplars-can shape perceptions about risk and may resonate differently among demographic subpopulations. Drawing on data from eight focus groups (N = 63) with smokers and nonsmokers from vulnerable populations, this qualitative study explores whether people considered exemplars in their reactions to and evaluations of U.S. graphic health warning labels initially proposed by the Food and Drug Administration. Participants made reference to prior and concurrent mass media messages and exemplars during the focus groups and used demographic cues in making sense of the images on the warning labels. Participants were particularly sensitive to age of the exemplars and how it might affect label effectiveness and beliefs about smoking. Race and socioeconomic status also were salient for some participants. We recommend that exemplars and exemplification be considered when selecting and evaluating graphic health warnings for tobacco labels and associated media campaigns.
Krummenauer, Frank; Storkebaum, Kristin; Dick, H Burkhard
2003-01-01
The evaluation of new diagnostic measurement devices allows intraindividual comparison with an established standard method. However, reports in journal articles often omit the adequate incorporation of the intraindividual design into the graphic representation. This article illustrates the drawbacks and the possible erroneous conclusions caused by this misleading practice in terms of recent method comparison data resulting from axial length measurement in 220 consecutive patients by both applanation ultrasound and partial coherence interferometry. Graphic representation of such method comparison data should be based on boxplots for intraindividual differences or on Bland-Altman plots. Otherwise, severe deviations between the measurement devices could be erroneously ignored and false-positive conclusions on the concordance of the instruments could result. Graphic representation of method comparison data should sensitively incorporate the underlying study design for intraindividual comparison.
Rodgers, Joseph Lee; Beasley, William Howard; Schuelke, Matthew
2014-01-01
Many data structures, particularly time series data, are naturally seasonal, cyclical, or otherwise circular. Past graphical methods for time series have focused on linear plots. In this article, we move graphical analysis onto the circle. We focus on 2 particular methods, one old and one new. Rose diagrams are circular histograms and can be produced in several different forms using the RRose software system. In addition, we propose, develop, illustrate, and provide software support for a new circular graphical method, called Wrap-Around Time Series Plots (WATS Plots), which is a graphical method useful to support time series analyses in general but in particular in relation to interrupted time series designs. We illustrate the use of WATS Plots with an interrupted time series design evaluating the effect of the Oklahoma City bombing on birthrates in Oklahoma County during the 10 years surrounding the bombing of the Murrah Building in Oklahoma City. We compare WATS Plots with linear time series representations and overlay them with smoothing and error bands. Each method is shown to have advantages in relation to the other; in our example, the WATS Plots more clearly show the existence and effect size of the fertility differential.
Lifetime Prediction of IGBT in a STATCOM Using Modified-Graphical Rainflow Counting Algorithm
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gopi Reddy, Lakshmi Reddy; Tolbert, Leon M; Ozpineci, Burak
Rainflow algorithms are one of the best counting methods used in fatigue and failure analysis [17]. There have been many approaches to the rainflow algorithm, some proposing modifications. Graphical Rainflow Method (GRM) was proposed recently with a claim of faster execution times [10]. However, the steps of the graphical method of rainflow algorithm, when implemented, do not generate the same output as the four-point or ASTM standard algorithm. A modified graphical method is presented and discussed in this paper to overcome the shortcomings of graphical rainflow algorithm. A fast rainflow algorithm based on four-point algorithm but considering point comparison thanmore » range comparison is also presented. A comparison between the performances of the common rainflow algorithms [6-10], including the proposed methods, in terms of execution time, memory used, and efficiency, complexity, and load sequences is presented. Finally, the rainflow algorithm is applied to temperature data of an IGBT in assessing the lifetime of a STATCOM operating for power factor correction of the load. From 5-minute data load profiles available, the lifetime is estimated to be at 3.4 years.« less
Nanoparticles in Constanta-North Wastewater Treatment Plant
NASA Astrophysics Data System (ADS)
Panaitescu, I. M.; Panaitescu, Fanel-Viorel L.; Panaitescu, Ileana-Irina F. V.
2015-02-01
In this paper we describe the route of the nanoparticles in the WWTP and demonstrate how to use the simulation flow sensitivity analysis within STOATTM program to evaluate the effect of variation of the constant, "k" in the equation v= kCh settling on fixed concentration of nanoparticles in sewage water from a primary tank of physical-biological stage. Wastewater treatment facilities are designed to remove conventional pollutants from sanitary waste. Major processes of treatment includes: a) physical treatment-remove suspended large solids by settling or sedimentation and eliminate floating greases; b) biological treatment-degradation or consumption of the dissolved organic matter using the means of cultivated in activated sludge or the trickling filters; c) chemical treatment-remove other matters by the means of chemical addition or destroying pathogenic organisms through disinfection; d) advanced treatment- removing specific constituents using processes such as activated carbon, membrane separation, or ion exchange. Particular treatment processes are: a) sedimentation; b) coagulation and flocculation; c) activated sludge; d) sand filters; e) membrane separation; f) disinfection. Methods are: 1) using the STOATTM program with input and output data for primary tank and parameters of wastewater. 2) generating a data file for influent using a sinusoidal model and we accepted defaults STOATTM data. 3) After this, getting spreadsheet data for various characteristics of wastewater for 48 hours:flow, temperature, pH, volatile fatty acids, soluble BOD, COD inert soluble particulate BOD, COD inert particles, volatile solids, volatile solids, ammonia, nitrate and soluble organic nitrogen. Findings and Results:1.Graphics after 48 hour;. 2.Graphics for parameters - flow,temperature, pH/units hours; 3.Graphics of nanoparticles; 4. Graphics of others volatile and non-volatile solids; 5. Timeseries data and summary statistics. Biodegradation of nanoparticles is the breakdown of organic molecules that may cause changes in the physical structure or the surface characteristic of the material.
Breast ultrasound computed tomography using waveform inversion with source encoding
NASA Astrophysics Data System (ADS)
Wang, Kun; Matthews, Thomas; Anis, Fatima; Li, Cuiping; Duric, Neb; Anastasio, Mark A.
2015-03-01
Ultrasound computed tomography (USCT) holds great promise for improving the detection and management of breast cancer. Because they are based on the acoustic wave equation, waveform inversion-based reconstruction methods can produce images that possess improved spatial resolution properties over those produced by ray-based methods. However, waveform inversion methods are computationally demanding and have not been applied widely in USCT breast imaging. In this work, source encoding concepts are employed to develop an accelerated USCT reconstruction method that circumvents the large computational burden of conventional waveform inversion methods. This method, referred to as the waveform inversion with source encoding (WISE) method, encodes the measurement data using a random encoding vector and determines an estimate of the speed-of-sound distribution by solving a stochastic optimization problem by use of a stochastic gradient descent algorithm. Computer-simulation studies are conducted to demonstrate the use of the WISE method. Using a single graphics processing unit card, each iteration can be completed within 25 seconds for a 128 × 128 mm2 reconstruction region. The results suggest that the WISE method maintains the high spatial resolution of waveform inversion methods while significantly reducing the computational burden.
GPU-accelerated FDTD modeling of radio-frequency field-tissue interactions in high-field MRI.
Chi, Jieru; Liu, Feng; Weber, Ewald; Li, Yu; Crozier, Stuart
2011-06-01
The analysis of high-field RF field-tissue interactions requires high-performance finite-difference time-domain (FDTD) computing. Conventional CPU-based FDTD calculations offer limited computing performance in a PC environment. This study presents a graphics processing unit (GPU)-based parallel-computing framework, producing substantially boosted computing efficiency (with a two-order speedup factor) at a PC-level cost. Specific details of implementing the FDTD method on a GPU architecture have been presented and the new computational strategy has been successfully applied to the design of a novel 8-element transceive RF coil system at 9.4 T. Facilitated by the powerful GPU-FDTD computing, the new RF coil array offers optimized fields (averaging 25% improvement in sensitivity, and 20% reduction in loop coupling compared with conventional array structures of the same size) for small animal imaging with a robust RF configuration. The GPU-enabled acceleration paves the way for FDTD to be applied for both detailed forward modeling and inverse design of MRI coils, which were previously impractical.
NASA Astrophysics Data System (ADS)
Dallacqua, Ashley Kaye
This year-long, ethnographic study documents the use of comics and graphic novels as academic literature across the curriculum in a suburban middle school. Because the use of this medium in classrooms is relatively new, it is a process that has not been extensively documented. While comics and graphic novels can provide a complex and valuable experience for readers, they can also be challenging to both students and teachers. In particular, this dissertation documents the tensions that surfaced as comics and graphic novels were integrated into a curriculum. This study is situated in a middle school entrenched in neoliberal ideologies, with focuses on high-stakes testing, a standardized curriculum, and individual, rather than collaborative work. Yet, the faculty in this middle school was also inviting nontraditional texts into classrooms, and operating in tension with a neoliberal agenda. By focusing on teaching and learning literacy practices with comics and graphic novels and talk about those practices, this study also addresses negative assumptions and hesitancies around such texts being used for academic purposes. Participants included seventh grade teachers and students engaged in working with and talking about comics. This research considers how comics and graphic novels were welcomed into this school, as well as impacts around time and space, and positioning. All of these themes point back to how comics and graphic novels were working within and against normative structures in this school. This study is positioned to consider conventional literacy practices and how teaching and learning with comics and graphic novels supports and disrupts those practices. Serving as an example and a starting point for bringing this dynamic medium into classrooms, this study fills a significant gap, supporting and challenging traditional literacies practices and analyzing potential for new ways of operating in a school.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Xia; Liang, Pan; Huang, Hong-Sheng
2014-04-01
Graphical abstract: LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphor was obtained by calcining the precursor which was synthesized by boric acid melting method. It (a) exhibits much stronger PL intensity than that (b) prepared by conventional solid state reaction method. - Highlights: • A calcining precursor method was used for preparation of LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphor. • Precursor was prepared by boric acid melting method. • The luminescence intensity of LaB{sub 3}O{sub 6}:Eu{sup 3+} was enhanced by the present method. - Abstract: The LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphors were prepared by calcining the precursors which were synthesized by boric acid meltingmore » method using rare earth oxide and boric acid as raw materials, and they were characterized by EDS, XRD, IR, SEM and PL. The influences of reaction temperature for the preparation of precursor and subsequent calcination temperature and time of precursor on the luminescence properties of LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphor were investigated. The results showed that the LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphors with maximum luminescent intensity were obtained by calcining precursor at 1000 °C for 6 h, in which the precursor was prepared at 200 °C for 72 h. Compared with the conventional high temperature solid-state reaction method, the pure LaB{sub 3}O{sub 6}:Eu{sup 3+} phosphor can be obtained at relatively lower calcination temperature by the precursor method and exhibits much stronger emission intensity.« less
System dynamic modeling: an alternative method for budgeting.
Srijariya, Witsanuchai; Riewpaiboon, Arthorn; Chaikledkaew, Usa
2008-03-01
To construct, validate, and simulate a system dynamic financial model and compare it against the conventional method. The study was a cross-sectional analysis of secondary data retrieved from the National Health Security Office (NHSO) in the fiscal year 2004. The sample consisted of all emergency patients who received emergency services outside their registered hospital-catchments area. The dependent variable used was the amount of reimbursed money. Two types of model were constructed, namely, the system dynamic model using the STELLA software and the multiple linear regression model. The outputs of both methods were compared. The study covered 284,716 patients from various levels of providers. The system dynamic model had the capability of producing various types of outputs, for example, financial and graphical analyses. For the regression analysis, statistically significant predictors were composed of service types (outpatient or inpatient), operating procedures, length of stay, illness types (accident or not), hospital characteristics, age, and hospital location (adjusted R(2) = 0.74). The total budget arrived at from using the system dynamic model and regression model was US$12,159,614.38 and US$7,301,217.18, respectively, whereas the actual NHSO reimbursement cost was US$12,840,805.69. The study illustrated that the system dynamic model is a useful financial management tool, although it is not easy to construct. The model is not only more accurate in prediction but is also more capable of analyzing large and complex real-world situations than the conventional method.
Exploiting current-generation graphics hardware for synthetic-scene generation
NASA Astrophysics Data System (ADS)
Tanner, Michael A.; Keen, Wayne A.
2010-04-01
Increasing seeker frame rate and pixel count, as well as the demand for higher levels of scene fidelity, have driven scene generation software for hardware-in-the-loop (HWIL) and software-in-the-loop (SWIL) testing to higher levels of parallelization. Because modern PC graphics cards provide multiple computational cores (240 shader cores for a current NVIDIA Corporation GeForce and Quadro cards), implementation of phenomenology codes on graphics processing units (GPUs) offers significant potential for simultaneous enhancement of simulation frame rate and fidelity. To take advantage of this potential requires algorithm implementation that is structured to minimize data transfers between the central processing unit (CPU) and the GPU. In this paper, preliminary methodologies developed at the Kinetic Hardware In-The-Loop Simulator (KHILS) will be presented. Included in this paper will be various language tradeoffs between conventional shader programming, Compute Unified Device Architecture (CUDA) and Open Computing Language (OpenCL), including performance trades and possible pathways for future tool development.
W-curve alignments for HIV-1 genomic comparisons.
Cork, Douglas J; Lembark, Steven; Tovanabutra, Sodsai; Robb, Merlin L; Kim, Jerome H
2010-06-01
The W-curve was originally developed as a graphical visualization technique for viewing DNA and RNA sequences. Its ability to render features of DNA also makes it suitable for computational studies. Its main advantage in this area is utilizing a single-pass algorithm for comparing the sequences. Avoiding recursion during sequence alignments offers advantages for speed and in-process resources. The graphical technique also allows for multiple models of comparison to be used depending on the nucleotide patterns embedded in similar whole genomic sequences. The W-curve approach allows us to compare large numbers of samples quickly. We are currently tuning the algorithm to accommodate quirks specific to HIV-1 genomic sequences so that it can be used to aid in diagnostic and vaccine efforts. Tracking the molecular evolution of the virus has been greatly hampered by gap associated problems predominantly embedded within the envelope gene of the virus. Gaps and hypermutation of the virus slow conventional string based alignments of the whole genome. This paper describes the W-curve algorithm itself, and how we have adapted it for comparison of similar HIV-1 genomes. A treebuilding method is developed with the W-curve that utilizes a novel Cylindrical Coordinate distance method and gap analysis method. HIV-1 C2-V5 env sequence regions from a Mother/Infant cohort study are used in the comparison. The output distance matrix and neighbor results produced by the W-curve are functionally equivalent to those from Clustal for C2-V5 sequences in the mother/infant pairs infected with CRF01_AE. Significant potential exists for utilizing this method in place of conventional string based alignment of HIV-1 genomes, such as Clustal X. With W-curve heuristic alignment, it may be possible to obtain clinically useful results in a short time-short enough to affect clinical choices for acute treatment. A description of the W-curve generation process, including a comparison technique of aligning extremes of the curves to effectively phase-shift them past the HIV-1 gap problem, is presented. Besides yielding similar neighbor-joining phenogram topologies, most Mother and Infant C2-V5 sequences in the cohort pairs geometrically map closest to each other, indicating that W-curve heuristics overcame any gap problem.
Application of 2D graphic representation of protein sequence based on Huffman tree method.
Qi, Zhao-Hui; Feng, Jun; Qi, Xiao-Qin; Li, Ling
2012-05-01
Based on Huffman tree method, we propose a new 2D graphic representation of protein sequence. This representation can completely avoid loss of information in the transfer of data from a protein sequence to its graphic representation. The method consists of two parts. One is about the 0-1 codes of 20 amino acids by Huffman tree with amino acid frequency. The amino acid frequency is defined as the statistical number of an amino acid in the analyzed protein sequences. The other is about the 2D graphic representation of protein sequence based on the 0-1 codes. Then the applications of the method on ten ND5 genes and seven Escherichia coli strains are presented in detail. The results show that the proposed model may provide us with some new sights to understand the evolution patterns determined from protein sequences and complete genomes. Copyright © 2012 Elsevier Ltd. All rights reserved.
Xu, Xiuqing; Yang, Xiuhan; Martin, Steven J; Mes, Edwin; Chen, Junlan; Meunier, David M
2018-08-17
Accurate measurement of molecular weight averages (M¯ n, M¯ w, M¯ z ) and molecular weight distributions (MWD) of polyether polyols by conventional SEC (size exclusion chromatography) is not as straightforward as it would appear. Conventional calibration with polystyrene (PS) standards can only provide PS apparent molecular weights which do not provide accurate estimates of polyol molecular weights. Using polyethylene oxide/polyethylene glycol (PEO/PEG) for molecular weight calibration could improve the accuracy, but the retention behavior of PEO/PEG is not stable in THF-based (tetrahydrofuran) SEC systems. In this work, two approaches for calibration curve conversion with narrow PS and polyol molecular weight standards were developed. Equations to convert PS-apparent molecular weight to polyol-apparent molecular weight were developed using both a rigorous mathematical analysis and graphical plot regression method. The conversion equations obtained by the two approaches were in good agreement. Factors influencing the conversion equation were investigated. It was concluded that the separation conditions such as column batch and operating temperature did not have significant impact on the conversion coefficients and a universal conversion equation could be obtained. With this conversion equation, more accurate estimates of molecular weight averages and MWDs for polyether polyols can be achieved from conventional PS-THF SEC calibration. Moreover, no additional experimentation is required to convert historical PS equivalent data to reasonably accurate molecular weight results. Copyright © 2018. Published by Elsevier B.V.
Orthorectification by Using Gpgpu Method
NASA Astrophysics Data System (ADS)
Sahin, H.; Kulur, S.
2012-07-01
Thanks to the nature of the graphics processing, the newly released products offer highly parallel processing units with high-memory bandwidth and computational power of more than teraflops per second. The modern GPUs are not only powerful graphic engines but also they are high level parallel programmable processors with very fast computing capabilities and high-memory bandwidth speed compared to central processing units (CPU). Data-parallel computations can be shortly described as mapping data elements to parallel processing threads. The rapid development of GPUs programmability and capabilities attracted the attentions of researchers dealing with complex problems which need high level calculations. This interest has revealed the concepts of "General Purpose Computation on Graphics Processing Units (GPGPU)" and "stream processing". The graphic processors are powerful hardware which is really cheap and affordable. So the graphic processors became an alternative to computer processors. The graphic chips which were standard application hardware have been transformed into modern, powerful and programmable processors to meet the overall needs. Especially in recent years, the phenomenon of the usage of graphics processing units in general purpose computation has led the researchers and developers to this point. The biggest problem is that the graphics processing units use different programming models unlike current programming methods. Therefore, an efficient GPU programming requires re-coding of the current program algorithm by considering the limitations and the structure of the graphics hardware. Currently, multi-core processors can not be programmed by using traditional programming methods. Event procedure programming method can not be used for programming the multi-core processors. GPUs are especially effective in finding solution for repetition of the computing steps for many data elements when high accuracy is needed. Thus, it provides the computing process more quickly and accurately. Compared to the GPUs, CPUs which perform just one computing in a time according to the flow control are slower in performance. This structure can be evaluated for various applications of computer technology. In this study covers how general purpose parallel programming and computational power of the GPUs can be used in photogrammetric applications especially direct georeferencing. The direct georeferencing algorithm is coded by using GPGPU method and CUDA (Compute Unified Device Architecture) programming language. Results provided by this method were compared with the traditional CPU programming. In the other application the projective rectification is coded by using GPGPU method and CUDA programming language. Sample images of various sizes, as compared to the results of the program were evaluated. GPGPU method can be used especially in repetition of same computations on highly dense data, thus finding the solution quickly.
Graphic Novels, Web Comics, and Creator Blogs: Examining Product and Process
ERIC Educational Resources Information Center
Carter, James Bucky
2011-01-01
Young adult literature (YAL) of the late 20th and early 21st century is exploring hybrid forms with growing regularity by embracing textual conventions from sequential art, video games, film, and more. As well, Web-based technologies have given those who consume YAL more immediate access to authors, their metacognitive creative processes, and…
Facsimile Transmission of Microforms.
1983-12-30
display terminals, high speed printers, conventional facsimile receivers, and/or graphics COM recorders. Microforms designed for storage, retrieval...author and Whould not be construed as an official Department of the Army position, policy, or decision, unless so designated by other documentation...beconstrued as an official Department of the Army position, policy, or decision, unless so designated by other documentation. ,, -- UNCLASSIFIED SECURITY
Alternative Representations for Algebraic Problem Solving: When Are Graphs Better than Equations?
ERIC Educational Resources Information Center
Mielicki, Marta K.; Wiley, Jennifer
2016-01-01
Successful algebraic problem solving entails adaptability of solution methods using different representations. Prior research has suggested that students are more likely to prefer symbolic solution methods (equations) over graphical ones, even when graphical methods should be more efficient. However, this research has not tested how representation…
A User Study on Tactile Graphic Generation Methods
ERIC Educational Resources Information Center
Krufka, S. E.; Barner, K. E.
2006-01-01
Methods to automatically convert graphics into tactile representations have been recently investigated, creating either raised-line or relief images. In particular, we briefly review one raised-line method where important features are emphasized. This paper focuses primarily on the effects of such emphasis and on comparing both raised-line and…
Nguyen, Van-Giang; Lee, Soo-Jin
2016-07-01
Iterative reconstruction from Compton scattered data is known to be computationally more challenging than that from conventional line-projection based emission data in that the gamma rays that undergo Compton scattering are modeled as conic projections rather than line projections. In conventional tomographic reconstruction, to parallelize the projection and backprojection operations using the graphics processing unit (GPU), approximated methods that use an unmatched pair of ray-tracing forward projector and voxel-driven backprojector have been widely used. In this work, we propose a new GPU-accelerated method for Compton camera reconstruction which is more accurate by using exactly matched pair of projector and backprojector. To calculate conic forward projection, we first sample the cone surface into conic rays and accumulate the intersecting chord lengths of the conic rays passing through voxels using a fast ray-tracing method (RTM). For conic backprojection, to obtain the true adjoint of the conic forward projection, while retaining the computational efficiency of the GPU, we use a voxel-driven RTM which is essentially the same as the standard RTM used for the conic forward projector. Our simulation results show that, while the new method is about 3 times slower than the approximated method, it is still about 16 times faster than the CPU-based method without any loss of accuracy. The net conclusion is that our proposed method is guaranteed to retain the reconstruction accuracy regardless of the number of iterations by providing a perfectly matched projector-backprojector pair, which makes iterative reconstruction methods for Compton imaging faster and more accurate. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
ERIC Educational Resources Information Center
Alhajri, Salman
2016-01-01
Purpose: this paper investigates the effectiveness of teaching methods used in graphic design pedagogy in both analogue and digital education systems. Methodology and approach: the paper is based on theoretical study using a qualitative, case study approach. Comparison between the digital teaching methods and traditional teaching methods was…
Gaussian graphical modeling reveals specific lipid correlations in glioblastoma cells
NASA Astrophysics Data System (ADS)
Mueller, Nikola S.; Krumsiek, Jan; Theis, Fabian J.; Böhm, Christian; Meyer-Bäse, Anke
2011-06-01
Advances in high-throughput measurements of biological specimens necessitate the development of biologically driven computational techniques. To understand the molecular level of many human diseases, such as cancer, lipid quantifications have been shown to offer an excellent opportunity to reveal disease-specific regulations. The data analysis of the cell lipidome, however, remains a challenging task and cannot be accomplished solely based on intuitive reasoning. We have developed a method to identify a lipid correlation network which is entirely disease-specific. A powerful method to correlate experimentally measured lipid levels across the various samples is a Gaussian Graphical Model (GGM), which is based on partial correlation coefficients. In contrast to regular Pearson correlations, partial correlations aim to identify only direct correlations while eliminating indirect associations. Conventional GGM calculations on the entire dataset can, however, not provide information on whether a correlation is truly disease-specific with respect to the disease samples and not a correlation of control samples. Thus, we implemented a novel differential GGM approach unraveling only the disease-specific correlations, and applied it to the lipidome of immortal Glioblastoma tumor cells. A large set of lipid species were measured by mass spectrometry in order to evaluate lipid remodeling as a result to a combination of perturbation of cells inducing programmed cell death, while the other perturbations served solely as biological controls. With the differential GGM, we were able to reveal Glioblastoma-specific lipid correlations to advance biomedical research on novel gene therapies.
Gonthier, Gerard
2007-01-01
A graphical method that uses continuous water-level and barometric-pressure data was developed to estimate barometric efficiency. A plot of nearly continuous water level (on the y-axis), as a function of nearly continuous barometric pressure (on the x-axis), will plot as a line curved into a series of connected elliptical loops. Each loop represents a barometric-pressure fluctuation. The negative of the slope of the major axis of an elliptical loop will be the ratio of water-level change to barometric-pressure change, which is the sum of the barometric efficiency plus the error. The negative of the slope of the preferred orientation of many elliptical loops is an estimate of the barometric efficiency. The slope of the preferred orientation of many elliptical loops is approximately the median of the slopes of the major axes of the elliptical loops. If water-level change that is not caused by barometric-pressure change does not correlate with barometric-pressure change, the probability that the error will be greater than zero will be the same as the probability that it will be less than zero. As a result, the negative of the median of the slopes for many loops will be close to the barometric efficiency. The graphical method provided a rapid assessment of whether a well was affected by barometric-pressure change and also provided a rapid estimate of barometric efficiency. The graphical method was used to assess which wells at Air Force Plant 6, Marietta, Georgia, had water levels affected by barometric-pressure changes during a 2003 constant-discharge aquifer test. The graphical method was also used to estimate barometric efficiency. Barometric-efficiency estimates from the graphical method were compared to those of four other methods: average of ratios, median of ratios, Clark, and slope. The two methods (the graphical and median-of-ratios methods) that used the median values of water-level change divided by barometric-pressure change appeared to be most resistant to error caused by barometric-pressure-independent water-level change. The graphical method was particularly resistant to large amounts of barometric-pressure-independent water-level change, having an average and standard deviation of error for control wells that was less than one-quarter that of the other four methods. When using the graphical method, it is advisable that more than one person select the slope or that the same person fits the same data several times to minimize the effect of subjectivity. Also, a long study period should be used (at least 60 days) to ensure that loops affected by large amounts of barometric-pressure-independent water-level change do not significantly contribute to error in the barometric-efficiency estimate.
Development of a novel cell sorting method that samples population diversity in flow cytometry.
Osborne, Geoffrey W; Andersen, Stacey B; Battye, Francis L
2015-11-01
Flow cytometry based electrostatic cell sorting is an important tool in the separation of cell populations. Existing instruments can sort single cells into multi-well collection plates, and keep track of cell of origin and sorted well location. However currently single sorted cell results reflect the population distribution and fail to capture the population diversity. Software was designed that implements a novel sorting approach, "Slice and Dice Sorting," that links a graphical representation of a multi-well plate to logic that ensures that single cells are sampled and sorted from all areas defined by the sort region/s. Therefore the diversity of the total population is captured, and the more frequently occurring or rarer cell types are all sampled. The sorting approach was tested computationally, and using functional cell based assays. Computationally we demonstrate that conventional single cell sorting can sample as little as 50% of the population diversity dependant on the population distribution, and that Slice and Dice sorting samples much more of the variety present within a cell population. We then show by sorting single cells into wells using the Slice and Dice sorting method that there are cells sorted using this method that would be either rarely sorted, or not sorted at all using conventional single cell sorting approaches. The present study demonstrates a novel single cell sorting method that samples much more of the population diversity than current methods. It has implications in clonal selection, stem cell sorting, single cell sequencing and any areas where population heterogeneity is of importance. © 2015 International Society for Advancement of Cytometry.
Engineering Graphics in Education: Programming and Ready Programs.
ERIC Educational Resources Information Center
Audi, M. S.
1987-01-01
Suggests a method of integrating teaching microcomputer graphics in engineering curricula without encroaching on the fundamental engineering courses. Includes examples of engineering graphics produced by commercial programs and others produced by high-level language programing in a limited credit hour segment of an educational program. (CW)
Umoquit, Muriah J; Dobrow, Mark J; Lemieux-Charles, Louise; Ritvo, Paul G; Urbach, David R; Wodchis, Walter P
2008-08-08
This paper focuses on measuring the efficiency and effectiveness of two diagramming methods employed in key informant interviews with clinicians and health care administrators. The two methods are 'participatory diagramming', where the respondent creates a diagram that assists in their communication of answers, and 'graphic elicitation', where a researcher-prepared diagram is used to stimulate data collection. These two diagramming methods were applied in key informant interviews and their value in efficiently and effectively gathering data was assessed based on quantitative measures and qualitative observations. Assessment of the two diagramming methods suggests that participatory diagramming is an efficient method for collecting data in graphic form, but may not generate the depth of verbal response that many qualitative researchers seek. In contrast, graphic elicitation was more intuitive, better understood and preferred by most respondents, and often provided more contemplative verbal responses, however this was achieved at the expense of more interview time. Diagramming methods are important for eliciting interview data that are often difficult to obtain through traditional verbal exchanges. Subject to the methodological limitations of the study, our findings suggest that while participatory diagramming and graphic elicitation have specific strengths and weaknesses, their combined use can provide complementary information that would not likely occur with the application of only one diagramming method. The methodological insights gained by examining the efficiency and effectiveness of these diagramming methods in our study should be helpful to other researchers considering their incorporation into qualitative research designs.
Graphic representation of niche width and its application to salt marsh littoral foraminifera
DOE Office of Scientific and Technical Information (OSTI.GOV)
Muller, William A.
A method for representing the Hutchinsonian niche graphically has been designed and tested. The method provides a means for comparing the niche width of 3 species of foraminifera: Allogromia laticollaris, Rosalina leei and Spiroloculina hyalina.
A New Pattern of Getting Nasty Number in Graphical Method
NASA Astrophysics Data System (ADS)
Sumathi, P.; Indhumathi, N.
2018-04-01
This paper proposed a new technique of getting nasty numbers using graphical method in linear programming problem and it has been proved for various Linear programming problems. And also some characterisation of nasty numbers is discussed in this paper.
LookSeq: a browser-based viewer for deep sequencing data.
Manske, Heinrich Magnus; Kwiatkowski, Dominic P
2009-11-01
Sequencing a genome to great depth can be highly informative about heterogeneity within an individual or a population. Here we address the problem of how to visualize the multiple layers of information contained in deep sequencing data. We propose an interactive AJAX-based web viewer for browsing large data sets of aligned sequence reads. By enabling seamless browsing and fast zooming, the LookSeq program assists the user to assimilate information at different levels of resolution, from an overview of a genomic region to fine details such as heterogeneity within the sample. A specific problem, particularly if the sample is heterogeneous, is how to depict information about structural variation. LookSeq provides a simple graphical representation of paired sequence reads that is more revealing about potential insertions and deletions than are conventional methods.
Processing-in-Memory Enabled Graphics Processors for 3D Rendering
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xie, Chenhao; Song, Shuaiwen; Wang, Jing
2017-02-06
The performance of 3D rendering of Graphics Processing Unit that convents 3D vector stream into 2D frame with 3D image effects significantly impact users’ gaming experience on modern computer systems. Due to the high texture throughput in 3D rendering, main memory bandwidth becomes a critical obstacle for improving the overall rendering performance. 3D stacked memory systems such as Hybrid Memory Cube (HMC) provide opportunities to significantly overcome the memory wall by directly connecting logic controllers to DRAM dies. Based on the observation that texel fetches significantly impact off-chip memory traffic, we propose two architectural designs to enable Processing-In-Memory based GPUmore » for efficient 3D rendering.« less
CCP4i2: the new graphical user interface to the CCP4 program suite.
Potterton, Liz; Agirre, Jon; Ballard, Charles; Cowtan, Kevin; Dodson, Eleanor; Evans, Phil R; Jenkins, Huw T; Keegan, Ronan; Krissinel, Eugene; Stevenson, Kyle; Lebedev, Andrey; McNicholas, Stuart J; Nicholls, Robert A; Noble, Martin; Pannu, Navraj S; Roth, Christian; Sheldrick, George; Skubak, Pavol; Turkenburg, Johan; Uski, Ville; von Delft, Frank; Waterman, David; Wilson, Keith; Winn, Martyn; Wojdyr, Marcin
2018-02-01
The CCP4 (Collaborative Computational Project, Number 4) software suite for macromolecular structure determination by X-ray crystallography groups brings together many programs and libraries that, by means of well established conventions, interoperate effectively without adhering to strict design guidelines. Because of this inherent flexibility, users are often presented with diverse, even divergent, choices for solving every type of problem. Recently, CCP4 introduced CCP4i2, a modern graphical interface designed to help structural biologists to navigate the process of structure determination, with an emphasis on pipelining and the streamlined presentation of results. In addition, CCP4i2 provides a framework for writing structure-solution scripts that can be built up incrementally to create increasingly automatic procedures.
NASA Technical Reports Server (NTRS)
Ballester, P.
1992-01-01
MIDAS (Munich Image Data Analysis System) is the image processing system developed at ESO for astronomical data reduction. MIDAS is used for off-line data reduction at ESO and many astronomical institutes all over Europe. In addition to a set of general commands, enabling to process and analyze images, catalogs, graphics and tables, MIDAS includes specialized packages dedicated to astronomical applications or to specific ESO instruments. Several graphical interfaces are available in the MIDAS environment: XHelp provides an interactive help facility, and XLong and XEchelle enable data reduction of long-slip and echelle spectra. GUI builders facilitate the development of interfaces. All ESO interfaces comply to the ESO User Interfaces Common Conventions which secures an identical look and feel for telescope operations, data analysis, and archives.
ERIC Educational Resources Information Center
Thompson, Russel L.
Homoscedasticity is an important assumption of linear regression. This paper explains what it is and why it is important to the researcher. Graphical and mathematical methods for testing the homoscedasticity assumption are demonstrated. Sources of homoscedasticity and types of homoscedasticity are discussed, and methods for correction are…
Jackin, Boaz Jessie; Watanabe, Shinpei; Ootsu, Kanemitsu; Ohkawa, Takeshi; Yokota, Takashi; Hayasaki, Yoshio; Yatagai, Toyohiko; Baba, Takanobu
2018-04-20
A parallel computation method for large-size Fresnel computer-generated hologram (CGH) is reported. The method was introduced by us in an earlier report as a technique for calculating Fourier CGH from 2D object data. In this paper we extend the method to compute Fresnel CGH from 3D object data. The scale of the computation problem is also expanded to 2 gigapixels, making it closer to real application requirements. The significant feature of the reported method is its ability to avoid communication overhead and thereby fully utilize the computing power of parallel devices. The method exhibits three layers of parallelism that favor small to large scale parallel computing machines. Simulation and optical experiments were conducted to demonstrate the workability and to evaluate the efficiency of the proposed technique. A two-times improvement in computation speed has been achieved compared to the conventional method, on a 16-node cluster (one GPU per node) utilizing only one layer of parallelism. A 20-times improvement in computation speed has been estimated utilizing two layers of parallelism on a very large-scale parallel machine with 16 nodes, where each node has 16 GPUs.
John Herschel's Graphical Method
NASA Astrophysics Data System (ADS)
Hankins, Thomas L.
2011-01-01
In 1833 John Herschel published an account of his graphical method for determining the orbits of double stars. He had hoped to be the first to determine such orbits, but Felix Savary in France and Johann Franz Encke in Germany beat him to the punch using analytical methods. Herschel was convinced, however, that his graphical method was much superior to analytical methods, because it used the judgment of the hand and eye to correct the inevitable errors of observation. Line graphs of the kind used by Herschel became common only in the 1830s, so Herschel was introducing a new method. He also found computation fatiguing and devised a "wheeled machine" to help him out. Encke was skeptical of Herschel's methods. He said that he lived for calculation and that the English would be better astronomers if they calculated more. It is difficult to believe that the entire Scientific Revolution of the 17th century took place without graphs and that only a few examples appeared in the 18th century. Herschel promoted the use of graphs, not only in astronomy, but also in the study of meteorology and terrestrial magnetism. Because he was the most prominent scientist in England, Herschel's advocacy greatly advanced graphical methods.
NASA Technical Reports Server (NTRS)
Marshall, S. E.; Bernhard, R.
1984-01-01
A survey of the most widely used methods for visualizing acoustic phenomena is presented. Emphasis is placed on acoustic processes in the audible frequencies. Many visual problems are analyzed on computer graphic systems. A brief description of the current technology in computer graphics is included. The visualization technique survey will serve as basis for recommending an optimum scheme for displaying acoustic fields on computer graphic systems.
A Graphical Approach to Quantitative Structural Geology.
ERIC Educational Resources Information Center
De Paor, Declan G.
1986-01-01
Describes how computer graphic methods can be used in teaching structural geology. Describes the design of a graphics workstation for the Apple microcomputer. Includes a listing of commands used with software to plot structures in a digitized form. Argues for the establishment of computer laboratories for structural geology classes. (TW)
2015-01-01
1 3.0 Methods, Assumptions, and Procedures ...18 4.6.3. LineUp Web... Procedures A search of the internet looking at web sites specializing in graphics, graphics engines, web browser applications, and games was conducted to
Traditional Engineering Graphics versus Computer-Aided Drafting: A View from Academe.
ERIC Educational Resources Information Center
Foster, Robert J.
1987-01-01
Argues for a legitimate role of manually expressed engineering graphics within engineering education as a needed support for computer-assisted drafting work. Discusses what and how students should learn as well as trends in engineering graphics education. Compares and contrasts manual and computer drafting methods. (CW)
Learning with Interactive Computer Graphics in the Undergraduate Neuroscience Classroom
ERIC Educational Resources Information Center
Pani, John R.; Chariker, Julia H.; Naaz, Farah; Mattingly, William; Roberts, Joshua; Sephton, Sandra E.
2014-01-01
Instruction of neuroanatomy depends on graphical representation and extended self-study. As a consequence, computer-based learning environments that incorporate interactive graphics should facilitate instruction in this area. The present study evaluated such a system in the undergraduate neuroscience classroom. The system used the method of…
Interactive graphics to demonstrate health risks: formative development and qualitative evaluation
Ancker, Jessica S.; Chan, Connie; Kukafka, Rita
2015-01-01
Background Recent findings suggest that interactive game-like graphics might be useful in communicating probabilities. We developed a prototype for a risk communication module, focusing on eliciting users’ preferences for different interactive graphics and assessing usability and user interpretations. Methods Focus groups and iterative design methods. Results Feedback from five focus groups was used to design the graphics. The final version displayed a matrix of square buttons; clicking on any button allowed the user to see whether the stick figure underneath was affected by the health outcome. When participants used this interaction to learn about a risk, they expressed more emotional responses, both positive and negative, than when viewing any static graphic or numerical description of a risk. Their responses included relief about small risks and concern about large risks. The groups also commented on static graphics: arranging the figures affected by disease randomly throughout a group of figures made it more difficult to judge the proportion affected but was described as more realistic. Conclusions Interactive graphics appear to have potential for expressing risk magnitude as well as the affective feeling of risk. Quantitative studies are planned to assess the effect on perceived risks and estimated risk magnitudes. PMID:19657926
Pratt, Nathan S; Ellison, Brenna D; Benjamin, Aaron S; Nakamura, Manabu T
2016-01-01
Consumers have difficulty using nutrition information. We hypothesized that graphically delivering information of select nutrients relative to a target would allow individuals to process information in time-constrained settings more effectively than numerical information. Objectives of the study were to determine the efficacy of the graphical method in (1) improving memory of nutrient information and (2) improving consumer purchasing behavior in a restaurant. Values of fiber and protein per calorie were 2-dimensionally plotted alongside a target box. First, a randomized cued recall experiment was conducted (n=63). Recall accuracy of nutrition information improved by up to 43% when shown graphically instead of numerically. Second, the impact of graphical nutrition signposting on diner choices was tested in a cafeteria. Saturated fat and sodium information was also presented using color coding. Nutrient content of meals (n=362) was compared between 3 signposting phases: graphical, nutrition facts panels (NFP), or no nutrition label. Graphical signposting improved nutrient content of purchases in the intended direction, whereas NFP had no effect compared with the baseline. Calories ordered from total meals, entrées, and sides were significantly less during graphical signposting than no-label and NFP periods. For total meal and entrées, protein per calorie purchased was significantly higher and saturated fat significantly lower during graphical signposting than the other phases. Graphical signposting remained a predictor of calories and protein per calorie purchased in regression modeling. These findings demonstrate that graphically presenting nutrition information makes that information more available for decision making and influences behavior change in a realistic setting. Copyright © 2016 Elsevier Inc. All rights reserved.
Graphical Evaluation of Hierarchical Clustering Schemes. Technical Report No. 1.
ERIC Educational Resources Information Center
Halff, Henry M.
Graphical methods for evaluating the fit of Johnson's hierarchical clustering schemes are presented together with an example. These evaluation methods examine the extent to which the clustering algorithm can minimize the overlap of the distributions of intracluster and intercluster distances. (Author)
NASA Astrophysics Data System (ADS)
Astionenko, I. O.; Litvinenko, O. I.; Osipova, N. V.; Tuluchenko, G. Ya.; Khomchenko, A. N.
2016-10-01
Recently the interpolation bases of the hierarchical type have been used for the problem solving of the approximation of multiple arguments functions (such as in the finite-element method). In this work the cognitive graphical method of constructing of the hierarchical form bases on the serendipity finite elements is suggested, which allowed to get the alternative bases on a biquadratic finite element from the serendipity family without internal knots' inclusion. The cognitive-graphic method allowed to improve the known interpolation procedure of Taylor and to get the modified elements with irregular arrangement of knots. The proposed procedures are universal and are spread in the area of finite-elements.
Umoquit, Muriah J; Dobrow, Mark J; Lemieux-Charles, Louise; Ritvo, Paul G; Urbach, David R; Wodchis, Walter P
2008-01-01
Background This paper focuses on measuring the efficiency and effectiveness of two diagramming methods employed in key informant interviews with clinicians and health care administrators. The two methods are 'participatory diagramming', where the respondent creates a diagram that assists in their communication of answers, and 'graphic elicitation', where a researcher-prepared diagram is used to stimulate data collection. Methods These two diagramming methods were applied in key informant interviews and their value in efficiently and effectively gathering data was assessed based on quantitative measures and qualitative observations. Results Assessment of the two diagramming methods suggests that participatory diagramming is an efficient method for collecting data in graphic form, but may not generate the depth of verbal response that many qualitative researchers seek. In contrast, graphic elicitation was more intuitive, better understood and preferred by most respondents, and often provided more contemplative verbal responses, however this was achieved at the expense of more interview time. Conclusion Diagramming methods are important for eliciting interview data that are often difficult to obtain through traditional verbal exchanges. Subject to the methodological limitations of the study, our findings suggest that while participatory diagramming and graphic elicitation have specific strengths and weaknesses, their combined use can provide complementary information that would not likely occur with the application of only one diagramming method. The methodological insights gained by examining the efficiency and effectiveness of these diagramming methods in our study should be helpful to other researchers considering their incorporation into qualitative research designs. PMID:18691410
Process for making carbon foam
Klett, James W.
2000-01-01
The process obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications.
Graphic Poetry: How to Help Students Get the Most out of Pictures
ERIC Educational Resources Information Center
Chiang, River Ya-ling
2013-01-01
This paper attempts to give an account of some innovative work in paintings and modern poetry and to show how modern poets, such as Jane Flanders and Anne Sexton, the two American poets in particular, express and develop radically new conventions for their respective arts. Also elaborated are how such changes in artistic techniques are related to…
Evaluation of an integrated graphical display to promote acute change detection in ICU patients
Anders, Shilo; Albert, Robert; Miller, Anne; Weinger, Matthew B.; Doig, Alexa K.; Behrens, Michael; Agutter, Jim
2012-01-01
Objective The purpose of this study was to evaluate ICU nurses’ ability to detect patient change using an integrated graphical information display (IGID) versus a conventional tabular ICU patient information display (i.e. electronic chart). Design Using participants from two different sites, we conducted a repeated measures simulator-based experiment to assess ICU nurses’ ability to detect abnormal patient variables using a novel IGID versus a conventional tabular information display. Patient scenarios and display presentations were fully counterbalanced. Measurements We measured percent correct detection of abnormal patient variables, nurses’ perceived workload (NASA-TLX), and display usability ratings. Results 32 ICU nurses (87% female, median age of 29 years, and median ICU experience of 2.5 years) using the IGID detected more abnormal variables compared to the tabular display [F (1,119)=13.0, p < 0.05]. There was a significant main effect of site [F (1, 119)=14.2], with development site participants doing better. There were no significant differences in nurses’ perceived workload. The IGID display was rated as more usable than the conventional display, [F (1, 60)=31.7]. Conclusion Overall, nurses reported more important physiological information with the novel IGID than tabular display. Moreover, the finding of site differences may reflect local influences in work practice and involvement in iterative display design methodology. Information displays developed using user-centered design should accommodate the full diversity of the intended user population across use sites. PMID:22534099
Distributed computation of graphics primitives on a transputer network
NASA Technical Reports Server (NTRS)
Ellis, Graham K.
1988-01-01
A method is developed for distributing the computation of graphics primitives on a parallel processing network. Off-the-shelf transputer boards are used to perform the graphics transformations and scan-conversion tasks that would normally be assigned to a single transputer based display processor. Each node in the network performs a single graphics primitive computation. Frequently requested tasks can be duplicated on several nodes. The results indicate that the current distribution of commands on the graphics network shows a performance degradation when compared to the graphics display board alone. A change to more computation per node for every communication (perform more complex tasks on each node) may cause the desired increase in throughput.
Graphical Method for Determining Projectile Trajectory
ERIC Educational Resources Information Center
Moore, J. C.; Baker, J. C.; Franzel, L.; McMahon, D.; Songer, D.
2010-01-01
We present a nontrigonometric graphical method for predicting the trajectory of a projectile when the angle and initial velocity are known. Students enrolled in a general education conceptual physics course typically have weak backgrounds in trigonometry, making inaccessible the standard analytical calculation of projectile range. Furthermore,…
Data graphing methods, articles of manufacture, and computing devices
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wong, Pak Chung; Mackey, Patrick S.; Cook, Kristin A.
Data graphing methods, articles of manufacture, and computing devices are described. In one aspect, a method includes accessing a data set, displaying a graphical representation including data of the data set which is arranged according to a first of different hierarchical levels, wherein the first hierarchical level represents the data at a first of a plurality of different resolutions which respectively correspond to respective ones of the hierarchical levels, selecting a portion of the graphical representation wherein the data of the portion is arranged according to the first hierarchical level at the first resolution, modifying the graphical representation by arrangingmore » the data of the portion according to a second of the hierarchal levels at a second of the resolutions, and after the modifying, displaying the graphical representation wherein the data of the portion is arranged according to the second hierarchal level at the second resolution.« less
Peng, Fei; Li, Jiao-ting; Long, Min
2015-03-01
To discriminate the acquisition pipelines of digital images, a novel scheme for the identification of natural images and computer-generated graphics is proposed based on statistical and textural features. First, the differences between them are investigated from the view of statistics and texture, and 31 dimensions of feature are acquired for identification. Then, LIBSVM is used for the classification. Finally, the experimental results are presented. The results show that it can achieve an identification accuracy of 97.89% for computer-generated graphics, and an identification accuracy of 97.75% for natural images. The analyses also demonstrate the proposed method has excellent performance, compared with some existing methods based only on statistical features or other features. The method has a great potential to be implemented for the identification of natural images and computer-generated graphics. © 2014 American Academy of Forensic Sciences.
Brain-computer interface: changes in performance using virtual reality techniques.
Ron-Angevin, Ricardo; Díaz-Estrella, Antonio
2009-01-09
The ability to control electroencephalographic (EEG) signals when different mental tasks are carried out would provide a method of communication for people with serious motor function problems. This system is known as a brain-computer interface (BCI). Due to the difficulty of controlling one's own EEG signals, a suitable training protocol is required to motivate subjects, as it is necessary to provide some type of visual feedback allowing subjects to see their progress. Conventional systems of feedback are based on simple visual presentations, such as a horizontal bar extension. However, virtual reality is a powerful tool with graphical possibilities to improve BCI-feedback presentation. The objective of the study is to explore the advantages of the use of feedback based on virtual reality techniques compared to conventional systems of feedback. Sixteen untrained subjects, divided into two groups, participated in the experiment. A group of subjects was trained using a BCI system, which uses conventional feedback (bar extension), and another group was trained using a BCI system, which submits subjects to a more familiar environment, such as controlling a car to avoid obstacles. The obtained results suggest that EEG behaviour can be modified via feedback presentation. Significant differences in classification error rates between both interfaces were obtained during the feedback period, confirming that an interface based on virtual reality techniques can improve the feedback control, specifically for untrained subjects.
An interactive graphics system to facilitate finite element structural analysis
NASA Technical Reports Server (NTRS)
Burk, R. C.; Held, F. H.
1973-01-01
The characteristics of an interactive graphics systems to facilitate the finite element method of structural analysis are described. The finite element model analysis consists of three phases: (1) preprocessing (model generation), (2) problem solution, and (3) postprocessing (interpretation of results). The advantages of interactive graphics to finite element structural analysis are defined.
Influence of Colour on Acquisition and Generalisation of Graphic Symbols
ERIC Educational Resources Information Center
Hetzroni, O. E.; Ne'eman, A.
2013-01-01
Background: Children with autism may benefit from using graphic symbols for their communication, language and literacy development. The purpose of this study was to investigate the influence of colour versus grey-scale displays on the identification of graphic symbols using a computer-based intervention. Method: An alternating treatment design was…
Fast algorithm for the rendering of three-dimensional surfaces
NASA Astrophysics Data System (ADS)
Pritt, Mark D.
1994-02-01
It is often desirable to draw a detailed and realistic representation of surface data on a computer graphics display. One such representation is a 3D shaded surface. Conventional techniques for rendering shaded surfaces are slow, however, and require substantial computational power. Furthermore, many techniques suffer from aliasing effects, which appear as jagged lines and edges. This paper describes an algorithm for the fast rendering of shaded surfaces without aliasing effects. It is much faster than conventional ray tracing and polygon-based rendering techniques and is suitable for interactive use. On an IBM RISC System/6000TM workstation it renders a 1000 X 1000 surface in about 7 seconds.
Graphic report of the results from propensity score method analyses.
Shrier, Ian; Pang, Menglan; Platt, Robert W
2017-08-01
To increase transparency in studies reporting propensity scores by using graphical methods that clearly illustrate (1) the number of participant exclusions that occur as a consequence of the analytic strategy and (2) whether treatment effects are constant or heterogeneous across propensity scores. We applied graphical methods to a real-world pharmacoepidemiologic study that evaluated the effect of initiating statin medication on the 1-year all-cause mortality post-myocardial infarction. We propose graphical methods to show the consequences of trimming and matching on the exclusion of participants from the analysis. We also propose the use of meta-analytical forest plots to show the magnitude of effect heterogeneity. A density plot with vertical lines demonstrated the proportion of subjects excluded because of trimming. A frequency plot with horizontal lines demonstrated the proportion of subjects excluded because of matching. An augmented forest plot illustrates the amount of effect heterogeneity present in the data. Our proposed techniques present additional and useful information that helps readers understand the sample that is analyzed with propensity score methods and whether effect heterogeneity is present. Copyright © 2017 Elsevier Inc. All rights reserved.
Pasion, Editha; Good, Levell; Tizon, Jisebelle; Krieger, Staci; O'Kier, Catherine; Taylor, Nicole; Johnson, Jennifer; Horton, Carrie M; Peterson, Mary
2010-11-01
To determine if the monitor cursor-line feature on bedside monitors is accurate for measuring central venous and pulmonary artery pressures in cardiac surgery patients. Central venous and pulmonary artery pressures were measured via 3 methods (end-expiratory graphic recording, monitor cursor-line display, and monitor digital display) in a convenience sample of postoperative cardiac surgery patients. Pressures were measured twice during both mechanical ventilation and spontaneous breathing. Analysis of variance was used to determine differences between measurement methods and the percentage of monitor pressures that differed by 4 mm Hg or more from the measurement obtained from the graphic recording. Significance level was set at P less than .05. Twenty-five patients were studied during mechanical ventilation (50 measurements) and 21 patients during spontaneous breathing (42 measurements). Measurements obtained via the 3 methods did not differ significantly for either type of pressure (P > .05). Graphically recorded pressures and measurements obtained via the monitor cursor-line or digital display methods differed by 4 mm Hg or more in 4% and 6% of measurements, respectively, during mechanical ventilation and 4% and 11%, respectively, during spontaneous breathing. The monitor cursor-line method for measuring central venous and pulmonary artery pressures may be a reasonable alternative to the end-expiratory graphic recording method in hemodynamically stable, postoperative cardiac surgery patients. Use of the digital display on the bedside monitor may result in larger discrepancies from the graphically recorded pressures than when the cursor-line method is used, particularly in spontaneously breathing patients.
Contemporary computerized methods for logging planning
Chris B. LeDoux
1988-01-01
Contemporary harvest planning graphic software is highlighted with practical applications. Planning results from a production study of the Clearwater Cable Yarder are summarized. Application of the planning methods to evaluation of proposed silvicultural treatments is included. Results show that 3-dimensional graphic analysis of proposed harvesting or silvicultural...
A new improved graphical and quantitative method for detecting bias in meta-analysis.
Furuya-Kanamori, Luis; Barendregt, Jan J; Doi, Suhail A R
2018-04-04
Detection of publication and related biases remains suboptimal and threatens the validity and interpretation of meta-analytical findings. When bias is present, it usually differentially affects small and large studies manifesting as an association between precision and effect size and therefore visual asymmetry of conventional funnel plots. This asymmetry can be quantified and Egger's regression is, by far, the most widely used statistical measure for quantifying funnel plot asymmetry. However, concerns have been raised about both the visual appearance of funnel plots and the sensitivity of Egger's regression to detect such asymmetry, particularly when the number of studies is small. In this article, we propose a new graphical method, the Doi plot, to visualize asymmetry and also a new measure, the LFK index, to detect and quantify asymmetry of study effects in Doi plots. We demonstrate that the visual representation of asymmetry was better for the Doi plot when compared with the funnel plot. We also show that the diagnostic accuracy of the LFK index in discriminating between asymmetry due to simulated publication bias versus chance or no asymmetry was also better with the LFK index which had areas under the receiver operating characteristic curve of 0.74-0.88 with simulations of meta-analyses with five, 10, 15, and 20 studies. The Egger's regression result had lower areas under the receiver operating characteristic curve values of 0.58-0.75 across the same simulations. The LFK index also had a higher sensitivity (71.3-72.1%) than the Egger's regression result (18.5-43.0%). We conclude that the methods proposed in this article can markedly improve the ability of researchers to detect bias in meta-analysis.
A review of contemporary methods for the presentation of scientific uncertainty.
Makinson, K A; Hamby, D M; Edwards, J A
2012-12-01
Graphic methods for displaying uncertainty are often the most concise and informative way to communicate abstract concepts. Presentation methods currently in use for the display and interpretation of scientific uncertainty are reviewed. Numerous subjective and objective uncertainty display methods are presented, including qualitative assessments, node and arrow diagrams, standard statistical methods, box-and-whisker plots,robustness and opportunity functions, contribution indexes, probability density functions, cumulative distribution functions, and graphical likelihood functions.
Zhang, Airong; Zhang, Song; Bian, Cuirong
2018-02-01
Cortical bone provides the main form of support in humans and other vertebrates against various forces. Thus, capturing its mechanical properties is important. In this study, the mechanical properties of cortical bone were investigated by using automated ball indentation and graphics processing at both the macroscopic and microstructural levels under dry conditions. First, all polished samples were photographed under a metallographic microscope, and the area ratio of the circumferential lamellae and osteons was calculated through the graphics processing method. Second, fully-computer-controlled automated ball indentation (ABI) tests were performed to explore the micro-mechanical properties of the cortical bone at room temperature and a constant indenter speed. The indentation defects were examined with a scanning electron microscope. Finally, the macroscopic mechanical properties of the cortical bone were estimated with the graphics processing method and mixture rule. Combining ABI and graphics processing proved to be an effective tool to obtaining the mechanical properties of the cortical bone, and the indenter size had a significant effect on the measurement. The methods presented in this paper provide an innovative approach to acquiring the macroscopic mechanical properties of cortical bone in a nondestructive manner. Copyright © 2017 Elsevier Ltd. All rights reserved.
A "Large and Graceful Sinuosity": John Herschel's Graphical Method
NASA Astrophysics Data System (ADS)
Hankins, Thomas L.
2006-12-01
In 1833 John Herschel published a graphical method for determining the orbits of double stars. He argued that this method, which depended on human judgment rather than mathematical analysis, gave better results than computation, given the uncertainty in the data. Herschel found that astronomy and terrestrial physics were especially suitable for graphical treatment, and he expected that graphs would soon become important in all areas of science. He argued with William Whewell and James D. Forbes over the process of induction, over the application of probability, and over the moral content of science. Graphs entered into all these debates, but because they constituted a method, not a metaphysics, they were acceptable to most practicing scientists and became increasingly popular throughout the nineteenth century.
The effect of using graphic organizers in the teaching of standard biology
NASA Astrophysics Data System (ADS)
Pepper, Wade Louis, Jr.
This study was conducted to determine if the use of graphic organizers in the teaching of standard biology would increase student achievement, involvement and quality of activities. The subjects were 10th grade standard biology students in a large southern inner city high school. The study was conducted over a six-week period in an instructional setting using action research as the investigative format. After calculation of the homogeneity between classes, random selection was used to determine the graphic organizer class and the control class. The graphic organizer class was taught unit material through a variety of instructional methods along with the use of teacher generated graphic organizers. The control class was taught the same unit material using the same instructional methods, but without the use of graphic organizers. Data for the study were gathered from in-class written assignments, teacher-generated tests and text-generated tests, and rubric scores of an out-of-class written assignment and project. Also, data were gathered from student reactions, comments, observations and a teacher's research journal. Results were analyzed using descriptive statistics and qualitative interpretation. By comparing statistical results, it was determined that the use of graphic organizers did not make a statistically significant difference in the understanding of biological concepts and retention of factual information. Furthermore, the use of graphic organizers did not make a significant difference in motivating students to fulfill all class assignments with quality efforts and products. However, based upon student reactions and comments along with observations by the researcher, graphic organizers were viewed by the students as a favorable and helpful instructional tool. In lieu of statistical results, student gains from instructional activities using graphic organizers were positive and merit the continuation of their use as an instructional tool.
ERIC Educational Resources Information Center
Bessey, Barbara L.; And Others
Graphical methods for displaying data, as well as available computer software and hardware, are reviewed. The authors have emphasized the types of graphs which are most relevant to the needs of the National Center for Education Statistics (NCES) and its readers. The following types of graphs are described: tabulations, stem-and-leaf displays,…
Introduction to 3D Graphics through Excel
ERIC Educational Resources Information Center
Benacka, Jan
2013-01-01
The article presents a method of explaining the principles of 3D graphics through making a revolvable and sizable orthographic parallel projection of cuboid in Excel. No programming is used. The method was tried in fourteen 90 minute lessons with 181 participants, which were Informatics teachers, undergraduates of Applied Informatics and gymnasium…
NASA Technical Reports Server (NTRS)
Lewis, Clayton; Wilde, Nick
1989-01-01
Space construction will require heavy investment in the development of a wide variety of user interfaces for the computer-based tools that will be involved at every stage of construction operations. Using today's technology, user interface development is very expensive for two reasons: (1) specialized and scarce programming skills are required to implement the necessary graphical representations and complex control regimes for high-quality interfaces; (2) iteration on prototypes is required to meet user and task requirements, since these are difficult to anticipate with current (and foreseeable) design knowledge. We are attacking this problem by building a user interface development tool based on extensions to the spreadsheet model of computation. The tool provides high-level support for graphical user interfaces and permits dynamic modification of interfaces, without requiring conventional programming concepts and skills.
CCP4i2: the new graphical user interface to the CCP4 program suite
Potterton, Liz; Ballard, Charles; Dodson, Eleanor; Evans, Phil R.; Keegan, Ronan; Krissinel, Eugene; Stevenson, Kyle; Lebedev, Andrey; McNicholas, Stuart J.; Noble, Martin; Pannu, Navraj S.; Roth, Christian; Sheldrick, George; Skubak, Pavol; Uski, Ville
2018-01-01
The CCP4 (Collaborative Computational Project, Number 4) software suite for macromolecular structure determination by X-ray crystallography groups brings together many programs and libraries that, by means of well established conventions, interoperate effectively without adhering to strict design guidelines. Because of this inherent flexibility, users are often presented with diverse, even divergent, choices for solving every type of problem. Recently, CCP4 introduced CCP4i2, a modern graphical interface designed to help structural biologists to navigate the process of structure determination, with an emphasis on pipelining and the streamlined presentation of results. In addition, CCP4i2 provides a framework for writing structure-solution scripts that can be built up incrementally to create increasingly automatic procedures. PMID:29533233
A graphical weather system design for the NASA transport systems research vehicle B-737
NASA Technical Reports Server (NTRS)
Scanlon, Charles H.
1992-01-01
A graphical weather system was designed for testing in the NASA Transport Systems Research Vehicle B-737 airplane and simulator. The purpose of these tests was to measure the impact of graphical weather products on aircrew decision processes, weather situation awareness, reroute clearances, workload, and weather monitoring. The flight crew graphical weather interface is described along with integration of the weather system with the flight navigation system, and data link transmission methods for sending weather data to the airplane.
Groundwater Resources Assessment under the Pressures of Humanity and Climate Changes
Bret Bruce; Diana Allen; Henrique Chaves; Gordon Grant; Gualbert Oude Essink; Henk Kooi; Ian White; Jason Gurdak; Jay Famiglietti; Jose Luis Martin-Bordes; Kevin Hiscock; Matthew Rodell; Neno Kukuric; Peter B. McMahon; Richard Taylor; Timothy Green; Yoseph Yechieli
2008-01-01
Given the vision and mission statements for GRAPHIC above, this document provides an updated framework for the GRAPHIC program. The approach to addressing global issues under the GRAPHIC umbrella involves case studies designed to cover a broad range of the identified Subjects, Methods, and Regions. Interdependencies of factors and processes affecting subsurface water...
Utilizing a Graphic Organizer for Promoting Pupils' Argumentation
ERIC Educational Resources Information Center
Hsieh, Fu-Pei; Lee, Sung-Tao
2011-01-01
The purpose of this study was utilizing a GO (graphic organizer) for promoting pupils' argumentation. The method of case study was employed. A total of eight fifth grade pupils from two classes were assigned (n = 4, two high achievers, two low achievers) with GOI (graphic organizer instruction), and the others (n = 4, 2 high achievers, 2 low…
Interplay of Computer and Paper-Based Sketching in Graphic Design
ERIC Educational Resources Information Center
Pan, Rui; Kuo, Shih-Ping; Strobel, Johannes
2013-01-01
The purpose of this study is to investigate student designers' attitude and choices towards the use of computers and paper sketches when involved in a graphic design process. 65 computer graphic technology undergraduates participated in this research. A mixed method study with survey and in-depth interviews was applied to answer the research…
Guo, Hongbin; Renaut, Rosemary A; Chen, Kewei; Reiman, Eric M
2010-01-01
Graphical analysis methods are widely used in positron emission tomography quantification because of their simplicity and model independence. But they may, particularly for reversible kinetics, lead to bias in the estimated parameters. The source of the bias is commonly attributed to noise in the data. Assuming a two-tissue compartmental model, we investigate the bias that originates from modeling error. This bias is an intrinsic property of the simplified linear models used for limited scan durations, and it is exaggerated by random noise and numerical quadrature error. Conditions are derived under which Logan's graphical method either over- or under-estimates the distribution volume in the noise-free case. The bias caused by modeling error is quantified analytically. The presented analysis shows that the bias of graphical methods is inversely proportional to the dissociation rate. Furthermore, visual examination of the linearity of the Logan plot is not sufficient for guaranteeing that equilibrium has been reached. A new model which retains the elegant properties of graphical analysis methods is presented, along with a numerical algorithm for its solution. We perform simulations with the fibrillar amyloid β radioligand [11C] benzothiazole-aniline using published data from the University of Pittsburgh and Rotterdam groups. The results show that the proposed method significantly reduces the bias due to modeling error. Moreover, the results for data acquired over a 70 minutes scan duration are at least as good as those obtained using existing methods for data acquired over a 90 minutes scan duration. PMID:20493196
An interactive medical image segmentation framework using iterative refinement.
Kalshetti, Pratik; Bundele, Manas; Rahangdale, Parag; Jangra, Dinesh; Chattopadhyay, Chiranjoy; Harit, Gaurav; Elhence, Abhay
2017-04-01
Segmentation is often performed on medical images for identifying diseases in clinical evaluation. Hence it has become one of the major research areas. Conventional image segmentation techniques are unable to provide satisfactory segmentation results for medical images as they contain irregularities. They need to be pre-processed before segmentation. In order to obtain the most suitable method for medical image segmentation, we propose MIST (Medical Image Segmentation Tool), a two stage algorithm. The first stage automatically generates a binary marker image of the region of interest using mathematical morphology. This marker serves as the mask image for the second stage which uses GrabCut to yield an efficient segmented result. The obtained result can be further refined by user interaction, which can be done using the proposed Graphical User Interface (GUI). Experimental results show that the proposed method is accurate and provides satisfactory segmentation results with minimum user interaction on medical as well as natural images. Copyright © 2017 Elsevier Ltd. All rights reserved.
Amorphous titania modified with boric acid for selective capture of glycoproteins.
Jin, Shanxia; Liu, Liping; Zhou, Ping
2018-05-22
Amorphous titania was modified with boric acid, and the resulting material was characterized by scanning electron microscopy, Fourier transform infrared spectroscopy, X-ray powder diffraction and X-ray photoelectron spectrometry. The new material, in contrast to conventional boronate affinity materials containing boronic acid ligands, bears boric acid groups. It is shown to exhibit high specificity for glycoproteins, and this was applied to design a method for solid phase extraction of glycoproteins as shown for ribonuclease B, horse radish peroxidase and ovalbumin. Glycoproteins were captured under slightly alkaline environment and released in acidic solutions. The glycoproteins extracted were detected by matrix-assisted laser desorption/ionization time-of-flight mass spectrometry. The binding capacities for ribonuclease B, horse radish peroxidase and ovalbumin typically are 9.3, 26.0 and 53.0 mg ∙ g -1 , respectively. The method was successfully applied to the selective enrichment of ovalbumin from egg white. Graphical abstract Schematic presentation of the capture of glycoproteins by amorphous titania modified with boric acid.
Semantically-Rigorous Systems Engineering Modeling Using Sysml and OWL
NASA Technical Reports Server (NTRS)
Jenkins, J. Steven; Rouquette, Nicolas F.
2012-01-01
The Systems Modeling Language (SysML) has found wide acceptance as a standard graphical notation for the domain of systems engineering. SysML subsets and extends the Unified Modeling Language (UML) to define conventions for expressing structural, behavioral, and analytical elements, and relationships among them. SysML-enabled modeling tools are available from multiple providers, and have been used for diverse projects in military aerospace, scientific exploration, and civil engineering. The Web Ontology Language (OWL) has found wide acceptance as a standard notation for knowledge representation. OWL-enabled modeling tools are available from multiple providers, as well as auxiliary assets such as reasoners and application programming interface libraries, etc. OWL has been applied to diverse projects in a wide array of fields. While the emphasis in SysML is on notation, SysML inherits (from UML) a semantic foundation that provides for limited reasoning and analysis. UML's partial formalization (FUML), however, does not cover the full semantics of SysML, which is a substantial impediment to developing high confidence in the soundness of any conclusions drawn therefrom. OWL, by contrast, was developed from the beginning on formal logical principles, and consequently provides strong support for verification of consistency and satisfiability, extraction of entailments, conjunctive query answering, etc. This emphasis on formal logic is counterbalanced by the absence of any graphical notation conventions in the OWL standards. Consequently, OWL has had only limited adoption in systems engineering. The complementary strengths and weaknesses of SysML and OWL motivate an interest in combining them in such a way that we can benefit from the attractive graphical notation of SysML and the formal reasoning of OWL. This paper describes an approach to achieving that combination.
Peng, Bo; Wang, Yuqi; Hall, Timothy J; Jiang, Jingfeng
2017-04-01
Our primary objective of this paper was to extend a previously published 2-D coupled subsample tracking algorithm for 3-D speckle tracking in the framework of ultrasound breast strain elastography. In order to overcome heavy computational cost, we investigated the use of a graphic processing unit (GPU) to accelerate the 3-D coupled subsample speckle tracking method. The performance of the proposed GPU implementation was tested using a tissue-mimicking phantom and in vivo breast ultrasound data. The performance of this 3-D subsample tracking algorithm was compared with the conventional 3-D quadratic subsample estimation algorithm. On the basis of these evaluations, we concluded that the GPU implementation of this 3-D subsample estimation algorithm can provide high-quality strain data (i.e., high correlation between the predeformation and the motion-compensated postdeformation radio frequency echo data and high contrast-to-noise ratio strain images), as compared with the conventional 3-D quadratic subsample algorithm. Using the GPU implementation of the 3-D speckle tracking algorithm, volumetric strain data can be achieved relatively fast (approximately 20 s per volume [2.5 cm ×2.5 cm ×2.5 cm]).
Johnson, Quentin R; Lindsay, Richard J; Shen, Tongye
2018-02-21
A computational method which extracts the dominant motions from an ensemble of biomolecular conformations via a correlation analysis of residue-residue contacts is presented. The algorithm first renders the structural information into contact matrices, then constructs the collective modes based on the correlated dynamics of a selected set of dynamic contacts. Associated programs can bridge the results for further visualization using graphics software. The aim of this method is to provide an analysis of conformations of biopolymers from the contact viewpoint. It may assist a systematical uncovering of conformational switching mechanisms existing in proteins and biopolymer systems in general by statistical analysis of simulation snapshots. In contrast to conventional correlation analyses of Cartesian coordinates (such as distance covariance analysis and Cartesian principal component analysis), this program also provides an alternative way to locate essential collective motions in general. Herein, we detail the algorithm in a stepwise manner and comment on the importance of the method as applied to decoding allosteric mechanisms. © 2018 Wiley Periodicals, Inc. © 2018 Wiley Periodicals, Inc.
Direct 4D reconstruction of parametric images incorporating anato-functional joint entropy.
Tang, Jing; Kuwabara, Hiroto; Wong, Dean F; Rahmim, Arman
2010-08-07
We developed an anatomy-guided 4D closed-form algorithm to directly reconstruct parametric images from projection data for (nearly) irreversible tracers. Conventional methods consist of individually reconstructing 2D/3D PET data, followed by graphical analysis on the sequence of reconstructed image frames. The proposed direct reconstruction approach maintains the simplicity and accuracy of the expectation-maximization (EM) algorithm by extending the system matrix to include the relation between the parametric images and the measured data. A closed-form solution was achieved using a different hidden complete-data formulation within the EM framework. Furthermore, the proposed method was extended to maximum a posterior reconstruction via incorporation of MR image information, taking the joint entropy between MR and parametric PET features as the prior. Using realistic simulated noisy [(11)C]-naltrindole PET and MR brain images/data, the quantitative performance of the proposed methods was investigated. Significant improvements in terms of noise versus bias performance were demonstrated when performing direct parametric reconstruction, and additionally upon extending the algorithm to its Bayesian counterpart using the MR-PET joint entropy measure.
Peng, Kuan; He, Ling; Zhu, Ziqiang; Tang, Jingtian; Xiao, Jiaying
2013-12-01
Compared with commonly used analytical reconstruction methods, the frequency-domain finite element method (FEM) based approach has proven to be an accurate and flexible algorithm for photoacoustic tomography. However, the FEM-based algorithm is computationally demanding, especially for three-dimensional cases. To enhance the algorithm's efficiency, in this work a parallel computational strategy is implemented in the framework of the FEM-based reconstruction algorithm using a graphic-processing-unit parallel frame named the "compute unified device architecture." A series of simulation experiments is carried out to test the accuracy and accelerating effect of the improved method. The results obtained indicate that the parallel calculation does not change the accuracy of the reconstruction algorithm, while its computational cost is significantly reduced by a factor of 38.9 with a GTX 580 graphics card using the improved method.
NASA Astrophysics Data System (ADS)
Park, Seyoun; Robinson, Adam; Quon, Harry; Kiess, Ana P.; Shen, Colette; Wong, John; Plishker, William; Shekhar, Raj; Lee, Junghoon
2016-03-01
In this paper, we propose a CT-CBCT registration method to accurately predict the tumor volume change based on daily cone-beam CTs (CBCTs) during radiotherapy. CBCT is commonly used to reduce patient setup error during radiotherapy, but its poor image quality impedes accurate monitoring of anatomical changes. Although physician's contours drawn on the planning CT can be automatically propagated to daily CBCTs by deformable image registration (DIR), artifacts in CBCT often cause undesirable errors. To improve the accuracy of the registration-based segmentation, we developed a DIR method that iteratively corrects CBCT intensities by local histogram matching. Three popular DIR algorithms (B-spline, demons, and optical flow) with the intensity correction were implemented on a graphics processing unit for efficient computation. We evaluated their performances on six head and neck (HN) cancer cases. For each case, four trained scientists manually contoured the nodal gross tumor volume (GTV) on the planning CT and every other fraction CBCTs to which the propagated GTV contours by DIR were compared. The performance was also compared with commercial image registration software based on conventional mutual information (MI), VelocityAI (Varian Medical Systems Inc.). The volume differences (mean±std in cc) between the average of the manual segmentations and automatic segmentations are 3.70+/-2.30 (B-spline), 1.25+/-1.78 (demons), 0.93+/-1.14 (optical flow), and 4.39+/-3.86 (VelocityAI). The proposed method significantly reduced the estimation error by 9% (B-spline), 38% (demons), and 51% (optical flow) over the results using VelocityAI. Although demonstrated only on HN nodal GTVs, the results imply that the proposed method can produce improved segmentation of other critical structures over conventional methods.
A k-Space Method for Moderately Nonlinear Wave Propagation
Jing, Yun; Wang, Tianren; Clement, Greg T.
2013-01-01
A k-space method for moderately nonlinear wave propagation in absorptive media is presented. The Westervelt equation is first transferred into k-space via Fourier transformation, and is solved by a modified wave-vector time-domain scheme. The present approach is not limited to forward propagation or parabolic approximation. One- and two-dimensional problems are investigated to verify the method by comparing results to analytic solutions and finite-difference time-domain (FDTD) method. It is found that to obtain accurate results in homogeneous media, the grid size can be as little as two points per wavelength, and for a moderately nonlinear problem, the Courant–Friedrichs–Lewy number can be as large as 0.4. Through comparisons with the conventional FDTD method, the k-space method for nonlinear wave propagation is shown here to be computationally more efficient and accurate. The k-space method is then employed to study three-dimensional nonlinear wave propagation through the skull, which shows that a relatively accurate focusing can be achieved in the brain at a high frequency by sending a low frequency from the transducer. Finally, implementations of the k-space method using a single graphics processing unit shows that it required about one-seventh the computation time of a single-core CPU calculation. PMID:22899114
ERIC Educational Resources Information Center
Zebehazy, Kim T.; Wilton, Adam P.
2014-01-01
Introduction: This study investigated the perceptions and practices of teachers of students with visual impairments in Canada and the United States regarding graphics (both tactile and print) that are used by students with visual impairments. Questions focused on quality, importance, and instruction in the use of graphics. Methods: An electronic…
ERIC Educational Resources Information Center
Schlosser, Ralf W.; Koul, Rajinder; Shane, Howard; Sorce, James; Brock, Kristofer; Harmon, Ashley; Moerlein, Dorothy; Hearn, Emilia
2014-01-01
Purpose: The effects of animation on naming and identification of graphic symbols for verbs and prepositions were studied in 2 graphic symbol sets in preschoolers. Method: Using a 2 × 2 × 2 × 3 completely randomized block design, preschoolers across three age groups were randomly assigned to combinations of symbol set (Autism Language Program…
Panesar, Sukhmeet S; Rao, Christopher; Vecht, Joshua A; Mirza, Saqeb B; Netuveli, Gopalakrishnan; Morris, Richard; Rosenthal, Joe; Darzi, Ara; Athanasiou, Thanos
2009-10-01
Meta-analyses may be prone to generating misleading results because of a paucity of experimental studies (especially in surgery); publication bias; and heterogeneity in study design, intervention and the patient population of included studies. When investigating a specific clinical or scientific question on which several relevant meta-analyses may have been published, value judgments must be applied to determine which analysis represents the most robust evidence. These value judgments should be specifically acknowledged. We designed the Veritas plot to explicitly explore important elements of quality and to facilitate decision-making by highlighting specific areas in which meta-analyses are found to be deficient. Furthermore, as a graphic tool, it may be more intuitive than when similar data are presented in a tabular or text format. The Veritas plot is an adaption of the radar plot, a graphic tool for the description of multiattribute data. Key elements of meta-analytical quality such as heterogeneity, publication bias and study design are assessed. Existing qualitative methods such as the Assessment of Multiple Systematic Reviews (AMSTAR) tool have been incorporated in addition to important considerations when interpreting surgical meta-analyses such as the year of publication and population characteristics. To demonstrate the potential of the Veritas plot to inform clinical practice, we apply the Veritas plot to the meta-analytical literature comparing the incidence of 30-day stroke in off-pump coronary artery bypass surgery and conventional coronary artery bypass surgery. We demonstrate that a visually-stimulating and practical evidence-synthesis tool can direct the clinician and scientist to a particular meta-analytical study to inform clinical practice. The Veritas plot is also cumulative and allowed us to assess the quality of evidence over time. We have presented a practical graphic application for scientists and clinicians to identify and interpret variability in meta-analyses. Although further validation of the Veritas plot is required, it may have the potential to contribute to the implementation of evidence-based practice.
Morin, Roger H.; Olsen, Harold W.; Nelson, Karl R.; Gill, James D.
1989-01-01
A graphical method has been developed for determining the coefficient of consolidation from the transient phases of a flow-pump permeability test. The flow pump can be used to infuse fluid into or withdraw fluid from a laboratory sediment specimen at a constant volumetric rate in order to obtain data that can be used to calculate permeability using Darcy's law. Representative type-curve solutions to the associated forced-flow and pressure-decay models are derived. These curves provide the basis for graphically evaluating the permeability k, the coefficient of consolidation cv, and the coefficient of volume change mv. The curve-matching technique is easy and rapid. Values of k, cv and mv for a laterally confined kaolinite specimen were determined by this graphical method and appear to be in reasonably good agreement with numerically derived estimates (within 20%). Discrepancies between the two sets of results seem to be largely a function of data quality.
Graphical function mapping as a new way to explore cause-and-effect chains
Evans, Mary Anne
2016-01-01
Graphical function mapping provides a simple method for improving communication within interdisciplinary research teams and between scientists and nonscientists. This article introduces graphical function mapping using two examples and discusses its usefulness. Function mapping projects the outcome of one function into another to show the combined effect. Using this mathematical property in a simpler, even cartoon-like, graphical way allows the rapid combination of multiple information sources (models, empirical data, expert judgment, and guesses) in an intuitive visual to promote further discussion, scenario development, and clear communication.
Kole, J S; Beekman, F J
2006-02-21
Statistical reconstruction methods offer possibilities to improve image quality as compared with analytical methods, but current reconstruction times prohibit routine application in clinical and micro-CT. In particular, for cone-beam x-ray CT, the use of graphics hardware has been proposed to accelerate the forward and back-projection operations, in order to reduce reconstruction times. In the past, wide application of this texture hardware mapping approach was hampered owing to limited intrinsic accuracy. Recently, however, floating point precision has become available in the latest generation commodity graphics cards. In this paper, we utilize this feature to construct a graphics hardware accelerated version of the ordered subset convex reconstruction algorithm. The aims of this paper are (i) to study the impact of using graphics hardware acceleration for statistical reconstruction on the reconstructed image accuracy and (ii) to measure the speed increase one can obtain by using graphics hardware acceleration. We compare the unaccelerated algorithm with the graphics hardware accelerated version, and for the latter we consider two different interpolation techniques. A simulation study of a micro-CT scanner with a mathematical phantom shows that at almost preserved reconstructed image accuracy, speed-ups of a factor 40 to 222 can be achieved, compared with the unaccelerated algorithm, and depending on the phantom and detector sizes. Reconstruction from physical phantom data reconfirms the usability of the accelerated algorithm for practical cases.
He, Yue-Jing; Hung, Wei-Chih; Lai, Zhe-Ping
2016-01-01
In this study, a numerical simulation method was employed to investigate and analyze superstructure fiber Bragg gratings (SFBGs) with five duty cycles (50%, 33.33%, 14.28%, 12.5%, and 10%). This study focuses on demonstrating the relevance between design period and spectral characteristics of SFBGs (in the form of graphics) for SFBGs of all duty cycles. Compared with complicated and hard-to-learn conventional coupled-mode theory, the result of the present study may assist beginner and expert designers in understanding the basic application aspects, optical characteristics, and design techniques of SFBGs, thereby indirectly lowering the physical concepts and mathematical skills required for entering the design field. To effectively improve the accuracy of overall computational performance and numerical calculations and to shorten the gap between simulation results and actual production, this study integrated a perfectly matched layer (PML), perfectly reflecting boundary (PRB), object meshing method (OMM), and boundary meshing method (BMM) into the finite element method (FEM) and eigenmode expansion method (EEM). The integrated method enables designers to easily and flexibly design optical fiber communication systems that conform to the specific spectral characteristic by using the simulation data in this paper, which includes bandwidth, number of channels, and band gap size. PMID:26861322
NASA Astrophysics Data System (ADS)
He, Yue-Jing; Hung, Wei-Chih; Syu, Cheng-Jyun
2017-12-01
The finite-element method (FEM) and eigenmode expansion method (EEM) were adopted to analyze the guided modes and spectrum of phase-shift fiber Bragg grating at five phase-shift degrees (including zero, 1/4π, 1/2π, 3/4π, and π). In previous studies on optical fiber grating, conventional coupled-mode theory was crucial. This theory contains abstruse knowledge about physics and complex computational processes, and thus is challenging for users. Therefore, a numerical simulation method was coupled with a simple and rigorous design procedure to help beginners and users to overcome difficulty in entering the field; in addition, graphical simulation results were presented. To reduce the difference between the simulated context and the actual context, a perfectly matched layer and perfectly reflecting boundary were added to the FEM and the EEM. When the FEM was used for grid cutting, the object meshing method and the boundary meshing method proposed in this study were used to effectively enhance computational accuracy and substantially reduce the time required for simulation. In summary, users can use the simulation results in this study to easily and rapidly design an optical fiber communication system and optical sensors with spectral characteristics.
A Work Station For Control Of Changing Systems
NASA Technical Reports Server (NTRS)
Mandl, Daniel J.
1988-01-01
Touch screen and microcomputer enable flexible control of complicated systems. Computer work station equipped to produce graphical displays used as command panel and status indicator for command-and-control system. Operator uses images of control buttons displayed on touch screen to send prestored commands. Use of prestored library of commands reduces incidence of errors. If necessary, operator uses conventional keyboard to enter commands in real time to handle unforeseeable situations.
Surgical Tooth Implants, Combat and Field.
1984-07-15
and identify by block number) --- This Annual Report summarizes progress to date on a long-term implant study of a serrated ceramic dental implant...upper two parts of the implant, post and core and crown, are conventional metaT materials. A series of graded dental implants have been produced to...throughout the experimental period. Periodic radio- graphic analyses of dental implants verify this observation. Gross and microscopic patho- logic analyses
Distributed Computerized Catalog System
NASA Technical Reports Server (NTRS)
Borgen, Richard L.; Wagner, David A.
1995-01-01
DarkStar Distributed Catalog System describes arbitrary data objects in unified manner, providing end users with versatile, yet simple search mechanism for locating and identifying objects. Provides built-in generic and dynamic graphical user interfaces. Design of system avoids some of problems of standard DBMS, and system provides more flexibility than do conventional relational data bases, or object-oriented data bases. Data-collection lattice partly hierarchical representation of relationships among collections, subcollections, and data objects.
VORTAB - A data-tablet method of developing input data for the VORLAX program
NASA Technical Reports Server (NTRS)
Denn, F. M.
1979-01-01
A method of developing an input data file for use in the aerodynamic analysis of a complete airplane with the VORLAX computer program is described. The hardware consists of an interactive graphics terminal equipped with a graphics tablet. Software includes graphics routines from the Tektronix PLOT 10 package as well as the VORTAB program described. The user determines the size and location of each of the major panels for the aircraft before using the program. Data is entered both from the terminal keyboard and the graphics tablet. The size of the resulting data file is dependent on the complexity of the model and can vary from ten to several hundred card images. After the data are entered, two programs READB and PLOTB, are executed which plot the configuration allowing visual inspection of the model.
Wahab, M Farooq; Patel, Darshan C; Armstrong, Daniel W
2017-08-04
Most peak shapes obtained in separation science depart from linearity for various reasons such as thermodynamic, kinetic, or flow based effects. An indication of the nature of asymmetry often helps in problem solving e.g. in column overloading, slurry packing, buffer mismatch, and extra-column band broadening. However, existing tests for symmetry/asymmetry only indicate the skewness in excess (tail or front) and not the presence of both. Two simple graphical approaches are presented to analyze peak shapes typically observed in gas, liquid, and supercritical fluid chromatography as well as capillary electrophoresis. The derivative test relies on the symmetry of the inflection points and the maximum and minimum values of the derivative. The Gaussian test is a constrained curve fitting approach and determines the residuals. The residual pattern graphically allows the user to assess the problematic regions in a given peak, e.g., concurrent tailing or fronting, something which cannot be easily done with other current methods. The template provided in MS Excel automates this process. The total peak shape analysis extracts the peak parameters from the upper sections (>80% height) of the peak rather than the half height as is done conventionally. A number of situations are presented and the utility of this approach in solving practical problems is demonstrated. Copyright © 2017 Elsevier B.V. All rights reserved.
Hiilamo, Heikki; Crosbie, Eric; Glantz, Stanton A
2013-01-01
Objective To analyse the evolution and diffusion of health warnings on cigarette packs around the world, including tobacco industry attempts to block this diffusion. Methods We analysed tobacco industry documents and public sources to construct a database on the global evolution and diffusion of health warning labels from 1966 to 2012, and also analysed industry strategies. Results Health warning labels, especially labels with graphic elements, threaten the tobacco industry because they are a low-cost, effective measure to reduce smoking. Multinational tobacco companies did not object to voluntary innocuous warnings with ambiguous health messages, in part because they saw them as offering protection from lawsuits and local packaging regulations. The companies worked systematically at the international level to block or weaken warnings once stronger more specific warnings began to appear in the 1970s. Since 1985 in Iceland, the tobacco industry has been aware of the effectiveness of graphic health warning labels (GWHL). The industry launched an all-out attack in the early 1990s to prevent GHWLs, and was successful in delaying GHWLs internationally for nearly 10 years. Conclusions Beginning in 2005, as a result of the World Health Organisation Framework Convention on Tobacco Control (FCTC), GHWLs began to spread. Effective implementation of FCTC labelling provisions has stimulated diffusion of strong health warning labels despite industry opposition. PMID:23092884
A Guide to the Literature on Learning Graphical Models
NASA Technical Reports Server (NTRS)
Buntine, Wray L.; Friedland, Peter (Technical Monitor)
1994-01-01
This literature review discusses different methods under the general rubric of learning Bayesian networks from data, and more generally, learning probabilistic graphical models. Because many problems in artificial intelligence, statistics and neural networks can be represented as a probabilistic graphical model, this area provides a unifying perspective on learning. This paper organizes the research in this area along methodological lines of increasing complexity.
Area-delay trade-offs of texture decompressors for a graphics processing unit
NASA Astrophysics Data System (ADS)
Novoa Súñer, Emilio; Ituero, Pablo; López-Vallejo, Marisa
2011-05-01
Graphics Processing Units have become a booster for the microelectronics industry. However, due to intellectual property issues, there is a serious lack of information on implementation details of the hardware architecture that is behind GPUs. For instance, the way texture is handled and decompressed in a GPU to reduce bandwidth usage has never been dealt with in depth from a hardware point of view. This work addresses a comparative study on the hardware implementation of different texture decompression algorithms for both conventional (PCs and video game consoles) and mobile platforms. Circuit synthesis is performed targeting both a reconfigurable hardware platform and a 90nm standard cell library. Area-delay trade-offs have been extensively analyzed, which allows us to compare the complexity of decompressors and thus determine suitability of algorithms for systems with limited hardware resources.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hatakeyama, Keisuke, E-mail: hatakeyamak@pref.tottori.jp; Okuda, Masukazu; Kuki, Takahiro
2012-12-15
Graphical abstract: Display Omitted Highlights: ► The photocatalytic property of a silver orthophosphate (Ag{sub 3}PO{sub 4}) was investigated for humic acid degradation. ► The Ag{sub 3}PO{sub 4} shows high photocatalytic activity under visible light. ► The photocatalytic activity was greatly improved by employing the precipitation method. -- Abstract: In order to remove dissolved organic matter such as humic acid from water, a silver orthophosphate (Ag{sub 3}PO{sub 4}) was newly employed as a heterogeneous photocatalyst. Here, Ag{sub 3}PO{sub 4} was prepared by simple ion-exchange and precipitation methods, and the physico-chemical properties were characterized by X-ray diffraction, ultraviolet–visible diffuse reflectance spectroscopy, scanningmore » electron microscopy, particle distribution measurements and Brunauer–Emmett–Teller (BET) analysis. The degradation of humic acid was faster over Ag{sub 3}PO{sub 4} catalyst than over conventional TiO{sub 2} (P-25). The total photocatalytic properties were improved by employing not an ion-exchange method but a precipitation method; humic acid degradation was performed with a removal ratio of dissolved organic carbon of 75% under visible light (λ = 451 nm) for 2-h irradiation.« less
Yang, Hyeri; Na, Jihye; Jang, Won-Hee; Jung, Mi-Sook; Jeon, Jun-Young; Heo, Yong; Yeo, Kyung-Wook; Jo, Ji-Hoon; Lim, Kyung-Min; Bae, SeungJin
2015-05-05
Mouse local lymph node assay (LLNA, OECD TG429) is an alternative test replacing conventional guinea pig tests (OECD TG406) for the skin sensitization test but the use of a radioisotopic agent, (3)H-thymidine, deters its active dissemination. New non-radioisotopic LLNA, LLNA:BrdU-FCM employs a non-radioisotopic analog, 5-bromo-2'-deoxyuridine (BrdU) and flow cytometry. For an analogous method, OECD TG429 performance standard (PS) advises that two reference compounds be tested repeatedly and ECt(threshold) values obtained must fall within acceptable ranges to prove within- and between-laboratory reproducibility. However, this criteria is somewhat arbitrary and sample size of ECt is less than 5, raising concerns about insufficient reliability. Here, we explored various statistical methods to evaluate the reproducibility of LLNA:BrdU-FCM with stimulation index (SI), the raw data for ECt calculation, produced from 3 laboratories. Descriptive statistics along with graphical representation of SI was presented. For inferential statistics, parametric and non-parametric methods were applied to test the reproducibility of SI of a concurrent positive control and the robustness of results were investigated. Descriptive statistics and graphical representation of SI alone could illustrate the within- and between-laboratory reproducibility. Inferential statistics employing parametric and nonparametric methods drew similar conclusion. While all labs passed within- and between-laboratory reproducibility criteria given by OECD TG429 PS based on ECt values, statistical evaluation based on SI values showed that only two labs succeeded in achieving within-laboratory reproducibility. For those two labs that satisfied the within-lab reproducibility, between-laboratory reproducibility could be also attained based on inferential as well as descriptive statistics. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Vivier, L.
2013-07-01
Usually, the tangent line is considered to be a calculus notion. However, it is also a graphical and an algebraic notion. The graphical frame, where our primary conceptions are conceived, could give rise to algebraic methods to obtain the tangent line to a curve. In this pre-calculus perspective, two methods are described and discussed according to their potential for secondary students and teacher training.
The End of the Rainbow? Color Schemes for Improved Data Graphics
NASA Astrophysics Data System (ADS)
Light, Adam; Bartlein, Patrick J.
2004-10-01
Modern computer displays and printers enable the widespread use of color in scientific communication, but the expertise for designing effective graphics has not kept pace with the technology for producing them. Historically, even the most prestigious publications have tolerated high defect rates in figures and illustrations, and technological advances that make creating and reproducing graphics easier do not appear to have decreased the frequency of errors. Flawed graphics consequently beget more flawed graphics as authors emulate published examples. Color has the potential to enhance communication, but design mistakes can result in color figures that are less effective than gray scale displays of the same data. Empirical research on human subjects can build a fundamental understanding of visual perception and scientific methods can be used to evaluate existing designs, but creating effective data graphics is a design task and not fundamentally a scientific pursuit. Like writing well, creating good data graphics requires a combination of formal knowledge and artistic sensibility tempered by experience: a combination of ``substance, statistics, and design''.
Giang, Kim Bao; Chung, Le Hong; Minh, Hoang Van; Kien, Vu Duy; Giap, Vu Van; Hinh, Nguyen Duc; Cuong, Nguyen Manh; Manh, Pham Duc; Duc, Ha Anh; Yang, Jui-Chen
2016-01-01
Graphic health warnings (GHW) on tobacco packages have proven to be effective in increasing quit attempts among smokers and reducing initial smoking among adolescents. This research aimed to examine the relative importance of different attributes of graphic health warnings on tobacco packages in Viet Nam. A discrete choice experimental (DCE) design was applied with a conditional logit model. In addition, a ranking method was used to list from the least to the most dreadful GHW labels. With the results from DCE model, graphic type was shown to be the most important attribute, followed by cost and coverage area of GHW. The least important attribute was position of the GHW. Among 5 graphic types (internal lung cancer image, external damaged teeth, abstract image, human suffering image and text), the image of lung cancer was found to have the strongest influence on both smokers and non-smokers. With ranking method, the image of throat cancer and heart diseases were considered the most dreadful images. GHWs should be designed with these attributes in mind, to maximise influence on purchase among both smokers and non-smokers.
Visualizing Astronomical Data with Blender
NASA Astrophysics Data System (ADS)
Kent, Brian R.
2014-01-01
We present methods for using the 3D graphics program Blender in the visualization of astronomical data. The software's forte for animating 3D data lends itself well to use in astronomy. The Blender graphical user interface and Python scripting capabilities can be utilized in the generation of models for data cubes, catalogs, simulations, and surface maps. We review methods for data import, 2D and 3D voxel texture applications, animations, camera movement, and composite renders. Rendering times can be improved by using graphic processing units (GPUs). A number of examples are shown using the software features most applicable to various kinds of data paradigms in astronomy.
LinkWinds: An Approach to Visual Data Analysis
NASA Technical Reports Server (NTRS)
Jacobson, Allan S.
1992-01-01
The Linked Windows Interactive Data System (LinkWinds) is a prototype visual data exploration and analysis system resulting from a NASA/JPL program of research into graphical methods for rapidly accessing, displaying and analyzing large multivariate multidisciplinary datasets. It is an integrated multi-application execution environment allowing the dynamic interconnection of multiple windows containing visual displays and/or controls through a data-linking paradigm. This paradigm, which results in a system much like a graphical spreadsheet, is not only a powerful method for organizing large amounts of data for analysis, but provides a highly intuitive, easy to learn user interface on top of the traditional graphical user interface.
Apparatus and method for implementing power saving techniques when processing floating point values
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Young Moon; Park, Sang Phill
An apparatus and method are described for reducing power when reading and writing graphics data. For example, one embodiment of an apparatus comprises: a graphics processor unit (GPU) to process graphics data including floating point data; a set of registers, at least one of the registers of the set partitioned to store the floating point data; and encode/decode logic to reduce a number of binary 1 values being read from the at least one register by causing a specified set of bit positions within the floating point data to be read out as 0s rather than 1s.
An application of interactive graphics to neutron spectrometry
NASA Technical Reports Server (NTRS)
Binney, S. E.
1972-01-01
The use of interactive graphics is presented as an attractive method for performing multi-parameter data analysis of proton recoil distributions to determine neutron spectra. Interactive graphics allows the user to view results on-line as the program is running and to maintain maximum control over the path along which the calculation will proceed. Other advantages include less time to obtain results and freedom from handling paper tapes and IBM cards.
Probabilistic Methods for Image Generation and Encoding.
1993-10-15
video and graphics lab at Georgia Tech, linking together Silicon Graphics workstations, a laser video recorder, a Betacam video recorder, scanner...computer laboratory at Georgia Tech, based on two Silicon Graphics Personal Iris workstations, a SONY laser video recorder, a SONY Betacam SP video...laser disk in component RGB form, with variable speed playback. From the laser recorder the images can be dubbed to the Betacam or the VHS recorder in
Purohit, Bharathi M; Singh, Abhinav; Dwivedi, Ashish
2017-03-01
The study aims to assess the reliability of video-graphic method as a tool to screen the dental caries among 12-year-old school children in a rural region of India. A total of 139 school children participated in the study. Visual tactile examinations were conducted using the Decayed, Missing, and Filled Teeth (DMFT) index. Simultaneously, standardized video recording of the oral cavity was performed. Sensitivity and specificity values were calculated for video-graphic assessment of dental caries. Bland-Altman plot was used to assess agreement between the two methods of caries assessment. Likelihood ratio (LR) and receiver-operating characteristic (ROC) curve were used to assess the predictive accuracy of the video-graphic method. Mean DMFT for the study population was 2.47 ± 2.01 and 2.46 ± 1.91 by visual tactile and video-graphic assessment (P = 0.76; > 0.05). Sensitivity and specificity values of 0.86 and 0.58 were established for video-graphic assessment. A fair degree of agreement was noted between the two methods with Intraclass correlation coefficient (ICC) value of 0.56. LR for video-graphic assessment was 2.05. Bland-Altman plot confirmed the level of agreement between the two assessment methods. The area under curve was 0.69 (CI 0.57, 0.80, P = 0.001). Teledentistry examination is comparable to clinical examination when screening for dental caries among school children. This study provides evidence that teledentistry may be used as an alternative screening tool for assessment of dental caries and is viable for remote consultation and treatment planning. Teledentistry offers to change the dynamics of dental care delivery and may effectively bridge the rural-urban oral health divide. © 2016 American Association of Public Health Dentistry.
Motion compensation in digital subtraction angiography using graphics hardware.
Deuerling-Zheng, Yu; Lell, Michael; Galant, Adam; Hornegger, Joachim
2006-07-01
An inherent disadvantage of digital subtraction angiography (DSA) is its sensitivity to patient motion which causes artifacts in the subtraction images. These artifacts could often reduce the diagnostic value of this technique. Automated, fast and accurate motion compensation is therefore required. To cope with this requirement, we first examine a method explicitly designed to detect local motions in DSA. Then, we implement a motion compensation algorithm by means of block matching on modern graphics hardware. Both methods search for maximal local similarity by evaluating a histogram-based measure. In this context, we are the first who have mapped an optimizing search strategy on graphics hardware while paralleling block matching. Moreover, we provide an innovative method for creating histograms on graphics hardware with vertex texturing and frame buffer blending. It turns out that both methods can effectively correct the artifacts in most case, as the hardware implementation of block matching performs much faster: the displacements of two 1024 x 1024 images can be calculated at 3 frames/s with integer precision or 2 frames/s with sub-pixel precision. Preliminary clinical evaluation indicates that the computation with integer precision could already be sufficient.
MGUPGMA: A Fast UPGMA Algorithm With Multiple Graphics Processing Units Using NCCL
Hua, Guan-Jie; Hung, Che-Lun; Lin, Chun-Yuan; Wu, Fu-Che; Chan, Yu-Wei; Tang, Chuan Yi
2017-01-01
A phylogenetic tree is a visual diagram of the relationship between a set of biological species. The scientists usually use it to analyze many characteristics of the species. The distance-matrix methods, such as Unweighted Pair Group Method with Arithmetic Mean and Neighbor Joining, construct a phylogenetic tree by calculating pairwise genetic distances between taxa. These methods have the computational performance issue. Although several new methods with high-performance hardware and frameworks have been proposed, the issue still exists. In this work, a novel parallel Unweighted Pair Group Method with Arithmetic Mean approach on multiple Graphics Processing Units is proposed to construct a phylogenetic tree from extremely large set of sequences. The experimental results present that the proposed approach on a DGX-1 server with 8 NVIDIA P100 graphic cards achieves approximately 3-fold to 7-fold speedup over the implementation of Unweighted Pair Group Method with Arithmetic Mean on a modern CPU and a single GPU, respectively. PMID:29051701
MGUPGMA: A Fast UPGMA Algorithm With Multiple Graphics Processing Units Using NCCL.
Hua, Guan-Jie; Hung, Che-Lun; Lin, Chun-Yuan; Wu, Fu-Che; Chan, Yu-Wei; Tang, Chuan Yi
2017-01-01
A phylogenetic tree is a visual diagram of the relationship between a set of biological species. The scientists usually use it to analyze many characteristics of the species. The distance-matrix methods, such as Unweighted Pair Group Method with Arithmetic Mean and Neighbor Joining, construct a phylogenetic tree by calculating pairwise genetic distances between taxa. These methods have the computational performance issue. Although several new methods with high-performance hardware and frameworks have been proposed, the issue still exists. In this work, a novel parallel Unweighted Pair Group Method with Arithmetic Mean approach on multiple Graphics Processing Units is proposed to construct a phylogenetic tree from extremely large set of sequences. The experimental results present that the proposed approach on a DGX-1 server with 8 NVIDIA P100 graphic cards achieves approximately 3-fold to 7-fold speedup over the implementation of Unweighted Pair Group Method with Arithmetic Mean on a modern CPU and a single GPU, respectively.
Methods and apparatus for graphical display and editing of flight plans
NASA Technical Reports Server (NTRS)
Gibbs, Michael J. (Inventor); Adams, Jr., Mike B. (Inventor); Chase, Karl L. (Inventor); Lewis, Daniel E. (Inventor); McCrobie, Daniel E. (Inventor); Omen, Debi Van (Inventor)
2002-01-01
Systems and methods are provided for an integrated graphical user interface which facilitates the display and editing of aircraft flight-plan data. A user (e.g., a pilot) located within the aircraft provides input to a processor through a cursor control device and receives visual feedback via a display produced by a monitor. The display includes various graphical elements associated with the lateral position, vertical position, flight-plan and/or other indicia of the aircraft's operational state as determined from avionics data and/or various data sources. Through use of the cursor control device, the user may modify the flight-plan and/or other such indicia graphically in accordance with feedback provided by the display. In one embodiment, the display includes a lateral view, a vertical profile view, and a hot-map view configured to simplify the display and editing of the aircraft's flight-plan data.
HeNCE: A Heterogeneous Network Computing Environment
Beguelin, Adam; Dongarra, Jack J.; Geist, George Al; ...
1994-01-01
Network computing seeks to utilize the aggregate resources of many networked computers to solve a single problem. In so doing it is often possible to obtain supercomputer performance from an inexpensive local area network. The drawback is that network computing is complicated and error prone when done by hand, especially if the computers have different operating systems and data formats and are thus heterogeneous. The heterogeneous network computing environment (HeNCE) is an integrated graphical environment for creating and running parallel programs over a heterogeneous collection of computers. It is built on a lower level package called parallel virtual machine (PVM).more » The HeNCE philosophy of parallel programming is to have the programmer graphically specify the parallelism of a computation and to automate, as much as possible, the tasks of writing, compiling, executing, debugging, and tracing the network computation. Key to HeNCE is a graphical language based on directed graphs that describe the parallelism and data dependencies of an application. Nodes in the graphs represent conventional Fortran or C subroutines and the arcs represent data and control flow. This article describes the present state of HeNCE, its capabilities, limitations, and areas of future research.« less
Considerations on the construction of a Powder Bed Fusion platform for Additive Manufacturing
NASA Astrophysics Data System (ADS)
Andersen, Sebastian Aagaard; Nielsen, Karl-Emil; Pedersen, David Bue; Nielsen, Jakob Skov
As the demand for moulds and other tools becomes increasingly specific and complex, an additive manufacturing approach to production is making its way to the industry through laser based consolidation of metal powder particles by a method known as powder bed fusion. This paper concerns a variety of design choices facilitating the development of an experimental powder bed fusion machine tool, capable of manufacturing metal parts with strength matching that of conventional manufactured parts and a complexity surpassing that of subtractive processes. To understand the different mechanisms acting within such an experimental machine tool, a fully open and customizable rig is constructed. Emphasizing modularity in the rig, allows alternation of lasers, scanner systems, optical elements, powder deposition, layer height, temperature, atmosphere, and powder type. Through a custom-made software platform, control of the process is achieved, which extends into a graphical user interface, easing adjustment of process parameters and the job file generation.
Nawyn, John P.; Sargent, B. Pierre; Hoopes, Barbara; Augenstein, Todd; Rowland, Kathleen M.; Barber, Nancy L.
2017-10-06
The Aggregate Water-Use Data System (AWUDS) is the database management system used to enter, store, and analyze state aggregate water-use data. It is part of the U.S. Geological Survey National Water Information System. AWUDS has a graphical user interface that facilitates data entry, revision, review, and approval. This document provides information on the basic functions of AWUDS and the steps for carrying out common tasks that are a part of compiling an aggregated dataset. Also included are explanations of terminology and descriptions of user-interface structure, procedures for using the AWUDS operations, and dataset-naming conventions. Information on water-use category definitions, data-collection methods, and data sources are found in the report “Guidelines for preparation of State water-use estimates,” available at https://pubs.er.usgs.gov/publication/ofr20171029.
CPU architecture for a fast and energy-saving calculation of convolution neural networks
NASA Astrophysics Data System (ADS)
Knoll, Florian J.; Grelcke, Michael; Czymmek, Vitali; Holtorf, Tim; Hussmann, Stephan
2017-06-01
One of the most difficult problem in the use of artificial neural networks is the computational capacity. Although large search engine companies own specially developed hardware to provide the necessary computing power, for the conventional user only remains the state of the art method, which is the use of a graphic processing unit (GPU) as a computational basis. Although these processors are well suited for large matrix computations, they need massive energy. Therefore a new processor on the basis of a field programmable gate array (FPGA) has been developed and is optimized for the application of deep learning. This processor is presented in this paper. The processor can be adapted for a particular application (in this paper to an organic farming application). The power consumption is only a fraction of a GPU application and should therefore be well suited for energy-saving applications.
Storage strategies of eddy-current FE-BI model for GPU implementation
NASA Astrophysics Data System (ADS)
Bardel, Charles; Lei, Naiguang; Udpa, Lalita
2013-01-01
In the past few years graphical processing units (GPUs) have shown tremendous improvements in computational throughput over standard CPU architecture. However, this comes at the cost of restructuring the algorithms to meet the strengths and drawbacks of this GPU architecture. A major drawback is the state of limited memory, and hence storage of FE stiffness matrices on the GPU is important. In contrast to storage on CPU the GPU storage format has significant influence on the overall performance. This paper presents an investigation of a storage strategy in the implementation of a two-dimensional finite element-boundary integral (FE-BI) model for Eddy current NDE applications, on GPU architecture. Specifically, the high dimensional matrices are manipulated by examining the matrix structure and optimally splitting into structurally independent component matrices for efficient storage and retrieval of each component. Results obtained using the proposed approach are compared to those of conventional CPU implementation for validating the method.
High-performance dynamic quantum clustering on graphics processors
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wittek, Peter, E-mail: peterwittek@acm.org
2013-01-15
Clustering methods in machine learning may benefit from borrowing metaphors from physics. Dynamic quantum clustering associates a Gaussian wave packet with the multidimensional data points and regards them as eigenfunctions of the Schroedinger equation. The clustering structure emerges by letting the system evolve and the visual nature of the algorithm has been shown to be useful in a range of applications. Furthermore, the method only uses matrix operations, which readily lend themselves to parallelization. In this paper, we develop an implementation on graphics hardware and investigate how this approach can accelerate the computations. We achieve a speedup of up tomore » two magnitudes over a multicore CPU implementation, which proves that quantum-like methods and acceleration by graphics processing units have a great relevance to machine learning.« less
Pitch-based carbon foam and composites
Klett, James W.
2001-01-01
A process for producing carbon foam or a composite is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications.
Pitch-based carbon foam and composites
Klett, James W.
2003-12-16
A process for producing carbon foam or a composite is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications.
Pitch-based carbon foam and composites
Klett, James W.
2003-12-02
A process for producing carbon foam or a composite is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications.
Pitch-based carbon foam and composites
Klett, James W.
2002-01-01
A process for producing carbon foam or a composite is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications.
van Eck, Herman J; Vos, Peter G; Valkonen, Jari P T; Uitdewilligen, Jan G A M L; Lensing, Hellen; de Vetten, Nick; Visser, Richard G F
2017-03-01
The method of graphical genotyping is applied to a panel of tetraploid potato cultivars to visualize haplotype sharing. The method allowed to map genes involved in virus and nematode resistance. The physical coordinates of the amount of linkage drag surrounding these genes are easily interpretable. Graphical genotyping is a visually attractive and easily interpretable method to represent genetic marker data. In this paper, the method is extended from diploids to a panel of tetraploid potato cultivars. Application of filters to select a subset of SNPs allows one to visualize haplotype sharing between individuals that also share a specific locus. The method is illustrated with cultivars resistant to Potato virus Y (PVY), while simultaneously selecting for the absence of the SNPs in susceptible clones. SNP data will then merge into an image which displays the coordinates of a distal genomic region on the northern arm of chromosome 11 where a specific haplotype is introgressed from the wild potato species S. stoloniferum (CPC 2093) carrying a gene (Ny (o,n)sto ) conferring resistance to two PVY strains, PVY O and PVY NTN . Graphical genotyping was also successful in showing the haplotypes on chromosome 12 carrying Ry-f sto , another resistance gene derived from S. stoloniferum conferring broad-spectrum resistance to PVY, as well as chromosome 5 haplotypes from S. vernei, with the Gpa5 locus involved in resistance against Globodera pallida cyst nematodes. The image also shows shortening of linkage drag by meiotic recombination of the introgression segment in more recent breeding material. Identity-by-descent was found to be a requirement for using graphical genotyping, which is proposed as a non-statistical alternative method for gene discovery, as compared with genome-wide association studies. The potential and limitations of the method are discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Herrero, M.; Martinez-Gallegos, S.; Labajos, F.M.
2011-11-15
Conventional and microwave heating routes have been used to prepare PET-LDH (polyethylene terephthalate-layered double hydroxide) composites with 1-10 wt% LDH by in situ polymerization. To enhance the compatibility between PET and the LDH, terephthalate or dodecyl sulphate had been previously intercalated in the LDH. PXRD and TEM were used to detect the degree of dispersion of the filler and the type of the polymeric composites obtained, and FTIR spectroscopy confirmed that the polymerization process had taken place. The thermal stability of these composites, as studied by thermogravimetric analysis, was enhanced when the microwave heating method was applied. Dodecyl sulphate wasmore » more effective than terephthalate to exfoliate the samples, which only occurred for the terephthalate ones under microwave irradiation. - Graphical abstract: Conventional and microwave heating routes were used to prepare PET-LDH (polyethylene terephthalate-layered double hydroxide) composites with 1-10 wt% LDH by in situ polymerization. To enhance the compatibility between PET and the LDH, terephthalate or dodecyl sulphate was previously intercalated into the LDH. The microwave process improves the dispersion and the thermal stability of nanocomposites due to the interaction of the microwave radiation and the dipolar properties of EG and the homogeneous heating. Highlights: > LDH-PET compatibility is enhanced by preintercalation of organic anions. > Dodecylsulphate performance is much better than that of terephthalate. > Microwave heating improves the thermal stability of the composites. > Microwave heating improves as well the dispersion of the inorganic phase.« less
García-Calvo, Raúl; Guisado, JL; Diaz-del-Rio, Fernando; Córdoba, Antonio; Jiménez-Morales, Francisco
2018-01-01
Understanding the regulation of gene expression is one of the key problems in current biology. A promising method for that purpose is the determination of the temporal dynamics between known initial and ending network states, by using simple acting rules. The huge amount of rule combinations and the nonlinear inherent nature of the problem make genetic algorithms an excellent candidate for finding optimal solutions. As this is a computationally intensive problem that needs long runtimes in conventional architectures for realistic network sizes, it is fundamental to accelerate this task. In this article, we study how to develop efficient parallel implementations of this method for the fine-grained parallel architecture of graphics processing units (GPUs) using the compute unified device architecture (CUDA) platform. An exhaustive and methodical study of various parallel genetic algorithm schemes—master-slave, island, cellular, and hybrid models, and various individual selection methods (roulette, elitist)—is carried out for this problem. Several procedures that optimize the use of the GPU’s resources are presented. We conclude that the implementation that produces better results (both from the performance and the genetic algorithm fitness perspectives) is simulating a few thousands of individuals grouped in a few islands using elitist selection. This model comprises 2 mighty factors for discovering the best solutions: finding good individuals in a short number of generations, and introducing genetic diversity via a relatively frequent and numerous migration. As a result, we have even found the optimal solution for the analyzed gene regulatory network (GRN). In addition, a comparative study of the performance obtained by the different parallel implementations on GPU versus a sequential application on CPU is carried out. In our tests, a multifold speedup was obtained for our optimized parallel implementation of the method on medium class GPU over an equivalent sequential single-core implementation running on a recent Intel i7 CPU. This work can provide useful guidance to researchers in biology, medicine, or bioinformatics in how to take advantage of the parallelization on massively parallel devices and GPUs to apply novel metaheuristic algorithms powered by nature for real-world applications (like the method to solve the temporal dynamics of GRNs). PMID:29662297
García-Calvo, Raúl; Guisado, J L; Diaz-Del-Rio, Fernando; Córdoba, Antonio; Jiménez-Morales, Francisco
2018-01-01
Understanding the regulation of gene expression is one of the key problems in current biology. A promising method for that purpose is the determination of the temporal dynamics between known initial and ending network states, by using simple acting rules. The huge amount of rule combinations and the nonlinear inherent nature of the problem make genetic algorithms an excellent candidate for finding optimal solutions. As this is a computationally intensive problem that needs long runtimes in conventional architectures for realistic network sizes, it is fundamental to accelerate this task. In this article, we study how to develop efficient parallel implementations of this method for the fine-grained parallel architecture of graphics processing units (GPUs) using the compute unified device architecture (CUDA) platform. An exhaustive and methodical study of various parallel genetic algorithm schemes-master-slave, island, cellular, and hybrid models, and various individual selection methods (roulette, elitist)-is carried out for this problem. Several procedures that optimize the use of the GPU's resources are presented. We conclude that the implementation that produces better results (both from the performance and the genetic algorithm fitness perspectives) is simulating a few thousands of individuals grouped in a few islands using elitist selection. This model comprises 2 mighty factors for discovering the best solutions: finding good individuals in a short number of generations, and introducing genetic diversity via a relatively frequent and numerous migration. As a result, we have even found the optimal solution for the analyzed gene regulatory network (GRN). In addition, a comparative study of the performance obtained by the different parallel implementations on GPU versus a sequential application on CPU is carried out. In our tests, a multifold speedup was obtained for our optimized parallel implementation of the method on medium class GPU over an equivalent sequential single-core implementation running on a recent Intel i7 CPU. This work can provide useful guidance to researchers in biology, medicine, or bioinformatics in how to take advantage of the parallelization on massively parallel devices and GPUs to apply novel metaheuristic algorithms powered by nature for real-world applications (like the method to solve the temporal dynamics of GRNs).
A graphical method is presented for displaying how binding proteins and other macromolecules interact with individual bases of nucleotide sequences. Characters representing the sequence are either oriented normally and placed above a line indicating favorable contact, or upside-down and placed below the line indicating unfavorable contact. The positive or negative height of
ERIC Educational Resources Information Center
Schwabe, Robert A.
Interest in Total Quality Management (TQM) at institutions of higher education has been stressed in recent years as an important area of activity for institutional researchers. Two previous AIR Forum papers have presented some of the statistical and graphical methods used for TQM. This paper, the third in the series, first discusses some of the…
2D-dynamic representation of DNA sequences as a graphical tool in bioinformatics
NASA Astrophysics Data System (ADS)
Bielińska-Wa̧Ż, D.; Wa̧Ż, P.
2016-10-01
2D-dynamic representation of DNA sequences is briefly reviewed. Some new examples of 2D-dynamic graphs which are the graphical tool of the method are shown. Using the examples of the complete genome sequences of the Zika virus it is shown that the present method can be applied for the study of the evolution of viral genomes.
A Java-Enabled Interactive Graphical Gas Turbine Propulsion System Simulator
NASA Technical Reports Server (NTRS)
Reed, John A.; Afjeh, Abdollah A.
1997-01-01
This paper describes a gas turbine simulation system which utilizes the newly developed Java language environment software system. The system provides an interactive graphical environment which allows the quick and efficient construction and analysis of arbitrary gas turbine propulsion systems. The simulation system couples a graphical user interface, developed using the Java Abstract Window Toolkit, and a transient, space- averaged, aero-thermodynamic gas turbine analysis method, both entirely coded in the Java language. The combined package provides analytical, graphical and data management tools which allow the user to construct and control engine simulations by manipulating graphical objects on the computer display screen. Distributed simulations, including parallel processing and distributed database access across the Internet and World-Wide Web (WWW), are made possible through services provided by the Java environment.
Ice-sheet modelling accelerated by graphics cards
NASA Astrophysics Data System (ADS)
Brædstrup, Christian Fredborg; Damsgaard, Anders; Egholm, David Lundbek
2014-11-01
Studies of glaciers and ice sheets have increased the demand for high performance numerical ice flow models over the past decades. When exploring the highly non-linear dynamics of fast flowing glaciers and ice streams, or when coupling multiple flow processes for ice, water, and sediment, researchers are often forced to use super-computing clusters. As an alternative to conventional high-performance computing hardware, the Graphical Processing Unit (GPU) is capable of massively parallel computing while retaining a compact design and low cost. In this study, we present a strategy for accelerating a higher-order ice flow model using a GPU. By applying the newest GPU hardware, we achieve up to 180× speedup compared to a similar but serial CPU implementation. Our results suggest that GPU acceleration is a competitive option for ice-flow modelling when compared to CPU-optimised algorithms parallelised by the OpenMP or Message Passing Interface (MPI) protocols.
The Evaluation of Two CDU Concepts and Their Effects on FMS Training
NASA Technical Reports Server (NTRS)
Abbott, Terence S.
1995-01-01
One of the biggest challenges for a pilot in the transition to a "glass" cockpit is understanding the Flight Management System (FMS). This is due to both the complex nature of the FMS and to the pilot-FMS interface. For these reasons, a large portion of transition training is devoted to the FMS. The intent of the current study was to examine the impact of the primary pilot-FMS interface, the Control Display Unit (CDU), on FMS training. The hypothesis of this study was that the interface design could have a significant impact on training. An FMS simulation was developed with two separate interfaces. One interface was similar to a current-generation design and the other was a multi-windows CDU based on graphical user interface techniques. For both application and evaluation reasons, constraints were applied to the graphical CDU design to maintain as much similarity as possible with the conventional CDU.
NASA Technical Reports Server (NTRS)
Hickey, J. S.
1983-01-01
The Mesoscale Analysis and Space Sensor (MASS) Data Management and Analysis System developed by Atsuko Computing International (ACI) on the MASS HP-1000 Computer System within the Systems Dynamics Laboratory of the Marshall Space Flight Center is described. The MASS Data Management and Analysis System was successfully implemented and utilized daily by atmospheric scientists to graphically display and analyze large volumes of conventional and satellite derived meteorological data. The scientists can process interactively various atmospheric data (Sounding, Single Level, Gird, and Image) by utilizing the MASS (AVE80) share common data and user inputs, thereby reducing overhead, optimizing execution time, and thus enhancing user flexibility, useability, and understandability of the total system/software capabilities. In addition ACI installed eight APPLE III graphics/imaging computer terminals in individual scientist offices and integrated them into the MASS HP-1000 Computer System thus providing significant enhancement to the overall research environment.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnstad, H.
The purpose of this meeting is to discuss the current and future HEP computing support and environments from the perspective of new horizons in accelerator, physics, and computing technologies. Topics of interest to the Meeting include (but are limited to): the forming of the HEPLIB world user group for High Energy Physic computing; mandate, desirables, coordination, organization, funding; user experience, international collaboration; the roles of national labs, universities, and industry; range of software, Monte Carlo, mathematics, physics, interactive analysis, text processors, editors, graphics, data base systems, code management tools; program libraries, frequency of updates, distribution; distributed and interactive computing, datamore » base systems, user interface, UNIX operating systems, networking, compilers, Xlib, X-Graphics; documentation, updates, availability, distribution; code management in large collaborations, keeping track of program versions; and quality assurance, testing, conventions, standards.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnstad, H.
The purpose of this meeting is to discuss the current and future HEP computing support and environments from the perspective of new horizons in accelerator, physics, and computing technologies. Topics of interest to the Meeting include (but are limited to): the forming of the HEPLIB world user group for High Energy Physic computing; mandate, desirables, coordination, organization, funding; user experience, international collaboration; the roles of national labs, universities, and industry; range of software, Monte Carlo, mathematics, physics, interactive analysis, text processors, editors, graphics, data base systems, code management tools; program libraries, frequency of updates, distribution; distributed and interactive computing, datamore » base systems, user interface, UNIX operating systems, networking, compilers, Xlib, X-Graphics; documentation, updates, availability, distribution; code management in large collaborations, keeping track of program versions; and quality assurance, testing, conventions, standards.« less
Implications of Graphic Cigarette Warning Labels on Smoking Behavior: An International Perspective.
Jung, Minsoo
2016-03-01
Graphic warning labels (GWLs) have been developed as a representative non-price policy to block such marketing. This study investigated the current state and effect of the global introduction of GWLs and examines the future tasks related to GWLs. We systematically reviewed literatures on GWL and a tobacco control strategy in the past fifteen years. The policy of enforcing GWLs has spread globally based on the Framework Convention on Tobacco Control. GWLs are more effective than text warnings and are implemented in over 70 countries. The policy has showed the impact of GWLs as a preventive effect on adolescents' smoking, inducement of smoking cessation, reduction in the amount of tobacco smoked, and reduction in smoking rates. The success of an anti-smoking policy can manifests itself as an effect of individual policies, the rise of tobacco prices, and the introduction of GWLs.
Hasegawa, Tomoyuki; Kojima, Haruna; Masu, Chisato; Fukushima, Yasuhiro; Kojima, Hironori; Konokawa, Kiminori; Isobe, Tomonori; Sato, Eisuke; Murayama, Hideo; Maruyama, Koichi; Umeda, Tokuo
2010-01-01
Physics-related subjects are important in the educational fields of radiological physics and technology. However, conventional teaching tools, for example texts, equations, and two-dimensional figures, are not very effective in attracting the interest of students. Therefore, we have created several multimedia educational materials covering radiological physics and technology. Each educational presentation includes several segments of high-quality computer-graphic animations designed to attract students' interest. We used personal computers (PCs) and commercial software to create and compile these. Undergraduate and graduate students and teachers and related professionals contributed to the design and creation of the educational materials as part of student research. The educational materials can be displayed on a PC monitor and manipulated with popular free software. Opinion surveys conducted in undergraduate courses at Kitasato University support the effectiveness of our educational tools in helping students gain a better understanding of the subjects offered and in raising their interest.
Use of natural user interfaces in water simulations
NASA Astrophysics Data System (ADS)
Donchyts, G.; Baart, F.; van Dam, A.; Jagers, B.
2013-12-01
Conventional graphical user interfaces, used to edit input and present results of earth science models, have seen little innovation for the past two decades. In most cases model data is presented and edited using 2D projections even when working with 3D data. The emergence of 3D motion sensing technologies, such as Microsoft Kinect and LEAP Motion, opens new possibilities for user interaction by adding more degrees of freedom compared to a classical way using mouse and keyboard. Here we investigate how interaction with hydrodynamic numerical models can be improved using these new technologies. Our research hypothesis (H1) states that properly designed 3D graphical user interface paired with the 3D motion sensor can significantly reduce the time required to setup and use numerical models. In this work we have used a LEAP motion controller combined with a shallow water flow model engine D-Flow Flexible Mesh. Interacting with numerical model using hands
Sun, F; Chen, J; Tong, Q; Zeng, S
2007-01-01
Management of drinking water safety is changing towards an integrated risk assessment and risk management approach that includes all processes in a water supply system from catchment to consumers. However, given the large number of water supply systems in China and the cost of implementing such a risk assessment procedure, there is a necessity to first conduct a strategic screening analysis at a national level. An integrated methodology of risk assessment and screening analysis is thus proposed to evaluate drinking water safety of a conventional water supply system. The violation probability, indicating drinking water safety, is estimated at different locations of a water supply system in terms of permanganate index, ammonia nitrogen, turbidity, residual chlorine and trihalomethanes. Critical parameters with respect to drinking water safety are then identified, based on which an index system is developed to prioritize conventional water supply systems in implementing a detailed risk assessment procedure. The evaluation results are represented as graphic check matrices for the concerned hazards in drinking water, from which the vulnerability of a conventional water supply system is characterized.
Graphics Flutter Analysis Methods, an interactive computing system at Lockheed-California Company
NASA Technical Reports Server (NTRS)
Radovcich, N. A.
1975-01-01
An interactive computer graphics system, Graphics Flutter Analysis Methods (GFAM), was developed to complement FAMAS, a matrix-oriented batch computing system, and other computer programs in performing complex numerical calculations using a fully integrated data management system. GFAM has many of the matrix operation capabilities found in FAMAS, but on a smaller scale, and is utilized when the analysis requires a high degree of interaction between the engineer and computer, and schedule constraints exclude the use of batch entry programs. Applications of GFAM to a variety of preliminary design, development design, and project modification programs suggest that interactive flutter analysis using matrix representations is a feasible and cost effective computing tool.
3D Modeling as Method for Construction and Analysis of Graphic Objects
NASA Astrophysics Data System (ADS)
Kheyfets, A. L.; Vasilieva, V. N.
2017-11-01
The use of 3D modeling when constructing and analyzing perspective projections and shadows is considered. The creation of photorealistic image is shown. The perspective of the construction project and characterization of its image are given as an example. The authors consider the construction of a dynamic block as a means of graphical information storage and automation of geometric constructions. The example of the dynamic block construction at creating a truss node is demonstrated. The constructions are considered as applied to the Auto-CAD software. The paper is aimed at improving the graphic methods of architectural design and improving the educational process when training the Bachelor’s degree students majoring in construction.
Zhang, Xiao-Fei; Ou-Yang, Le; Yan, Hong
2017-08-15
Understanding how gene regulatory networks change under different cellular states is important for revealing insights into network dynamics. Gaussian graphical models, which assume that the data follow a joint normal distribution, have been used recently to infer differential networks. However, the distributions of the omics data are non-normal in general. Furthermore, although much biological knowledge (or prior information) has been accumulated, most existing methods ignore the valuable prior information. Therefore, new statistical methods are needed to relax the normality assumption and make full use of prior information. We propose a new differential network analysis method to address the above challenges. Instead of using Gaussian graphical models, we employ a non-paranormal graphical model that can relax the normality assumption. We develop a principled model to take into account the following prior information: (i) a differential edge less likely exists between two genes that do not participate together in the same pathway; (ii) changes in the networks are driven by certain regulator genes that are perturbed across different cellular states and (iii) the differential networks estimated from multi-view gene expression data likely share common structures. Simulation studies demonstrate that our method outperforms other graphical model-based algorithms. We apply our method to identify the differential networks between platinum-sensitive and platinum-resistant ovarian tumors, and the differential networks between the proneural and mesenchymal subtypes of glioblastoma. Hub nodes in the estimated differential networks rediscover known cancer-related regulator genes and contain interesting predictions. The source code is at https://github.com/Zhangxf-ccnu/pDNA. szuouyl@gmail.com. Supplementary data are available at Bioinformatics online. © The Author (2017). Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com
Vexler, Albert; Yu, Jihnhee
2018-04-13
A common statistical doctrine supported by many introductory courses and textbooks is that t-test type procedures based on normally distributed data points are anticipated to provide a standard in decision-making. In order to motivate scholars to examine this convention, we introduce a simple approach based on graphical tools of receiver operating characteristic (ROC) curve analysis, a well-established biostatistical methodology. In this context, we propose employing a p-values-based method, taking into account the stochastic nature of p-values. We focus on the modern statistical literature to address the expected p-value (EPV) as a measure of the performance of decision-making rules. During the course of our study, we extend the EPV concept to be considered in terms of the ROC curve technique. This provides expressive evaluations and visualizations of a wide spectrum of testing mechanisms' properties. We show that the conventional power characterization of tests is a partial aspect of the presented EPV/ROC technique. We desire that this explanation of the EPV/ROC approach convinces researchers of the usefulness of the EPV/ROC approach for depicting different characteristics of decision-making procedures, in light of the growing interest regarding correct p-values-based applications.
NASA Astrophysics Data System (ADS)
Saini, Surender Singh; Sardana, Harish Kumar; Pattnaik, Shyam Sundar
2017-06-01
Conventional image editing software in combination with other techniques are not only difficult to apply to an image but also permits a user to perform some basic functions one at a time. However, image processing algorithms and photogrammetric systems are developed in the recent past for real-time pattern recognition applications. A graphical user interface (GUI) is developed which can perform multiple functions simultaneously for the analysis and estimation of geometric distortion in an image with reference to the corresponding distorted image. The GUI measure, record, and visualize the performance metric of X/Y coordinates of one image over the other. The various keys and icons provided in the utility extracts the coordinates of distortion free reference image and the image with geometric distortion. The error between these two corresponding points gives the measure of distortion and also used to evaluate the correction parameters for image distortion. As the GUI interface minimizes human interference in the process of geometric correction, its execution just requires use of icons and keys provided in the utility; this technique gives swift and accurate results as compared to other conventional methods for the measurement of the X/Y coordinates of an image.
Peng, Bo; Wang, Yuqi; Hall, Timothy J; Jiang, Jingfeng
2017-01-01
Our primary objective of this work was to extend a previously published 2D coupled sub-sample tracking algorithm for 3D speckle tracking in the framework of ultrasound breast strain elastography. In order to overcome heavy computational cost, we investigated the use of a graphic processing unit (GPU) to accelerate the 3D coupled sub-sample speckle tracking method. The performance of the proposed GPU implementation was tested using a tissue-mimicking (TM) phantom and in vivo breast ultrasound data. The performance of this 3D sub-sample tracking algorithm was compared with the conventional 3D quadratic sub-sample estimation algorithm. On the basis of these evaluations, we concluded that the GPU implementation of this 3D sub-sample estimation algorithm can provide high-quality strain data (i.e. high correlation between the pre- and the motion-compensated post-deformation RF echo data and high contrast-to-noise ratio strain images), as compared to the conventional 3D quadratic sub-sample algorithm. Using the GPU implementation of the 3D speckle tracking algorithm, volumetric strain data can be achieved relatively fast (approximately 20 seconds per volume [2.5 cm × 2.5 cm × 2.5 cm]). PMID:28166493
NASA Astrophysics Data System (ADS)
Popa, CL; Popa, V.
2016-11-01
This paper proposes a profiling method for the tool which generates the helical groove of male rotor, screw compressor component. The method is based on a complementary theorem of surfaces enveloping - "Substitute Family Circles Method”. The specific theorem of family circles of substitution has been applied using AUTOCAD graphics design environment facility. The frontal view of the male rotor, screw compressor component, has been determinate knowing the transverse profile of female rotor, and using this theorem of "Substitute Family Circle". The three-dimensional model of the rotor makes possible to apply the same theorem, leading to the surface of revolution enveloping the helical surface. An application will be also presented to determine the axial profile of the disk cutter, numeric and graphics, following the proposed algorithm.
Nuutinen, Mikko; Virtanen, Toni; Rummukainen, Olli; Häkkinen, Jukka
2016-03-01
This article presents VQone, a graphical experiment builder, written as a MATLAB toolbox, developed for image and video quality ratings. VQone contains the main elements needed for the subjective image and video quality rating process. This includes building and conducting experiments and data analysis. All functions can be controlled through graphical user interfaces. The experiment builder includes many standardized image and video quality rating methods. Moreover, it enables the creation of new methods or modified versions from standard methods. VQone is distributed free of charge under the terms of the GNU general public license and allows code modifications to be made so that the program's functions can be adjusted according to a user's requirements. VQone is available for download from the project page (http://www.helsinki.fi/psychology/groups/visualcognition/).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wahanani, Nursinta Adi, E-mail: sintaadi@batan.go.id; Natsir, Khairina, E-mail: sintaadi@batan.go.id; Hartini, Entin, E-mail: sintaadi@batan.go.id
Data processing software packages such as VSOP and MCNPX are softwares that has been scientifically proven and complete. The result of VSOP and MCNPX are huge and complex text files. In the analyze process, user need additional processing like Microsoft Excel to show informative result. This research develop an user interface software for output of VSOP and MCNPX. VSOP program output is used to support neutronic analysis and MCNPX program output is used to support burn-up analysis. Software development using iterative development methods which allow for revision and addition of features according to user needs. Processing time with this softwaremore » 500 times faster than with conventional methods using Microsoft Excel. PYTHON is used as a programming language, because Python is available for all major operating systems: Windows, Linux/Unix, OS/2, Mac, Amiga, among others. Values that support neutronic analysis are k-eff, burn-up and mass Pu{sup 239} and Pu{sup 241}. Burn-up analysis used the mass inventory values of actinide (Thorium, Plutonium, Neptunium and Uranium). Values are visualized in graphical shape to support analysis.« less
USGS GeoData Digital Raster Graphics
,
2001-01-01
Passive diffusion samplers have been tested at a number of sites where volatile organic compounds (VOC?s) are the principal contaminants in ground water. Test results generally show good agreement between concentrations of VOC?s in samples collected with diffusion samplers and concentrations in samples collected by purging the water from a well. Diffusion samplers offer several advantages over conventional and low-flow ground-water sampling procedures: ? Elimination of the need to purge a well before collecting a sample and to dispose of contaminated water. ? Elimination of cross-contamination of samples associated with sampling with non-dedicated pumps or sample delivery tubes. ? Reduction in sampling time by as much as 80 percent of that required for ?purge type? sampling methods. ? An increase in the frequency and spatial coverage of monitoring at a site because of the associated savings in time and money. The successful use of diffusion samplers depends on the following three primary factors: (1) understanding site conditions and contaminants of interest (defining sample objectives), (2) validating of results of diffusion samplers against more widely acknowledged sampling methods, and (3) applying diffusion samplers in the field.
Pattern-histogram-based temporal change detection using personal chest radiographs
NASA Astrophysics Data System (ADS)
Ugurlu, Yucel; Obi, Takashi; Hasegawa, Akira; Yamaguchi, Masahiro; Ohyama, Nagaaki
1999-05-01
An accurate and reliable detection of temporal changes from a pair of images has considerable interest in the medical science. Traditional registration and subtraction techniques can be applied to extract temporal differences when,the object is rigid or corresponding points are obvious. However, in radiological imaging, loss of the depth information, the elasticity of object, the absence of clearly defined landmarks and three-dimensional positioning differences constraint the performance of conventional registration techniques. In this paper, we propose a new method in order to detect interval changes accurately without using an image registration technique. The method is based on construction of so-called pattern histogram and comparison procedure. The pattern histogram is a graphic representation of the frequency counts of all allowable patterns in the multi-dimensional pattern vector space. K-means algorithm is employed to partition pattern vector space successively. Any differences in the pattern histograms imply that different patterns are involved in the scenes. In our experiment, a pair of chest radiographs of pneumoconiosis is employed and the changing histogram bins are visualized on both of the images. We found that the method can be used as an alternative way of temporal change detection, particularly when the precise image registration is not available.
Enhanced calculation of eigen-stress field and elastic energy in atomistic interdiffusion of alloys
NASA Astrophysics Data System (ADS)
Cecilia, José M.; Hernández-Díaz, A. M.; Castrillo, Pedro; Jiménez-Alonso, J. F.
2017-02-01
The structural evolution of alloys is affected by the elastic energy associated to eigen-stress fields. However, efficient calculations of the elastic energy in evolving geometries are actually a great challenge in promising atomistic simulation techniques such as Kinetic Monte Carlo (KMC) methods. In this paper, we report two complementary algorithms to calculate the eigen-stress field by linear superposition (a.k.a. LSA, Lineal Superposition Algorithm) and the elastic energy modification in atomistic interdiffusion of alloys (the Atom Exchange Elastic Energy Evaluation (AE4) Algorithm). LSA is shown to be appropriated for fast incremental stress calculation in highly nanostructured materials, whereas AE4 provides the required input for KMC and, additionally, it can be used to evaluate the accuracy of the eigen-stress field calculated by LSA. Consequently, they are suitable to be used on-the-fly with KMC. Both algorithms are massively parallel by their definition and thus well-suited for their parallelization on modern Graphics Processing Units (GPUs). Our computational studies confirm that we can obtain significant improvements compared to conventional Finite Element Methods, and the utilization of GPUs opens up new possibilities for the development of these methods in atomistic simulation of materials.
An extended Kalman filter for mouse tracking.
Choi, Hongjun; Kim, Mingi; Lee, Onseok
2018-05-19
Animal tracking is an important tool for observing behavior, which is useful in various research areas. Animal specimens can be tracked using dynamic models and observation models that require several types of data. Tracking mouse has several barriers due to the physical characteristics of the mouse, their unpredictable movement, and cluttered environments. Therefore, we propose a reliable method that uses a detection stage and a tracking stage to successfully track mouse. The detection stage detects the surface area of the mouse skin, and the tracking stage implements an extended Kalman filter to estimate the state variables of a nonlinear model. The changes in the overall shape of the mouse are tracked using an oval-shaped tracking model to estimate the parameters for the ellipse. An experiment is conducted to demonstrate the performance of the proposed tracking algorithm using six video images showing various types of movement, and the ground truth values for synthetic images are compared to the values generated by the tracking algorithm. A conventional manual tracking method is also applied to compare across eight experimenters. Furthermore, the effectiveness of the proposed tracking method is also demonstrated by applying the tracking algorithm with actual images of mouse. Graphical abstract.
Hotspot detection using image pattern recognition based on higher-order local auto-correlation
NASA Astrophysics Data System (ADS)
Maeda, Shimon; Matsunawa, Tetsuaki; Ogawa, Ryuji; Ichikawa, Hirotaka; Takahata, Kazuhiro; Miyairi, Masahiro; Kotani, Toshiya; Nojima, Shigeki; Tanaka, Satoshi; Nakagawa, Kei; Saito, Tamaki; Mimotogi, Shoji; Inoue, Soichi; Nosato, Hirokazu; Sakanashi, Hidenori; Kobayashi, Takumi; Murakawa, Masahiro; Higuchi, Tetsuya; Takahashi, Eiichi; Otsu, Nobuyuki
2011-04-01
Below 40nm design node, systematic variation due to lithography must be taken into consideration during the early stage of design. So far, litho-aware design using lithography simulation models has been widely applied to assure that designs are printed on silicon without any error. However, the lithography simulation approach is very time consuming, and under time-to-market pressure, repetitive redesign by this approach may result in the missing of the market window. This paper proposes a fast hotspot detection support method by flexible and intelligent vision system image pattern recognition based on Higher-Order Local Autocorrelation. Our method learns the geometrical properties of the given design data without any defects as normal patterns, and automatically detects the design patterns with hotspots from the test data as abnormal patterns. The Higher-Order Local Autocorrelation method can extract features from the graphic image of design pattern, and computational cost of the extraction is constant regardless of the number of design pattern polygons. This approach can reduce turnaround time (TAT) dramatically only on 1CPU, compared with the conventional simulation-based approach, and by distributed processing, this has proven to deliver linear scalability with each additional CPU.
Communication Design Education: Could Nine Reflections Be Sufficient?
ERIC Educational Resources Information Center
van der Waarde, Karel; Vroombout, Maurits
2012-01-01
Situation: Graphic design education is subject to substantial changes. Changes in professional practice and higher education aggravate insecurities about the contents and structure of courses, assessment criteria, relations between practice, research and theory and teaching methods. Assumption: Graphic design education (visual communication design…
A Graphic Anthropometric Aid for Seating and Workplace Design.
1984-04-01
required proportion of the pdf . Suppose that some attribute is distributed according to a bivariate Normal pdf of zero mean value and equal variances a...2 Note that circular contours. dran at the normaliwed radii presented above, will enclose the respective proportions of the bi artate Normal pdf ...INTRODUCTION 1 2. A TWO-DIMENSIONAL MODEL BASE 2 3. CONCEPT OF USE 4 4. VALIDATION OF THE TWO-DIMENSIONAL MODEL 8 4.1 Conventional Anthropometry 9 4.2
Eksborg, Staffan
2013-01-01
Pharmacokinetic studies are important for optimizing of drug dosing, but requires proper validation of the used pharmacokinetic procedures. However, simple and reliable statistical methods suitable for evaluation of the predictive performance of pharmacokinetic analysis are essentially lacking. The aim of the present study was to construct and evaluate a graphic procedure for quantification of predictive performance of individual and population pharmacokinetic compartment analysis. Original data from previously published pharmacokinetic compartment analyses after intravenous, oral, and epidural administration, and digitized data, obtained from published scatter plots of observed vs predicted drug concentrations from population pharmacokinetic studies using the NPEM algorithm and NONMEM computer program and Bayesian forecasting procedures, were used for estimating the predictive performance according to the proposed graphical method and by the method of Sheiner and Beal. The graphical plot proposed in the present paper proved to be a useful tool for evaluation of predictive performance of both individual and population compartment pharmacokinetic analysis. The proposed method is simple to use and gives valuable information concerning time- and concentration-dependent inaccuracies that might occur in individual and population pharmacokinetic compartment analysis. Predictive performance can be quantified by the fraction of concentration ratios within arbitrarily specified ranges, e.g. within the range 0.8-1.2.
Mono-stereo-autostereo: the evolution of 3-dimensional neurosurgical planning.
Stadie, Axel T; Kockro, Ralf A
2013-01-01
In the past decade, surgery planning has changed significantly. The main reason is the improvements in computer graphical rendering power and display technology, which turned the plain graphics of the mid-1990s into interactive stereoscopic objects. To report our experiences with 2 virtual reality systems used for planning neurosurgical operations. A series of 208 operations were planned with the Dextroscope (Bracco AMT, Singapore) requiring the use of liquid crystal display shutter glasses. The participating neurosurgeons answered a questionnaire after the planning procedure and postoperatively. In a second prospective series of 33 patients, we used an autostereoscopic monitor system (MD20-3-D; Setred SA, Sweden) to plan intracranial operations. A questionnaire regarding the value of surgery planning was answered preoperatively and postoperatively. The Dextroscope could be integrated into daily surgical routine. Surgeons regarded their understanding of the pathoanatomical situation as improved, leading to enhanced intraoperative orientation and confidence compared with conventional planning. The autostereoscopic Setred system was regarded as helpful in establishing the surgical strategy and analyzing the pathoanatomical situation compared with conventional planning. Both systems were perceived as a backup in case of failure of the standard navigation system. Improvement of display and interaction techniques adds to the realism of the planning process and enables precise structural understanding preoperatively. This minimizes intraoperative guesswork and exploratory dissection. Autostereoscopic display techniques will further increase the value and acceptance of 3-dimensional planning and intraoperative navigation.
Data Analysis with Graphical Models: Software Tools
NASA Technical Reports Server (NTRS)
Buntine, Wray L.
1994-01-01
Probabilistic graphical models (directed and undirected Markov fields, and combined in chain graphs) are used widely in expert systems, image processing and other areas as a framework for representing and reasoning with probabilities. They come with corresponding algorithms for performing probabilistic inference. This paper discusses an extension to these models by Spiegelhalter and Gilks, plates, used to graphically model the notion of a sample. This offers a graphical specification language for representing data analysis problems. When combined with general methods for statistical inference, this also offers a unifying framework for prototyping and/or generating data analysis algorithms from graphical specifications. This paper outlines the framework and then presents some basic tools for the task: a graphical version of the Pitman-Koopman Theorem for the exponential family, problem decomposition, and the calculation of exact Bayes factors. Other tools already developed, such as automatic differentiation, Gibbs sampling, and use of the EM algorithm, make this a broad basis for the generation of data analysis software.
Choi, Yeop; Eo, SuRak
2014-04-01
The vertical dimension of the palpebral fissure and the marginal reflex distance are conventionally used to assess the amount or degree of blepharoptosis, and levator function is assessed by measuring total upper lid excursion between the extremes of down-gaze and up-gaze. However, these are 1-dimensional measures obtained with a ruler, and the results obtained are dependent on examiner skill. Digital photographs were obtained of 692 patients before and after upper blepharoplasty. Visual iris-pupil complex percentage (VIP) was measured in the 1,305 eyes by digital calculation using Adobe Photoshop CS3 (Adobe Systems, Inc). Perioperative eye images in primary gaze were evaluated independently by 2 surgeons, 2 nurses, and a graphic designer, and after excluding 50 eyes which were nonconcordant and 29 eyes which revealed retracted upper lids, the remaining 1,305 eyes were classified into 4 major groups, that is, into excellent (n = 415), good (n = 435), subclinical (n = 270), and prominent ptosis (n = 185) groups. In addition, eyes were subdivided into 5 types according to the iris-pupil complex position within the palpebral fissure. Visual iris-pupil complex percentages were from 85% to 94% in the excellent, from 78% to 84% in the good, from 70% to 77% in the subclinical ptosis, and below 70% in the prominent ptosis group. Patients in the subclinical or prominent ptosis eye group required surgery for blepharoptosis. Iris-pupil complex relation to the palpebral opening was classified into 5 eye types, namely, standard (n = 961), scleral (n = 266), sinking (n = 151), retracted (n = 3), and fish (n = 3). The authors devised a new prospective measurement method for assessing blepharoptosis in a clinical setting. Graphical comparisons between the devised method of measuring VIP and mathematical estimations showed that the devised method is easier, more practical, and more precise for measuring degree of blepharoptosis from general population trends, and that VIP also provides a useful objective index for evaluating the postoperative results of blepharoptosis.
Graphic Warning Labels in Cigarette Advertisements: Recall and Viewing Patterns
Strasser, Andrew A.; Tang, Kathy Z.; Romer, Daniel; Jepson, Chris; Cappella, Joseph N.
2012-01-01
Background The Family Smoking Prevention and Control Act gave the U.S. Food and Drug Administration (FDA) legal authority to mandate graphic warning labels on cigarette advertising and packaging. The FDA requires that these graphic warning labels be embedded into cigarette advertising and packaging by September 2012. Purpose The aim of this study was to examine differences in recall and viewing patterns of text-only versus graphic cigarette warning labels; and, the association between viewing patterns and recall. Methods Participants (current daily smokers; N=200) were randomized to view a cigarette advertisement with either text-only or graphic warning labels. Viewing patterns were measured using eye-tracking, and recall was later assessed. Sessions were conducted between November 2008 and November 2009. Data analysis was conducted between March 2011 and July 2011. Results There was a significant difference in percentage correct recall of the warning label between those in the text-only versus graphic warning label condition, 50% versus 83% (χ2 =23.74, p=0.0001). Time to first view of the graphic warning label text, and dwell time duration (i.e., time spent looking) on the graphic image were significantly associated with correct recall. Warning labels that drew attention more quickly and resulted in longer dwell times were associated with better recall. Conclusions Graphic warning labels improve smokers’ recall of warning and health risks; they do so by drawing and holding attention. PMID:22704744
User's Guide for Flight Simulation Data Visualization Workstation
NASA Technical Reports Server (NTRS)
Kaplan, Joseph A.; Chen, Ronnie; Kenney, Patrick S.; Koval, Christopher M.; Hutchinson, Brian K.
1996-01-01
Today's modern flight simulation research produces vast amounts of time sensitive data. The meaning of this data can be difficult to assess while in its raw format. Therefore, a method of breaking the data down and presenting it to the user in a graphical format is necessary. Simulation Graphics (SimGraph) is intended as a data visualization software package that will incorporate simulation data into a variety of animated graphical displays for easy interpretation by the simulation researcher. This document is intended as an end user's guide.
The Gaussian Graphical Model in Cross-Sectional and Time-Series Data.
Epskamp, Sacha; Waldorp, Lourens J; Mõttus, René; Borsboom, Denny
2018-04-16
We discuss the Gaussian graphical model (GGM; an undirected network of partial correlation coefficients) and detail its utility as an exploratory data analysis tool. The GGM shows which variables predict one-another, allows for sparse modeling of covariance structures, and may highlight potential causal relationships between observed variables. We describe the utility in three kinds of psychological data sets: data sets in which consecutive cases are assumed independent (e.g., cross-sectional data), temporally ordered data sets (e.g., n = 1 time series), and a mixture of the 2 (e.g., n > 1 time series). In time-series analysis, the GGM can be used to model the residual structure of a vector-autoregression analysis (VAR), also termed graphical VAR. Two network models can then be obtained: a temporal network and a contemporaneous network. When analyzing data from multiple subjects, a GGM can also be formed on the covariance structure of stationary means-the between-subjects network. We discuss the interpretation of these models and propose estimation methods to obtain these networks, which we implement in the R packages graphicalVAR and mlVAR. The methods are showcased in two empirical examples, and simulation studies on these methods are included in the supplementary materials.
Stator and Rotor Flux Based Deadbeat Direct Torque Control of Induction Machines
NASA Technical Reports Server (NTRS)
Kenny, Barbara H.; Lorenz, Robert D.
2001-01-01
A new, deadbeat type of direct torque control is proposed, analyzed, and experimentally verified in this paper. The control is based on stator and rotor flux as state variables. This choice of state variables allows a graphical representation which is transparent and insightful. The graphical solution shows the effects of realistic considerations such as voltage and current limits. A position and speed sensorless implementation of the control, based on the self-sensing signal injection technique, is also demonstrated experimentally for low speed operation. The paper first develops the new, deadbeat DTC methodology and graphical representation of the new algorithm. It then evaluates feasibility via simulation and experimentally demonstrates performance of the new method with a laboratory prototype including the sensorless methods.
Stator and Rotor Flux Based Deadbeat Direct Torque Control of Induction Machines
NASA Technical Reports Server (NTRS)
Kenny, Barbara H.; Lorenz, Robert D.
2003-01-01
A new, deadbeat type of direct torque control is proposed, analyzed and experimentally verified in this paper. The control is based on stator and rotor flux as state variables. This choice of state variables allows a graphical representation which is transparent and insightful. The graphical solution shows the effects of realistic considerations such as voltage and current limits. A position and speed sensorless implementation of the control, based on the self-sensing signal injection technique, is also demonstrated experimentally for low speed operation. The paper first develops the new, deadbeat DTC methodology and graphical representation of the new algorithm. It then evaluates feasibility via simulation and experimentally demonstrates performance of the new method with a laboratory prototype including the sensorless methods.
Stator and Rotor Flux Based Deadbeat Direct Torque Control of Induction Machines. Revision 1
NASA Technical Reports Server (NTRS)
Kenny, Barbara H.; Lorenz, Robert D.
2002-01-01
A new, deadbeat type of direct torque control is proposed, analyzed, and experimentally verified in this paper. The control is based on stator and rotor flux as state variables. This choice of state variables allows a graphical representation which is transparent and insightful. The graphical solution shows the effects of realistic considerations such as voltage and current limits. A position and speed sensorless implementation of the control, based on the self-sensing signal injection technique, is also demonstrated experimentally for low speed operation. The paper first develops the new, deadbeat DTC methodology and graphical representation of the new algorithm. It then evaluates feasibility via simulation and experimentally demonstrates performance of the new method with a laboratory prototype including the sensorless methods.
fgui: A Method for Automatically Creating Graphical User Interfaces for Command-Line R Packages
Hoffmann, Thomas J.; Laird, Nan M.
2009-01-01
The fgui R package is designed for developers of R packages, to help rapidly, and sometimes fully automatically, create a graphical user interface for a command line R package. The interface is built upon the Tcl/Tk graphical interface included in R. The package further facilitates the developer by loading in the help files from the command line functions to provide context sensitive help to the user with no additional effort from the developer. Passing a function as the argument to the routines in the fgui package creates a graphical interface for the function, and further options are available to tweak this interface for those who want more flexibility. PMID:21625291
Physiological Feedback Method and System
NASA Technical Reports Server (NTRS)
Pope, Alan T. (Inventor); Severance, Kurt E. (Inventor)
2002-01-01
A method and system provide physiological feedback for a patient and/or physician. At least one physiological effect experienced by a body part of a patient is measured noninvasively. A three-dimensional graphics model serving as an analogous representation of the body part is altered in accordance with the measurements. A binocular image signal representative of the three-dimensional graphics model so-altered is displayed for the patient and/or physician in a virtual reality environment.
IMP: Interactive mass properties program. Volume 1: Program description
NASA Technical Reports Server (NTRS)
Stewart, W. A.
1976-01-01
A method of computing a weights and center of gravity analysis of a flight vehicle using interactive graphical capabilities of the Adage 340 computer is described. The equations used to calculate area, volume, and mass properties are based on elemental surface characteristics. The input/output methods employ the graphic support of the Adage computer. Several interactive program options are available for analyzing the mass properties of a vehicle. These options are explained.
Viallon, Vivian; Banerjee, Onureena; Jougla, Eric; Rey, Grégoire; Coste, Joel
2014-03-01
Looking for associations among multiple variables is a topical issue in statistics due to the increasing amount of data encountered in biology, medicine, and many other domains involving statistical applications. Graphical models have recently gained popularity for this purpose in the statistical literature. In the binary case, however, exact inference is generally very slow or even intractable because of the form of the so-called log-partition function. In this paper, we review various approximate methods for structure selection in binary graphical models that have recently been proposed in the literature and compare them through an extensive simulation study. We also propose a modification of one existing method, that is shown to achieve good performance and to be generally very fast. We conclude with an application in which we search for associations among causes of death recorded on French death certificates. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
On an additive partial correlation operator and nonparametric estimation of graphical models.
Lee, Kuang-Yao; Li, Bing; Zhao, Hongyu
2016-09-01
We introduce an additive partial correlation operator as an extension of partial correlation to the nonlinear setting, and use it to develop a new estimator for nonparametric graphical models. Our graphical models are based on additive conditional independence, a statistical relation that captures the spirit of conditional independence without having to resort to high-dimensional kernels for its estimation. The additive partial correlation operator completely characterizes additive conditional independence, and has the additional advantage of putting marginal variation on appropriate scales when evaluating interdependence, which leads to more accurate statistical inference. We establish the consistency of the proposed estimator. Through simulation experiments and analysis of the DREAM4 Challenge dataset, we demonstrate that our method performs better than existing methods in cases where the Gaussian or copula Gaussian assumption does not hold, and that a more appropriate scaling for our method further enhances its performance.
On an additive partial correlation operator and nonparametric estimation of graphical models
Li, Bing; Zhao, Hongyu
2016-01-01
Abstract We introduce an additive partial correlation operator as an extension of partial correlation to the nonlinear setting, and use it to develop a new estimator for nonparametric graphical models. Our graphical models are based on additive conditional independence, a statistical relation that captures the spirit of conditional independence without having to resort to high-dimensional kernels for its estimation. The additive partial correlation operator completely characterizes additive conditional independence, and has the additional advantage of putting marginal variation on appropriate scales when evaluating interdependence, which leads to more accurate statistical inference. We establish the consistency of the proposed estimator. Through simulation experiments and analysis of the DREAM4 Challenge dataset, we demonstrate that our method performs better than existing methods in cases where the Gaussian or copula Gaussian assumption does not hold, and that a more appropriate scaling for our method further enhances its performance. PMID:29422689
Evaluation of 3-D graphics software: A case study
NASA Technical Reports Server (NTRS)
Lores, M. E.; Chasen, S. H.; Garner, J. M.
1984-01-01
An efficient 3-D geometry graphics software package which is suitable for advanced design studies was developed. The advanced design system is called GRADE--Graphics for Advanced Design. Efficiency and ease of use are gained by sacrificing flexibility in surface representation. The immediate options were either to continue development of GRADE or to acquire a commercially available system which would replace or complement GRADE. Test cases which would reveal the ability of each system to satisfy the requirements were developed. A scoring method which adequately captured the relative capabilities of the three systems was presented. While more complex multi-attribute decision methods could be used, the selected method provides all the needed information without being so complex that it is difficult to understand. If the value factors are modestly perturbed, system Z is a clear winner based on its overall capabilities. System Z is superior in two vital areas: surfacing and ease of interface with application programs.
Graphical tensor product reduction scheme for the Lie algebras so(5) = sp(2) , su(3) , and g(2)
NASA Astrophysics Data System (ADS)
Vlasii, N. D.; von Rütte, F.; Wiese, U.-J.
2016-08-01
We develop in detail a graphical tensor product reduction scheme, first described by Antoine and Speiser, for the simple rank 2 Lie algebras so(5) = sp(2) , su(3) , and g(2) . This leads to an efficient practical method to reduce tensor products of irreducible representations into sums of such representations. For this purpose, the 2-dimensional weight diagram of a given representation is placed in a ;landscape; of irreducible representations. We provide both the landscapes and the weight diagrams for a large number of representations for the three simple rank 2 Lie algebras. We also apply the algebraic ;girdle; method, which is much less efficient for calculations by hand for moderately large representations. Computer code for reducing tensor products, based on the graphical method, has been developed as well and is available from the authors upon request.
Three-dimensional image signals: processing methods
NASA Astrophysics Data System (ADS)
Schiopu, Paul; Manea, Adrian; Craciun, Anca-Ileana; Craciun, Alexandru
2010-11-01
Over the years extensive studies have been carried out to apply coherent optics methods in real-time processing, communications and transmission image. This is especially true when a large amount of information needs to be processed, e.g., in high-resolution imaging. The recent progress in data-processing networks and communication systems has considerably increased the capacity of information exchange. We describe the results of literature investigation research of processing methods for the signals of the three-dimensional images. All commercially available 3D technologies today are based on stereoscopic viewing. 3D technology was once the exclusive domain of skilled computer-graphics developers with high-end machines and software. The images capture from the advanced 3D digital camera can be displayed onto screen of the 3D digital viewer with/ without special glasses. For this is needed considerable processing power and memory to create and render the complex mix of colors, textures, and virtual lighting and perspective necessary to make figures appear three-dimensional. Also, using a standard digital camera and a technique called phase-shift interferometry we can capture "digital holograms." These are holograms that can be stored on computer and transmitted over conventional networks. We present some research methods to process "digital holograms" for the Internet transmission and results.
Accelerating EPI distortion correction by utilizing a modern GPU-based parallel computation.
Yang, Yao-Hao; Huang, Teng-Yi; Wang, Fu-Nien; Chuang, Tzu-Chao; Chen, Nan-Kuei
2013-04-01
The combination of phase demodulation and field mapping is a practical method to correct echo planar imaging (EPI) geometric distortion. However, since phase dispersion accumulates in each phase-encoding step, the calculation complexity of phase modulation is Ny-fold higher than conventional image reconstructions. Thus, correcting EPI images via phase demodulation is generally a time-consuming task. Parallel computing by employing general-purpose calculations on graphics processing units (GPU) can accelerate scientific computing if the algorithm is parallelized. This study proposes a method that incorporates the GPU-based technique into phase demodulation calculations to reduce computation time. The proposed parallel algorithm was applied to a PROPELLER-EPI diffusion tensor data set. The GPU-based phase demodulation method reduced the EPI distortion correctly, and accelerated the computation. The total reconstruction time of the 16-slice PROPELLER-EPI diffusion tensor images with matrix size of 128 × 128 was reduced from 1,754 seconds to 101 seconds by utilizing the parallelized 4-GPU program. GPU computing is a promising method to accelerate EPI geometric correction. The resulting reduction in computation time of phase demodulation should accelerate postprocessing for studies performed with EPI, and should effectuate the PROPELLER-EPI technique for clinical practice. Copyright © 2011 by the American Society of Neuroimaging.
Picazas-Márquez, Nerea; Sierra, María; Nova, Clara; Moreno, Juan Manuel; Aboitiz, Nuria; de Rivas, Gema; Sierra, Miguel A; Martínez-Álvarez, Roberto; Gómez-Caballero, Esther
2016-09-01
A new class of compounds, mono- and bis-haloethylphosphonates (HAPs and bisHAPs, respectively), listed in Schedule 2.B.04 of the Chemical Weapons Convention (CWC), has been synthesized and studied by GC-MS with two aims. First, to improve the identification of this type of chemicals by the Organization for the Prohibition of Chemical Weapons, (OPCW). Second, to study the synergistic effect of halogen and silicon atoms in molecules undergoing mass spectrometry. Fragmentation patterns of trimethylsilyl derivatives of HAPs were found to depend on the nature of the halogen atom; this was in agreement with DFT-calculations. The data suggest that a novel intramolecular halogen transfer takes place during the fragmentation process. Graphical Abstract ᅟ.
NASA Astrophysics Data System (ADS)
Yamaguchi, Masahiro; Haneishi, Hideaki; Fukuda, Hiroyuki; Kishimoto, Junko; Kanazawa, Hiroshi; Tsuchida, Masaru; Iwama, Ryo; Ohyama, Nagaaki
2006-01-01
In addition to the great advancement of high-resolution and large-screen imaging technology, the issue of color is now receiving considerable attention as another aspect than the image resolution. It is difficult to reproduce the original color of subject in conventional imaging systems, and that obstructs the applications of visual communication systems in telemedicine, electronic commerce, and digital museum. To breakthrough the limitation of conventional RGB 3-primary systems, "Natural Vision" project aims at an innovative video and still-image communication technology with high-fidelity color reproduction capability, based on spectral information. This paper summarizes the results of NV project including the development of multispectral and multiprimary imaging technologies and the experimental investigations on the applications to medicine, digital archives, electronic commerce, and computer graphics.
3D graphics hardware accelerator programming methods for real-time visualization systems
NASA Astrophysics Data System (ADS)
Souetov, Andrew E.
2001-02-01
The paper deals with new approaches in software design for creating real-time applications that use modern graphics acceleration hardware. The growing complexity of such type of software compels programmers to use different types of CASE systems in design and development process. The subject under discussion is integration of such systems in a development process, their effective use, and the combination of these new methods with the necessity to produce optimal codes. A method of simulation integration and modeling tools in real-time software development cycle is described.
3D graphics hardware accelerator programming methods for real-time visualization systems
NASA Astrophysics Data System (ADS)
Souetov, Andrew E.
2000-02-01
The paper deals with new approaches in software design for creating real-time applications that use modern graphics acceleration hardware. The growing complexity of such type of software compels programmers to use different types of CASE systems in design and development process. The subject under discussion is integration of such systems in a development process, their effective use, and the combination of these new methods with the necessity to produce optimal codes. A method of simulation integration and modeling tools in real-time software development cycle is described.
A Linux Workstation for High Performance Graphics
NASA Technical Reports Server (NTRS)
Geist, Robert; Westall, James
2000-01-01
The primary goal of this effort was to provide a low-cost method of obtaining high-performance 3-D graphics using an industry standard library (OpenGL) on PC class computers. Previously, users interested in doing substantial visualization or graphical manipulation were constrained to using specialized, custom hardware most often found in computers from Silicon Graphics (SGI). We provided an alternative to expensive SGI hardware by taking advantage of third-party, 3-D graphics accelerators that have now become available at very affordable prices. To make use of this hardware our goal was to provide a free, redistributable, and fully-compatible OpenGL work-alike library so that existing bodies of code could simply be recompiled. for PC class machines running a free version of Unix. This should allow substantial cost savings while greatly expanding the population of people with access to a serious graphics development and viewing environment. This should offer a means for NASA to provide a spectrum of graphics performance to its scientists, supplying high-end specialized SGI hardware for high-performance visualization while fulfilling the requirements of medium and lower performance applications with generic, off-the-shelf components and still maintaining compatibility between the two.
SPROC: A multiple-processor DSP IC
NASA Technical Reports Server (NTRS)
Davis, R.
1991-01-01
A large, single-chip, multiple-processor, digital signal processing (DSP) integrated circuit (IC) fabricated in HP-Cmos34 is presented. The innovative architecture is best suited for analog and real-time systems characterized by both parallel signal data flows and concurrent logic processing. The IC is supported by a powerful development system that transforms graphical signal flow graphs into production-ready systems in minutes. Automatic compiler partitioning of tasks among four on-chip processors gives the IC the signal processing power of several conventional DSP chips.
Augmented Computer Mouse Would Measure Applied Force
NASA Technical Reports Server (NTRS)
Li, Larry C. H.
1993-01-01
Proposed computer mouse measures force of contact applied by user. Adds another dimension to two-dimensional-position-measuring capability of conventional computer mouse; force measurement designated to represent any desired continuously variable function of time and position, such as control force, acceleration, velocity, or position along axis perpendicular to computer video display. Proposed mouse enhances sense of realism and intuition in interaction between operator and computer. Useful in such applications as three-dimensional computer graphics, computer games, and mathematical modeling of dynamics.
TCP Packet Trace Analysis. M.S. Thesis
NASA Technical Reports Server (NTRS)
Shepard, Timothy J.
1991-01-01
Examination of a trace of packets collected from the network is often the only method available for diagnosing protocol performance problems in computer networks. This thesis explores the use of packet traces to diagnose performance problems of the transport protocol TCP. Unfortunately, manual examination of these traces can be so tedious that effective analysis is not possible. The primary contribution of this thesis is a graphical method of displaying the packet trace which greatly reduce, the tediousness of examining a packet trace. The graphical method is demonstrated by the examination of some packet traces of typical TCP connections. The performance of two different implementations of TCP sending data across a particular network path is compared. Traces many thousands of packets long are used to demonstrate how effectively the graphical method simplifies examination of long complicated traces. In the comparison of the two TCP implementations, the burstiness of the TCP transmitter appeared to be related to the achieved throughput. A method of quantifying this burstiness is presented and its possible relevance to understanding the performance of TCP is discussed.
The Graphic Representation of Structure in Similarity/Dissimilarity Matrices: Alternative Methods.
ERIC Educational Resources Information Center
Rudnitsky, Alan N.
Three approaches to the graphic representation of similarity and dissimilarity matrices are compared and contrasted. Specifically, Kruskal's multidimensional scaling, Johnson's hierarchical clustering, and Waern's graphing techniques are employed to depict, in two dimensions, data representing the structure of a set of botanical concepts. Each of…
Item Screening in Graphical Loglinear Rasch Models
ERIC Educational Resources Information Center
Kreiner, Svend; Christensen, Karl Bang
2011-01-01
In behavioural sciences, local dependence and DIF are common, and purification procedures that eliminate items with these weaknesses often result in short scales with poor reliability. Graphical loglinear Rasch models (Kreiner & Christensen, in "Statistical Methods for Quality of Life Studies," ed. by M. Mesbah, F.C. Cole & M.T.…
Applying Semiotic Theories to Graphic Design Education: An Empirical Study on Poster Design Teaching
ERIC Educational Resources Information Center
Yang, Chao-Ming; Hsu, Tzu-Fan
2015-01-01
The rationales behind design are dissimilar to those behind art. Establishing an adequate theoretical foundation for conducting design education can facilitate scientising design methods. Thus, from the perspectives of the semiotic theories proposed by Saussure and Peirce, we investigated graphic design curricula by performing teaching…
The Myth of Fair Prices: A Graphical Analysis.
ERIC Educational Resources Information Center
Yanchus, Dennis; de Vanssay, Xavier
2003-01-01
Illustrates how fair price policies can be explained to undergraduate students by applying simple graphical methods normally used in general equilibrium trade theory. Indicates that fair price strategies can be looked upon as a suboptimal device for redistributing the gains from trade as compared with a transfer of funds. (JEH)
Periodicity and Some Graphical Insights on the Tendency toward Empty, Half-full, and Full Subshells.
ERIC Educational Resources Information Center
Rich, Ronald L.; Suter, Robert W.
1988-01-01
Investigates ground state electron configurations for some common elements using graphical methods. Bases observed tendencies on following ideas: "occupancy of differing shells, occupancy of differing subshells within a given shell, double occupancy vs. single occupancy of an orbital, and quantum-mechanical exchange." (ML)
An Evaluative and Prescriptive Look at Graphics Research.
ERIC Educational Resources Information Center
Soderston, Candace
Graphics research history shows that some topics have been studied heavily while others have been almost entirely neglected. Furthermore, researchers have used many different methods of defining and measuring effects such as legibility and comprehension, and this, together with vagueness in reporting, makes it difficult to compare studies and draw…
Slide shows vs graphic tablet live drawing for anatomy teaching.
Alsaid, B
2016-12-01
Blackboard drawing is the traditional and still widely learned method for anatomy teachers. However, for practical reasons, more and more lessons are done using slide shows. New digital learning tools are developed to create a more attractive teaching method. The objective of this study was to compare the use of graphic tablet live drawing versus slide shows. Sixty-five second-year students of the Faculty of Medicine participated in this study during their first semester of 2013-2014 academic year. The selected lecture dealt about neuroanatomy; two brain sections were taught: median sagittal and transverse. The sagittal section was presented via a slide show. The transverse section was taught using a graphics tablet using drawing software. Students were evaluated three times: before the lecture, immediately after the lecture and 8 weeks later. Means were compared using a t-test. Scores were significantly higher immediately after the lecture and 8 weeks later tests in comparing the transverse section (using the graphics tablet) versus the sagittal section (using PowerPoint ® ). Student satisfaction regarding the use of the tablet was high. The graphics tablet is a usable and efficient drawing tool in anatomy teaching. This tool requires a specific teacher training and preparation. Copyright © 2016 Elsevier Masson SAS. All rights reserved.
Engineering visualization utilizing advanced animation
NASA Technical Reports Server (NTRS)
Sabionski, Gunter R.; Robinson, Thomas L., Jr.
1989-01-01
Engineering visualization is the use of computer graphics to depict engineering analysis and simulation in visual form from project planning through documentation. Graphics displays let engineers see data represented dynamically which permits the quick evaluation of results. The current state of graphics hardware and software generally allows the creation of two types of 3D graphics. The use of animated video as an engineering visualization tool is presented. The engineering, animation, and videography aspects of animated video production are each discussed. Specific issues include the integration of staffing expertise, hardware, software, and the various production processes. A detailed explanation of the animation process reveals the capabilities of this unique engineering visualization method. Automation of animation and video production processes are covered and future directions are proposed.
NASA Technical Reports Server (NTRS)
Coles, W. A.
1975-01-01
The CAD/CAM interactive computer graphics system was described; uses to which it has been put were shown, and current developments of the system were outlined. The system supports batch, time sharing, and fully interactive graphic processing. Engineers using the system may switch between these methods of data processing and problem solving to make the best use of the available resources. It is concluded that the introduction of on-line computing in the form of teletypes, storage tubes, and fully interactive graphics has resulted in large increases in productivity and reduced timescales in the geometric computing, numerical lofting and part programming areas, together with a greater utilization of the system in the technical departments.
Integration of Modelling and Graphics to Create an Infrared Signal Processing Test Bed
NASA Astrophysics Data System (ADS)
Sethi, H. R.; Ralph, John E.
1989-03-01
The work reported in this paper was carried out as part of a contract with MoD (PE) UK. It considers the problems associated with realistic modelling of a passive infrared system in an operational environment. Ideally all aspects of the system and environment should be integrated into a complete end-to-end simulation but in the past limited computing power has prevented this. Recent developments in workstation technology and the increasing availability of parallel processing techniques makes the end-to-end simulation possible. However the complexity and speed of such simulations means difficulties for the operator in controlling the software and understanding the results. These difficulties can be greatly reduced by providing an extremely user friendly interface and a very flexible, high power, high resolution colour graphics capability. Most system modelling is based on separate software simulation of the individual components of the system itself and its environment. These component models may have their own characteristic inbuilt assumptions and approximations, may be written in the language favoured by the originator and may have a wide variety of input and output conventions and requirements. The models and their limitations need to be matched to the range of conditions appropriate to the operational scenerio. A comprehensive set of data bases needs to be generated by the component models and these data bases must be made readily available to the investigator. Performance measures need to be defined and displayed in some convenient graphics form. Some options are presented for combining available hardware and software to create an environment within which the models can be integrated, and which provide the required man-machine interface, graphics and computing power. The impact of massively parallel processing and artificial intelligence will be discussed. Parallel processing will make real time end-to-end simulation possible and will greatly improve the graphical visualisation of the model output data. Artificial intelligence should help to enhance the man-machine interface.
Jeong, Kyeong-Min; Kim, Hee-Seung; Hong, Sung-In; Lee, Sung-Keun; Jo, Na-Young; Kim, Yong-Soo; Lim, Hong-Gi; Park, Jae-Hyeung
2012-10-08
Speed enhancement of integral imaging based incoherent Fourier hologram capture using a graphic processing unit is reported. Integral imaging based method enables exact hologram capture of real-existing three-dimensional objects under regular incoherent illumination. In our implementation, we apply parallel computation scheme using the graphic processing unit, accelerating the processing speed. Using enhanced speed of hologram capture, we also implement a pseudo real-time hologram capture and optical reconstruction system. The overall operation speed is measured to be 1 frame per second.
Advanced graphical user interface for multi-physics simulations using AMST
NASA Astrophysics Data System (ADS)
Hoffmann, Florian; Vogel, Frank
2017-07-01
Numerical modelling of particulate matter has gained much popularity in recent decades. Advanced Multi-physics Simulation Technology (AMST) is a state-of-the-art three dimensional numerical modelling technique combining the eX-tended Discrete Element Method (XDEM) with Computational Fluid Dynamics (CFD) and Finite Element Analysis (FEA) [1]. One major limitation of this code is the lack of a graphical user interface (GUI) meaning that all pre-processing has to be made directly in a HDF5-file. This contribution presents the first graphical pre-processor developed for AMST.
Interactive voxel graphics in virtual reality
NASA Astrophysics Data System (ADS)
Brody, Bill; Chappell, Glenn G.; Hartman, Chris
2002-06-01
Interactive voxel graphics in virtual reality poses significant research challenges in terms of interface, file I/O, and real-time algorithms. Voxel graphics is not so new, as it is the focus of a good deal of scientific visualization. Interactive voxel creation and manipulation is a more innovative concept. Scientists are understandably reluctant to manipulate data. They collect or model data. A scientific analogy to interactive graphics is the generation of initial conditions for some model. It is used as a method to test those models. We, however, are in the business of creating new data in the form of graphical imagery. In our endeavor, science is a tool and not an end. Nevertheless, there is a whole class of interactions and associated data generation scenarios that are natural to our way of working and that are also appropriate to scientific inquiry. Annotation by sketching or painting to point to and distinguish interesting and important information is very significant for science as well as art. Annotation in 3D is difficult without a good 3D interface. Interactive graphics in virtual reality is an appropriate approach to this problem.
Graphic tracings of condylar paths and measurements of condylar angles.
el-Gheriani, A S; Winstanley, R B
1989-01-01
A study was carried out to determine the accuracy of different methods of measuring condylar inclination from graphical recordings of condylar paths. Thirty subjects made protrusive mandibular movements while condylar inclination was recorded on a graph paper card. A mandibular facebow and intraoral central bearing plate facilitated the procedure. The first method proved to be too variable to be of value in measuring condylar angles. The spline curve fitting technique was shown to be accurate, but its use clinically may prove complex. The mathematical method was more practical and overcame the variability of the tangent method. Other conclusions regarding condylar inclination are outlined.
Classical Control System Design: A non-Graphical Method for Finding the Exact System Parameters
NASA Astrophysics Data System (ADS)
Hussein, Mohammed Tawfik
2008-06-01
The Root Locus method of control system design was developed in the 1940's. It is a set of rules that helps in sketching the path traced by the roots of the closed loop characteristic equation of the system, as a parameter such as a controller gain, k, is varied. The procedure provides approximate sketching guidelines. Designs on control systems using the method are therefore not exact. This paper aims at a non-graphical method for finding the exact system parameters to place a pair of complex conjugate poles on a specified damping ratio line. The overall procedure is based on the exact solution of complex equations on the PC using numerical methods.
Computer Graphics-aided systems analysis: application to well completion design
DOE Office of Scientific and Technical Information (OSTI.GOV)
Detamore, J.E.; Sarma, M.P.
1985-03-01
The development of an engineering tool (in the form of a computer model) for solving design and analysis problems related with oil and gas well production operations is discussed. The development of the method is based on integrating the concepts of ''Systems Analysis'' with the techniques of ''Computer Graphics''. The concepts behind the method are very general in nature. This paper, however, illustrates the application of the method in solving gas well completion design problems. The use of the method will save time and improve the efficiency of such design and analysis problems. The method can be extended to othermore » design and analysis aspects of oil and gas wells.« less
USDA-ARS?s Scientific Manuscript database
This research tested whether children could categorize foods more accurately and speedily when presented with child-generated rather than professionally-generated food categories; and whether a graphically appealing browse procedure similar to the Apple, Inc, "cover flow" graphical user interface ac...
A Case Study on Pre-Service Teachers Students' Interaction with Graphical Artefacts
ERIC Educational Resources Information Center
Olande, Oduor
2014-01-01
This study reports from a pre-service teacher's online learning and assessment activity on determining variability of two graphical artefacts. Using a critical-analytical perspective to data, the present study indicates that the prospective teachers surveyed showed awareness of relevant subject specific operators and methods; however, these seem…
The MATH--Open Source Application for Easier Learning of Numerical Mathematics
ERIC Educational Resources Information Center
Glaser-Opitz, Henrich; Budajová, Kristina
2016-01-01
The article introduces a software application (MATH) supporting an education of Applied Mathematics, with focus on Numerical Mathematics. The MATH is an easy to use tool supporting various numerical methods calculations with graphical user interface and integrated plotting tool for graphical representation written in Qt with extensive use of Qwt…
Comparing Graphical and Verbal Representations of Measurement Error in Test Score Reports
ERIC Educational Resources Information Center
Zwick, Rebecca; Zapata-Rivera, Diego; Hegarty, Mary
2014-01-01
Research has shown that many educators do not understand the terminology or displays used in test score reports and that measurement error is a particularly challenging concept. We investigated graphical and verbal methods of representing measurement error associated with individual student scores. We created four alternative score reports, each…
Observing Student Working Styles when Using Graphic Calculators to Solve Mathematics Problems
ERIC Educational Resources Information Center
Berry, J.; Graham, E.; Smith, A.
2006-01-01
Some research studies, many of which used quantitative methods, have suggested that graphics calculators can be used to effectively enhance the learning of mathematics. More recently research studies have started to explore students' styles of working as they solve problems with technology. This paper describes the use of a software application…
A Prototype HTML Training System for Graphic Communication Majors
ERIC Educational Resources Information Center
Runquist, Roger L.
2010-01-01
This design research demonstrates a prototype content management system capable of training graphic communication students in the creation of basic HTML web pages. The prototype serve as a method of helping students learn basic HTML structure and commands earlier in their academic careers. Exposure to the concepts of web page creation early in…
Inouye, David I.; Ravikumar, Pradeep; Dhillon, Inderjit S.
2016-01-01
We develop Square Root Graphical Models (SQR), a novel class of parametric graphical models that provides multivariate generalizations of univariate exponential family distributions. Previous multivariate graphical models (Yang et al., 2015) did not allow positive dependencies for the exponential and Poisson generalizations. However, in many real-world datasets, variables clearly have positive dependencies. For example, the airport delay time in New York—modeled as an exponential distribution—is positively related to the delay time in Boston. With this motivation, we give an example of our model class derived from the univariate exponential distribution that allows for almost arbitrary positive and negative dependencies with only a mild condition on the parameter matrix—a condition akin to the positive definiteness of the Gaussian covariance matrix. Our Poisson generalization allows for both positive and negative dependencies without any constraints on the parameter values. We also develop parameter estimation methods using node-wise regressions with ℓ1 regularization and likelihood approximation methods using sampling. Finally, we demonstrate our exponential generalization on a synthetic dataset and a real-world dataset of airport delay times. PMID:27563373
Validation of Mobility of Pedestrians with Low Vision Using Graphic Floor Signs and Voice Guides.
Omori, Kiyohiro; Yanagihara, Takao; Kitagawa, Hiroshi; Ikeda, Norihiro
2015-01-01
Some people with low vision or elderly persons tend to walk while watching a nearby floor, therefore, they often overlook or hard to read suspended signs. In this study, we propose two kinds of voice guides, and an experiment is conducted by participants with low vision using these voice guides and graphic floor signs in order to investigate effectiveness of these combinations. In clock position method (CP), each direction of near facilities are described in using an analogy of a 12-hour clock. Meanwhile, in numbering method (NU), near facilities are put the number in clockwise order, however, each direction are only illustrated in a crossing sign. As a result of an experiment, it is showed that both voice guides are effective for pedestrians with low vision. NU is used as a complement of graphic floor signs. Meanwhile, CP is used independently with graphic floor signs, however, there is a risk in the case of using in the environment where pedestrians are easy to mistake the reference direction defined by the sounding speaker.
Minh, Hoang Van; Chung, Le Hong; Giang, Kim Bao; Duc, Duong Minh; Hinh, Nguyen Duc; Mai, Vu Quynh; Cuong, Nguyen Manh; Manh, Pham Duc; Duc, Ha Anh; Yang, Jui-Chen
2016-01-01
Two years after implementation of the graphic health warning intervention in Vietnam, it is very important to evaluate the intervention's potential impact. The objective of this paper was to predict effects of graphic health warnings on cigarette packages, particularly in reducing cigarette demand and smoking-associated deaths in Vietnam. In this study, a discrete choice experiment (DCE) method was used to evaluate the potential impact of graphic tobacco health warnings on smoking demand. To predict the impact of GHWs on reducing premature deaths associated with smoking, we constructed different static models. We adapted the method developed by University of Toronto, Canada and found that GHWs had statistically significant impact on reducing cigarette demand (up to 10.1% through images of lung damage), resulting in an overall decrease of smoking prevalence in Vietnam. We also found that between 428,417- 646,098 premature deaths would be prevented as a result of the GHW intervention. The potential impact of the GHW labels on reducing premature smoking-associated deaths in Vietnam were shown to be stronger among lower socio-economic groups.
NASTRAN data generation and management using interactive graphics
NASA Technical Reports Server (NTRS)
Smootkatow, M.; Cooper, B. M.
1972-01-01
A method of using an interactive graphics device to generate a large portion of the input bulk data with visual checks of the structure and the card images is described. The generation starts from GRID and PBAR cards. The visual checks result from a three-dimensional display of the model in any rotated position. By detailing the steps, the time saving and cost effectiveness of this method may be judged, and its potential as a useful tool for the structural analyst may be established.
Lee, Young Han; Park, Eun Hae; Suh, Jin-Suck
2015-01-01
The objectives are: 1) to introduce a simple and efficient method for extracting region of interest (ROI) values from a Picture Archiving and Communication System (PACS) viewer using optical character recognition (OCR) software and a macro program, and 2) to evaluate the accuracy of this method with a PACS workstation. This module was designed to extract the ROI values on the images of the PACS, and created as a development tool by using open-source OCR software and an open-source macro program. The principal processes are as follows: (1) capture a region of the ROI values as a graphic file for OCR, (2) recognize the text from the captured image by OCR software, (3) perform error-correction, (4) extract the values including area, average, standard deviation, max, and min values from the text, (5) reformat the values into temporary strings with tabs, and (6) paste the temporary strings into the spreadsheet. This principal process was repeated for the number of ROIs. The accuracy of this module was evaluated on 1040 recognitions from 280 randomly selected ROIs of the magnetic resonance images. The input times of ROIs were compared between conventional manual method and this extraction module-assisted input method. The module for extracting ROI values operated successfully using the OCR and macro programs. The values of the area, average, standard deviation, maximum, and minimum could be recognized and error-corrected with AutoHotkey-coded module. The average input times using the conventional method and the proposed module-assisted method were 34.97 seconds and 7.87 seconds, respectively. A simple and efficient method for ROI value extraction was developed with open-source OCR and a macro program. Accurate inputs of various numbers from ROIs can be extracted with this module. The proposed module could be applied to the next generation of PACS or existing PACS that have not yet been upgraded. Copyright © 2015 AUR. Published by Elsevier Inc. All rights reserved.
Graphical determination of wall temperatures for heat transfers through walls of arbitrary shape
NASA Technical Reports Server (NTRS)
Lutz, Otto
1950-01-01
A graphical method is given which permits determining of the temperature distribution during heat transfer in arbitrarily shaped walls. Three examples show the application of the method. The further development of heat engines depends to a great extent on the control of the thermal stresses in the walls. The thermal stresses stem from the nonuniform temperature distribution in heat transfer through walls which are, for structural reasons, of various thicknesses and sometimes complicated shape. Thus, it is important to know the temperature distribution in these structural parts. Following, a method is given which permits solution of this problem.
[Automatic tracing of conversion scales from conventional units to the SI system of units].
Besozzi, M; Bianchi, P; Agrifoglio, L
1988-01-01
American medical journals, as the Journal of the American Medical Association (JAMA), and the American Journal of Clinical Pathology (AJCP), the Journal of the American Society of Clinical Pathologists (ASCP), are shifting to selected SI (Système International d'Unités) units for reporting measurements. Further discussion by the AMA, the ASCP and other organizations is required before consensus in the US medical community can be reached as to the extent of and time frame for conversion to SI for reporting clinical laboratory measurements: however this decision will certainly greatly speed up the process of conversion in European countries too. Transition to SI units will require the use of different reference ranges, and there will be a potential for serious misinterpretation of laboratory data unless well-planned educational programs are instituted before the change. A simple program written in Microsoft Basic for automatically tracing on one's personal computer (PC) monitor a dual scale, in the conventional and in the SI system of units, is presented here. The program may be easily implemented and run on every PC operating under MS-DOS, equipped with a CGA or an AT&T6300 graphic card: through the operating system the scales may also be printed on a dot-matrix graphic printer. We believe that this, and other tools of this kind, will be useful in the thorough educational process of those reading the reports, and will be an important factor in the success of conversion to SI reporting.
Structural identifiability of cyclic graphical models of biological networks with latent variables.
Wang, Yulin; Lu, Na; Miao, Hongyu
2016-06-13
Graphical models have long been used to describe biological networks for a variety of important tasks such as the determination of key biological parameters, and the structure of graphical model ultimately determines whether such unknown parameters can be unambiguously obtained from experimental observations (i.e., the identifiability problem). Limited by resources or technical capacities, complex biological networks are usually partially observed in experiment, which thus introduces latent variables into the corresponding graphical models. A number of previous studies have tackled the parameter identifiability problem for graphical models such as linear structural equation models (SEMs) with or without latent variables. However, the limited resolution and efficiency of existing approaches necessarily calls for further development of novel structural identifiability analysis algorithms. An efficient structural identifiability analysis algorithm is developed in this study for a broad range of network structures. The proposed method adopts the Wright's path coefficient method to generate identifiability equations in forms of symbolic polynomials, and then converts these symbolic equations to binary matrices (called identifiability matrix). Several matrix operations are introduced for identifiability matrix reduction with system equivalency maintained. Based on the reduced identifiability matrices, the structural identifiability of each parameter is determined. A number of benchmark models are used to verify the validity of the proposed approach. Finally, the network module for influenza A virus replication is employed as a real example to illustrate the application of the proposed approach in practice. The proposed approach can deal with cyclic networks with latent variables. The key advantage is that it intentionally avoids symbolic computation and is thus highly efficient. Also, this method is capable of determining the identifiability of each single parameter and is thus of higher resolution in comparison with many existing approaches. Overall, this study provides a basis for systematic examination and refinement of graphical models of biological networks from the identifiability point of view, and it has a significant potential to be extended to more complex network structures or high-dimensional systems.
NASA Orbital Debris Engineering Model ORDEM2008 (Beta Version)
NASA Technical Reports Server (NTRS)
Stansbery, Eugene G.; Krisko, Paula H.
2009-01-01
This is an interim document intended to accompany the beta-release of the ORDEM2008 model. As such it provides the user with a guide for its use, a list of its capabilities, a brief summary of model development, and appendices included to educate the user as to typical runtimes for different orbit configurations. More detailed documentation will be delivered with the final product. ORDEM2008 supersedes NASA's previous model - ORDEM2000. The availability of new sensor and in situ data, the re-analysis of older data, and the development of new analytical techniques, has enabled the construction of this more comprehensive and sophisticated model. Integrated with the software is an upgraded graphical user interface (GUI), which uses project-oriented organization and provides the user with graphical representations of numerous output data products. These range from the conventional average debris size vs. flux magnitude for chosen analysis orbits, to the more complex color-contoured two-dimensional (2-D) directional flux diagrams in terms of local spacecraft pitch and yaw.
Off-the-shelf real-time monitoring of satellite constellations in a visual 3-D environment
NASA Technical Reports Server (NTRS)
Schwuttke, Ursula M.; Hervias, Felipe; Cheng, Cecilia Han; Mactutis, Anthony; Angelino, Robert
1996-01-01
The multimission spacecraft analysis system (MSAS) data monitor is a generic software product for future real-time data monitoring and analysis. The system represents the status of a satellite constellation through the shape, color, motion and position of graphical objects floating in a three dimensional virtual reality environment. It may be used for the monitoring of large volumes of data, for viewing results in configurable displays, and for providing high level and detailed views of a constellation of monitored satellites. It is considered that the data monitor is an improvement on conventional graphic and text-based displays as it increases the amount of data that the operator can absorb in a given period, and can be installed and configured without the requirement for software development by the end user. The functionality of the system is described, including: the navigation abilities; the representation of alarms in the cybergrid; limit violation; real-time trend analysis, and alarm status indication.
Parallel Rendering of Large Time-Varying Volume Data
NASA Technical Reports Server (NTRS)
Garbutt, Alexander E.
2005-01-01
Interactive visualization of large time-varying 3D volume datasets has been and still is a great challenge to the modem computational world. It stretches the limits of the memory capacity, the disk space, the network bandwidth and the CPU speed of a conventional computer. In this SURF project, we propose to develop a parallel volume rendering program on SGI's Prism, a cluster computer equipped with state-of-the-art graphic hardware. The proposed program combines both parallel computing and hardware rendering in order to achieve an interactive rendering rate. We use 3D texture mapping and a hardware shader to implement 3D volume rendering on each workstation. We use SGI's VisServer to enable remote rendering using Prism's graphic hardware. And last, we will integrate this new program with ParVox, a parallel distributed visualization system developed at JPL. At the end of the project, we Will demonstrate remote interactive visualization using this new hardware volume renderer on JPL's Prism System using a time-varying dataset from selected JPL applications.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Von Dreele, Robert
One of the goals in developing GSAS-II was to expand from the capabilities of the original General Structure Analysis System (GSAS) which largely encompassed just structure refinement and post refinement analysis. GSAS-II has been written almost entirely in Python loaded with graphics, GUI and mathematical packages (matplotlib, pyOpenGL, wxpython, numpy and scipy). Thus, GSAS-II has a fully developed modern GUI as well as extensive graphical display of data and results. However, the structure and operation of Python has required new approaches to many of the algorithms used in crystal structure analysis. The extensions beyond GSAS include image calibration/integration as wellmore » as peak fitting and unit cell indexing for powder data which are precursors for structure solution. Structure solution within GSAS-II begins with either Pawley or LeBail extracted structure factors from powder data or those measured in a single crystal experiment. Both charge flipping and Monte Carlo-Simulated Annealing techniques are available; the former can be applied to (3+1) incommensurate structures as well as conventional 3D structures.« less
A study of computer graphics technology in application of communication resource management
NASA Astrophysics Data System (ADS)
Li, Jing; Zhou, Liang; Yang, Fei
2017-08-01
With the development of computer technology, computer graphics technology has been widely used. Especially, the success of object-oriented technology and multimedia technology promotes the development of graphics technology in the computer software system. Therefore, the computer graphics theory and application technology have become an important topic in the field of computer, while the computer graphics technology becomes more and more extensive in various fields of application. In recent years, with the development of social economy, especially the rapid development of information technology, the traditional way of communication resource management cannot effectively meet the needs of resource management. In this case, the current communication resource management is still using the original management tools and management methods, resource management equipment management and maintenance, which brought a lot of problems. It is very difficult for non-professionals to understand the equipment and the situation in communication resource management. Resource utilization is relatively low, and managers cannot quickly and accurately understand the resource conditions. Aimed at the above problems, this paper proposes to introduce computer graphics technology into the communication resource management. The introduction of computer graphics not only makes communication resource management more vivid, but also reduces the cost of resource management and improves work efficiency.
NASA Astrophysics Data System (ADS)
Lackner, Bettina C.; Kirchengast, Gottfried
2015-04-01
Besides written and spoken language, graphical displays play an important role in communicating scientific findings or explaining scientific methods, both within one and between various disciplines. Uncertainties and probabilities are generally difficult to communicate, especially via graphics. Graphics including uncertainty sometimes need detailed written or oral descriptions to be understood. "Good" graphics should ease scientific communication, especially amongst different disciplines. One key objective of the Doctoral Programme "Climate Change: Uncertainties, Thresholds and Coping Strategies" (http://dk-climate-change.uni-graz.at/en/), located at the University of Graz, is to reach a better understanding of climate change uncertainties by bridging research in multiple disciplines, including physical climate sciences, geosciences, systems and sustainability sciences, environmental economics, and climate ethics. This asks for efforts into the formulation of a "common language", not only as to words, but also as to graphics. The focus of this work is on two topics: (1) What different kinds of uncertainties (e.g., data uncertainty, model uncertainty) are included in the graphics of the recent IPCC reports of all three working groups (WGs) and in what ways do uncertainties get illustrated? (2) How are these graphically displayed uncertainties perceived by researchers of a similar research discipline and from researchers of different disciplines than the authors of the graphics? To answer the first question, the IPCC graphics including uncertainties are grouped and analyzed with respect to different kinds of uncertainties to filter out most of the commonly used types of displays. The graphics will also be analyzed with respect to their WG origin, as we assume that graphics from researchers rooted in, e.g., physical climate sciences and geosciences (mainly IPCC WG 1) differ from those of researchers rooted in, e.g., economics or system sciences (mainly WG 3). In a subsequent analysis, some basic types of graphics displaying uncertainty are selected to serve as input for the construction of "makeshift graphics" (displaying only the main features but including no detailed title or caption). These makeshift graphics are then used to assess how the displayed features are perceived and understood by researchers of various disciplines. In this initial study, this analysis will be based on results of a workshop including the wide diversity of researchers within the FWF-DK Climate Change. We will present first results of this work.
NASA Technical Reports Server (NTRS)
Anderson, Frederick; Biezad, Daniel J.
1994-01-01
This paper describes the Rapid Aircraft DynamIcs AssessmeNt (RADIAN) project - an integration of the Aircraft SYNThesis (ACSTNT) design code with the USAD DATCOM code that estimates stability derivatives. Both of these codes are available to universities. These programs are then linked to flight simulation and flight controller synthesis tools and resulting design is evaluated on a graphics workstation. The entire process reduces the preliminary design time by an order of magnitude and provides an initial handling qualities evaluation of the design coupled to a control law. The integrated design process is applicable to both conventional aircraft taken from current textbooks and to unconventional designs emphasizing agility and propulsive control of attitude. The interactive and concurrent nature of the design process has been well received by industry and by design engineers at NASA. The process is being implemented into the design curriculum and is being used by students who view it as a significant advance over prior methods.
Gustafsson, Nils; Culley, Siân; Ashdown, George; Owen, Dylan M.; Pereira, Pedro Matos; Henriques, Ricardo
2016-01-01
Despite significant progress, high-speed live-cell super-resolution studies remain limited to specialized optical setups, generally requiring intense phototoxic illumination. Here, we describe a new analytical approach, super-resolution radial fluctuations (SRRF), provided as a fast graphics processing unit-enabled ImageJ plugin. In the most challenging data sets for super-resolution, such as those obtained in low-illumination live-cell imaging with GFP, we show that SRRF is generally capable of achieving resolutions better than 150 nm. Meanwhile, for data sets similar to those obtained in PALM or STORM imaging, SRRF achieves resolutions approaching those of standard single-molecule localization analysis. The broad applicability of SRRF and its performance at low signal-to-noise ratios allows super-resolution using modern widefield, confocal or TIRF microscopes with illumination orders of magnitude lower than methods such as PALM, STORM or STED. We demonstrate this by super-resolution live-cell imaging over timescales ranging from minutes to hours. PMID:27514992
Object oriented studies into artificial space debris
NASA Technical Reports Server (NTRS)
Adamson, J. M.; Marshall, G.
1988-01-01
A prototype simulation is being developed under contract to the Royal Aerospace Establishment (RAE), Farnborough, England, to assist in the discrimination of artificial space objects/debris. The methodology undertaken has been to link Object Oriented programming, intelligent knowledge based system (IKBS) techniques and advanced computer technology with numeric analysis to provide a graphical, symbolic simulation. The objective is to provide an additional layer of understanding on top of conventional classification methods. Use is being made of object and rule based knowledge representation, multiple reasoning, truth maintenance and uncertainty. Software tools being used include Knowledge Engineering Environment (KEE) and SymTactics for knowledge representation. Hooks are being developed within the SymTactics framework to incorporate mathematical models describing orbital motion and fragmentation. Penetration and structural analysis can also be incorporated. SymTactics is an Object Oriented discrete event simulation tool built as a domain specific extension to the KEE environment. The tool provides facilities for building, debugging and monitoring dynamic (military) simulations.
Design and Evaluation of Perceptual-based Object Group Selection Techniques
NASA Astrophysics Data System (ADS)
Dehmeshki, Hoda
Selecting groups of objects is a frequent task in graphical user interfaces. It is required prior to many standard operations such as deletion, movement, or modification. Conventional selection techniques are lasso, rectangle selection, and the selection and de-selection of items through the use of modifier keys. These techniques may become time-consuming and error-prone when target objects are densely distributed or when the distances between target objects are large. Perceptual-based selection techniques can considerably improve selection tasks when targets have a perceptual structure, for example when arranged along a line. Current methods to detect such groups use ad hoc grouping algorithms that are not based on results from perception science. Moreover, these techniques do not allow selecting groups with arbitrary arrangements or permit modifying a selection. This dissertation presents two domain-independent perceptual-based systems that address these issues. Based on established group detection models from perception research, the proposed systems detect perceptual groups formed by the Gestalt principles of good continuation and proximity. The new systems provide gesture-based or click-based interaction techniques for selecting groups with curvilinear or arbitrary structures as well as clusters. Moreover, the gesture-based system is adapted for the graph domain to facilitate path selection. This dissertation includes several user studies that show the proposed systems outperform conventional selection techniques when targets form salient perceptual groups and are still competitive when targets are semi-structured.
NASA Astrophysics Data System (ADS)
O'Shea, Bethany; Jankowski, Jerzy
2006-12-01
The major ion composition of Great Artesian Basin groundwater in the lower Namoi River valley is relatively homogeneous in chemical composition. Traditional graphical techniques have been combined with multivariate statistical methods to determine whether subtle differences in the chemical composition of these waters can be delineated. Hierarchical cluster analysis and principal components analysis were successful in delineating minor variations within the groundwaters of the study area that were not visually identified in the graphical techniques applied. Hydrochemical interpretation allowed geochemical processes to be identified in each statistically defined water type and illustrated how these groundwaters differ from one another. Three main geochemical processes were identified in the groundwaters: ion exchange, precipitation, and mixing between waters from different sources. Both statistical methods delineated an anomalous sample suspected of being influenced by magmatic CO2 input. The use of statistical methods to complement traditional graphical techniques for waters appearing homogeneous is emphasized for all investigations of this type. Copyright
Stochastic Spectral Descent for Discrete Graphical Models
Carlson, David; Hsieh, Ya-Ping; Collins, Edo; ...
2015-12-14
Interest in deep probabilistic graphical models has in-creased in recent years, due to their state-of-the-art performance on many machine learning applications. Such models are typically trained with the stochastic gradient method, which can take a significant number of iterations to converge. Since the computational cost of gradient estimation is prohibitive even for modestly sized models, training becomes slow and practically usable models are kept small. In this paper we propose a new, largely tuning-free algorithm to address this problem. Our approach derives novel majorization bounds based on the Schatten- norm. Intriguingly, the minimizers of these bounds can be interpreted asmore » gradient methods in a non-Euclidean space. We thus propose using a stochastic gradient method in non-Euclidean space. We both provide simple conditions under which our algorithm is guaranteed to converge, and demonstrate empirically that our algorithm leads to dramatically faster training and improved predictive ability compared to stochastic gradient descent for both directed and undirected graphical models.« less
Graphical programming interface: A development environment for MRI methods.
Zwart, Nicholas R; Pipe, James G
2015-11-01
To introduce a multiplatform, Python language-based, development environment called graphical programming interface for prototyping MRI techniques. The interface allows developers to interact with their scientific algorithm prototypes visually in an event-driven environment making tasks such as parameterization, algorithm testing, data manipulation, and visualization an integrated part of the work-flow. Algorithm developers extend the built-in functionality through simple code interfaces designed to facilitate rapid implementation. This article shows several examples of algorithms developed in graphical programming interface including the non-Cartesian MR reconstruction algorithms for PROPELLER and spiral as well as spin simulation and trajectory visualization of a FLORET example. The graphical programming interface framework is shown to be a versatile prototyping environment for developing numeric algorithms used in the latest MR techniques. © 2014 Wiley Periodicals, Inc.
Freiberger, Manuel; Egger, Herbert; Liebmann, Manfred; Scharfetter, Hermann
2011-11-01
Image reconstruction in fluorescence optical tomography is a three-dimensional nonlinear ill-posed problem governed by a system of partial differential equations. In this paper we demonstrate that a combination of state of the art numerical algorithms and a careful hardware optimized implementation allows to solve this large-scale inverse problem in a few seconds on standard desktop PCs with modern graphics hardware. In particular, we present methods to solve not only the forward but also the non-linear inverse problem by massively parallel programming on graphics processors. A comparison of optimized CPU and GPU implementations shows that the reconstruction can be accelerated by factors of about 15 through the use of the graphics hardware without compromising the accuracy in the reconstructed images.
Analyzing women's roles through graphic representation of narratives.
Hall, Joanne M
2003-08-01
A 1992 triangulated international nursing study of women's health was reported. The researchers used the perspectives of feminism and symbolic interactionism, specifically role theory. A narrative analysis was done to clarify the concept of role integration. The narrative analysis was reported in 1992, but graphic/visual techniques used in the team dialogue process of narrative analysis were not reported due to space limitations. These techniques have not been reported elsewhere and thus remain innovative. Specific steps in the method are outlined here in detail as an audit trail. The process would be useful to other qualitative researchers as an exemplar of one novel way that verbal data can be abstracted visually/graphically. Suggestions are included for aspects of narrative, in addition to roles, that could be depicted graphically in qualitative research.
Advanced display object selection methods for enhancing user-computer productivity
NASA Technical Reports Server (NTRS)
Osga, Glenn A.
1993-01-01
The User-Interface Technology Branch at NCCOSC RDT&E Division has been conducting a series of studies to address the suitability of commercial off-the-shelf (COTS) graphic user-interface (GUI) methods for efficiency and performance in critical naval combat systems. This paper presents an advanced selection algorithm and method developed to increase user performance when making selections on tactical displays. The method has also been applied with considerable success to a variety of cursor and pointing tasks. Typical GUI's allow user selection by: (1) moving a cursor with a pointing device such as a mouse, trackball, joystick, touchscreen; and (2) placing the cursor on the object. Examples of GUI objects are the buttons, icons, folders, scroll bars, etc. used in many personal computer and workstation applications. This paper presents an improved method of selection and the theoretical basis for the significant performance gains achieved with various input devices tested. The method is applicable to all GUI styles and display sizes, and is particularly useful for selections on small screens such as notebook computers. Considering the amount of work-hours spent pointing and clicking across all styles of available graphic user-interfaces, the cost/benefit in applying this method to graphic user-interfaces is substantial, with the potential for increasing productivity across thousands of users and applications.
De Los Reyes, Andres; Augenstein, Tara M; Aldao, Amelia; Thomas, Sarah A; Daruwala, Samantha; Kline, Kathryn; Regan, Timothy
2015-01-01
Social stressor tasks induce adolescents' social distress as indexed by low-cost psychophysiological methods. Unknown is how to incorporate these methods within clinical assessments. Having assessors judge graphical depictions of psychophysiological data may facilitate detections of data patterns that may be difficult to identify using judgments about numerical depictions of psychophysiological data. Specifically, the Chernoff Face method involves graphically representing data using features on the human face (eyes, nose, mouth, and face shape). This method capitalizes on humans' abilities to discern subtle variations in facial features. Using adolescent heart rate norms and Chernoff Faces, we illustrated a method for implementing psychophysiology within clinical assessments of adolescent social anxiety. Twenty-two clinic-referred adolescents completed a social anxiety self-report and provided psychophysiological data using wireless heart rate monitors during a social stressor task. We graphically represented participants' psychophysiological data and normative adolescent heart rates. For each participant, two undergraduate coders made comparative judgments between the dimensions (eyes, nose, mouth, and face shape) of two Chernoff Faces. One Chernoff Face represented a participant's heart rate within a context (baseline, speech preparation, or speech-giving). The second Chernoff Face represented normative heart rate data matched to the participant's age. Using Chernoff Faces, coders reliably and accurately identified contextual variation in participants' heart rate responses to social stress. Further, adolescents' self-reported social anxiety symptoms predicted Chernoff Face judgments, and judgments could be differentiated by social stress context. Our findings have important implications for implementing psychophysiology within clinical assessments of adolescent social anxiety.
From least squares to multilevel modeling: A graphical introduction to Bayesian inference
NASA Astrophysics Data System (ADS)
Loredo, Thomas J.
2016-01-01
This tutorial presentation will introduce some of the key ideas and techniques involved in applying Bayesian methods to problems in astrostatistics. The focus will be on the big picture: understanding the foundations (interpreting probability, Bayes's theorem, the law of total probability and marginalization), making connections to traditional methods (propagation of errors, least squares, chi-squared, maximum likelihood, Monte Carlo simulation), and highlighting problems where a Bayesian approach can be particularly powerful (Poisson processes, density estimation and curve fitting with measurement error). The "graphical" component of the title reflects an emphasis on pictorial representations of some of the math, but also on the use of graphical models (multilevel or hierarchical models) for analyzing complex data. Code for some examples from the talk will be available to participants, in Python and in the Stan probabilistic programming language.
Quick estimate of oil discovery from gas-condensate reservoirs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sarem, A.M.
1966-10-24
A quick method of estimating the depletion performance of gas-condensate reservoirs is presented by graphical representations. The method is based on correlations reported in the literature and expresses recoverable liquid as function of gas reserves, producing gas-oil ratio, and initial and final reservoir pressures. The amount of recoverable liquid reserves (RLR) under depletion conditions, is estimated from an equation which is given. Where the liquid-reserves are in stock-tank barrels the gas reserves are in Mcf, with the arbitrary constant, N calculated from one graphical representation by dividing fractional oil recovery by the initial gas-oil ratio and multiplying 10U6D for convenience.more » An equation is given for estimating the coefficient C. These factors (N and C) can be determined from the graphical representations. An example calculation is included.« less
On an Additive Semigraphoid Model for Statistical Networks With Application to Pathway Analysis.
Li, Bing; Chun, Hyonho; Zhao, Hongyu
2014-09-01
We introduce a nonparametric method for estimating non-gaussian graphical models based on a new statistical relation called additive conditional independence, which is a three-way relation among random vectors that resembles the logical structure of conditional independence. Additive conditional independence allows us to use one-dimensional kernel regardless of the dimension of the graph, which not only avoids the curse of dimensionality but also simplifies computation. It also gives rise to a parallel structure to the gaussian graphical model that replaces the precision matrix by an additive precision operator. The estimators derived from additive conditional independence cover the recently introduced nonparanormal graphical model as a special case, but outperform it when the gaussian copula assumption is violated. We compare the new method with existing ones by simulations and in genetic pathway analysis.
A semiparametric graphical modelling approach for large-scale equity selection.
Liu, Han; Mulvey, John; Zhao, Tianqi
2016-01-01
We propose a new stock selection strategy that exploits rebalancing returns and improves portfolio performance. To effectively harvest rebalancing gains, we apply ideas from elliptical-copula graphical modelling and stability inference to select stocks that are as independent as possible. The proposed elliptical-copula graphical model has a latent Gaussian representation; its structure can be effectively inferred using the regularized rank-based estimators. The resulting algorithm is computationally efficient and scales to large data-sets. To show the efficacy of the proposed method, we apply it to conduct equity selection based on a 16-year health care stock data-set and a large 34-year stock data-set. Empirical tests show that the proposed method is superior to alternative strategies including a principal component analysis-based approach and the classical Markowitz strategy based on the traditional buy-and-hold assumption.
[Graphic images on cigarette packages not effective].
Kok, Gerjo; Peters, Gjalt-Jorn Y; Ruiter, Robert A C
2013-01-01
The Dutch Government intends to make graphic images on cigarette packages mandatory. However, contrary to other policy measures to reduce smoking, health warnings do not work. There is no acceptable evidence in favour of graphic images and behaviour change theories suggest methods of change that improve skills, self-efficacy and social support. Thus, theory- and evidence-based policy should focus on prohibiting the tobacco industry from glamourizing packaging and make health communications on packages mandatory. As to the type of communications to be used, theory and evidence suggest that warning of the negative consequences of smoking is not an effective approach. Rather, targeting the most important determinants of the initiation of smoking and its successful cessation - such as skills, self-efficacy and subjective norm - along with the most effective behaviour change methods appears to be the most expedient strategy.
A graphical vector autoregressive modelling approach to the analysis of electronic diary data
2010-01-01
Background In recent years, electronic diaries are increasingly used in medical research and practice to investigate patients' processes and fluctuations in symptoms over time. To model dynamic dependence structures and feedback mechanisms between symptom-relevant variables, a multivariate time series method has to be applied. Methods We propose to analyse the temporal interrelationships among the variables by a structural modelling approach based on graphical vector autoregressive (VAR) models. We give a comprehensive description of the underlying concepts and explain how the dependence structure can be recovered from electronic diary data by a search over suitable constrained (graphical) VAR models. Results The graphical VAR approach is applied to the electronic diary data of 35 obese patients with and without binge eating disorder (BED). The dynamic relationships for the two subgroups between eating behaviour, depression, anxiety and eating control are visualized in two path diagrams. Results show that the two subgroups of obese patients with and without BED are distinguishable by the temporal patterns which influence their respective eating behaviours. Conclusion The use of the graphical VAR approach for the analysis of electronic diary data leads to a deeper insight into patient's dynamics and dependence structures. An increasing use of this modelling approach could lead to a better understanding of complex psychological and physiological mechanisms in different areas of medical care and research. PMID:20359333
Eminaga, Okyaz; Hinkelammert, Reemt; Semjonow, Axel; Neumann, Joerg; Abbas, Mahmoud; Koepke, Thomas; Bettendorf, Olaf; Eltze, Elke; Dugas, Martin
2010-11-15
The pathology report of radical prostatectomy specimens plays an important role in clinical decisions and the prognostic evaluation in Prostate Cancer (PCa). The anatomical schema is a helpful tool to document PCa extension for clinical and research purposes. To achieve electronic documentation and analysis, an appropriate documentation model for anatomical schemas is needed. For this purpose we developed cMDX. The document architecture of cMDX was designed according to Open Packaging Conventions by separating the whole data into template data and patient data. Analogue custom XML elements were considered to harmonize the graphical representation (e.g. tumour extension) with the textual data (e.g. histological patterns). The graphical documentation was based on the four-layer visualization model that forms the interaction between different custom XML elements. Sensible personal data were encrypted with a 256-bit cryptographic algorithm to avoid misuse. In order to assess the clinical value, we retrospectively analysed the tumour extension in 255 patients after radical prostatectomy. The pathology report with cMDX can represent pathological findings of the prostate in schematic styles. Such reports can be integrated into the hospital information system. "cMDX" documents can be converted into different data formats like text, graphics and PDF. Supplementary tools like cMDX Editor and an analyser tool were implemented. The graphical analysis of 255 prostatectomy specimens showed that PCa were mostly localized in the peripheral zone (Mean: 73% ± 25). 54% of PCa showed a multifocal growth pattern. cMDX can be used for routine histopathological reporting of radical prostatectomy specimens and provide data for scientific analysis.
Acceleration of GPU-based Krylov solvers via data transfer reduction
Anzt, Hartwig; Tomov, Stanimire; Luszczek, Piotr; ...
2015-04-08
Krylov subspace iterative solvers are often the method of choice when solving large sparse linear systems. At the same time, hardware accelerators such as graphics processing units continue to offer significant floating point performance gains for matrix and vector computations through easy-to-use libraries of computational kernels. However, as these libraries are usually composed of a well optimized but limited set of linear algebra operations, applications that use them often fail to reduce certain data communications, and hence fail to leverage the full potential of the accelerator. In this study, we target the acceleration of Krylov subspace iterative methods for graphicsmore » processing units, and in particular the Biconjugate Gradient Stabilized solver that significant improvement can be achieved by reformulating the method to reduce data-communications through application-specific kernels instead of using the generic BLAS kernels, e.g. as provided by NVIDIA’s cuBLAS library, and by designing a graphics processing unit specific sparse matrix-vector product kernel that is able to more efficiently use the graphics processing unit’s computing power. Furthermore, we derive a model estimating the performance improvement, and use experimental data to validate the expected runtime savings. Finally, considering that the derived implementation achieves significantly higher performance, we assert that similar optimizations addressing algorithm structure, as well as sparse matrix-vector, are crucial for the subsequent development of high-performance graphics processing units accelerated Krylov subspace iterative methods.« less
Proton radiography in three dimensions: A proof of principle of a new technique
DOE Office of Scientific and Technical Information (OSTI.GOV)
Raytchev, Milen; Seco, Joao
2013-10-15
Purpose: Monte Carlo simulations were used to investigate a range of phantom configurations to establish enabling three-dimensional proton radiographic techniques.Methods: A large parameter space of stacked phantom geometries composed of tissue inhomogeneity materials such as lung, bone, and cartilage inserted within water background were simulated using a purposefully modified version of TOPAS, an application running on top of the GEANT4 Monte Carlo code. The phantoms were grouped in two classes, one with the inhomogeneity inserted only half-way in the lateral direction and another with complete inhomogeneity insertion. The former class was used to calculate the track count and the energymore » fluence of the protons as they exit the phantoms either having traversed the inhomogeneity or not. The latter class was used to calculate one yield value accounting for loss of protons due to physical processes only and another yield value accounting for deliberately discarded protons due to large scattering angles. A graphical fingerprinting method was developed to determine the inhomogeneity thickness and location within the phantom based on track count and energy fluence information. Two additional yield values extended this method to the general case which also determines the inhomogeneity material and the phantom thickness.Results: The graphical fingerprinting method was manually validated for two, and automatically tested for all, tissue materials using an exhaustive set of inhomogeneity geometries for 16 cm thick phantoms. Unique recognition of test phantom configurations was achieved in the large majority of cases. The method in the general case was further tested using an exhaustive set of inhomogeneity and phantom tissues and geometries where the phantom thicknesses ranged between 8 and 24 cm. Unique recognition of the test phantom configurations was achieved only for part of the phantom parameter space. The correlations between the remaining false positive recognitions were analyzed.Conclusions: The concept of 3D proton radiography for tissue inhomogeneities of simple geometries was established with the current work. In contrast to conventional 2D proton radiography, the main objective of the demonstrated 3D technique is not proton range. Rather, it is to measure the depth and thickness of an inhomogeneity located in an imaged geometry. Further work is needed to extend and apply the method to more complex geometries.« less
Witte, Anna Kristina; Fister, Susanne; Mester, Patrick; Schoder, Dagmar; Rossmanith, Peter
2016-11-01
Fast and reliable pathogen detection is an important issue for human health. Since conventional microbiological methods are rather slow, there is growing interest in detection and quantification using molecular methods. The droplet digital polymerase chain reaction (ddPCR) is a relatively new PCR method for absolute and accurate quantification without external standards. Using the Listeria monocytogenes specific prfA assay, we focused on the questions of whether the assay was directly transferable to ddPCR and whether ddPCR was suitable for samples derived from heterogeneous matrices, such as foodstuffs that often included inhibitors and a non-target bacterial background flora. Although the prfA assay showed suboptimal cluster formation, use of ddPCR for quantification of L. monocytogenes from pure bacterial cultures, artificially contaminated cheese, and naturally contaminated foodstuff was satisfactory over a relatively broad dynamic range. Moreover, results demonstrated the outstanding detection limit of one copy. However, while poorer DNA quality, such as resulting from longer storage, can impair ddPCR, internal amplification control (IAC) of prfA by ddPCR, that is integrated in the genome of L. monocytogenes ΔprfA, showed even slightly better quantification over a broader dynamic range. Graphical Abstract Evaluating the absolute quantification potential of ddPCR targeting Listeria monocytogenes prfA.
Managing Heuristics as a Method of Inquiry in Autobiographical Graphic Design Theses
ERIC Educational Resources Information Center
Ings, Welby
2011-01-01
This article draws on case studies undertaken in postgraduate research at AUT University, Auckland. It seeks to address a number of issues related to heuristic inquiries employed by graphic design students who use autobiographical approaches when developing research-based theses. For this type of thesis, heuristics as a system of inquiry may…
ERIC Educational Resources Information Center
Beal, Carole R.; Rosenblum, L. Penny
2015-01-01
Introduction: The present study evaluated the feasibility of using an iPad application or "app" for algebra-readiness mathematics, with accompanying braille materials and accessible graphics, when used in authentic educational settings. Methods: Twenty-nine students with visual impairments in grades 4-11 used the materials under the…
Factors Influencing the Effectiveness of Note Taking on Computer-Based Graphic Organizers
ERIC Educational Resources Information Center
Crooks, Steven M.; White, David R.; Barnard, Lucy
2007-01-01
Previous research on graphic organizer (GO) note taking has shown that this method is most effective when the GO is presented to the student partially complete with provided notes. This study extended prior research by investigating the effects of provided note type (summary vs. verbatim) and GO bite size (large vs. small) on the transfer…
Expected Utility Illustrated: A Graphical Analysis of Gambles with More than Two Possible Outcomes
ERIC Educational Resources Information Center
Chen, Frederick H.
2010-01-01
The author presents a simple geometric method to graphically illustrate the expected utility from a gamble with more than two possible outcomes. This geometric result gives economics students a simple visual aid for studying expected utility theory and enables them to analyze a richer set of decision problems under uncertainty compared to what…
ERIC Educational Resources Information Center
Doyle, Carole S.
This study examined the effectiveness of two approaches to enhancing the reading comprehension of learning disabled students in the social studies content area. An approach using the graphic organizer in the form of visual displays was compared to the traditional method in which students were presented content through lecture, text, and linear…
Color-Coded Graphic Organizers for Teaching Writing to Students with Learning Disabilities
ERIC Educational Resources Information Center
Ewoldt, Kathy B.; Morgan, Joseph John
2017-01-01
A commonly used method for supporting the writing of students with learning disabilities (LD), graphic organizers have been shown to effectively support instruction for students with LD in a variety of content areas (Dexter & Hughes, 2011). Students with LD often struggle with the process of developing their ideas into organized sentences; the…
Using a Computerised Graphics Package to Achieve a Technology-Oriented Classroom
ERIC Educational Resources Information Center
Aladejana, Francisca; Idowu, Lanre
2009-01-01
The present situation in Nigeria involves students of fine arts, a practical-oriented subject, being exposed to poor methods of teaching with consequent poor performances. This study examined the extent to which the use of a computerised graphics package could make the classroom technology-oriented and affect the performance of learners. This is…
Yue, Chen; Chen, Shaojie; Sair, Haris I; Airan, Raag; Caffo, Brian S
2015-09-01
Data reproducibility is a critical issue in all scientific experiments. In this manuscript, the problem of quantifying the reproducibility of graphical measurements is considered. The image intra-class correlation coefficient (I2C2) is generalized and the graphical intra-class correlation coefficient (GICC) is proposed for such purpose. The concept for GICC is based on multivariate probit-linear mixed effect models. A Markov Chain Monte Carlo EM (mcm-cEM) algorithm is used for estimating the GICC. Simulation results with varied settings are demonstrated and our method is applied to the KIRBY21 test-retest dataset.
The graphic cell method: a new look at digitizing geologic maps
Hanley, J.T.
1982-01-01
The graphic cell method is an alternative method of digitizing areal geologic information. It involves a discrete-point sampling scheme in which the computer establishes a matrix of cells over the map. Each cell and the whole cell is assigned the identity or value of the geologic information that is recognized at its center. Cell size may be changed to suit the needs of the user. The computer program resolves the matrix and identifies potential errors such as multiple assignments. Input includes the digitized boundaries of each geologic formation. This method should eliminate a primary bottleneck in the creation and testing of geomathematical models in such disciplines as resource appraisal. ?? 1982.
NASA Technical Reports Server (NTRS)
Jones, Robert T
1937-01-01
A simplified treatment of the application of Heaviside's operational methods to problems of airplane dynamics is given. Certain graphical methods and logarithmic formulas that lessen the amount of computation involved are explained. The problem representing a gust disturbance or control manipulation is taken up and it is pointed out that in certain cases arbitrary control manipulations may be dealt with as though they imposed specific constraints on the airplane, thus avoiding the necessity of any integration. The application of the calculations described in the text is illustrated by several examples chosen to show the use of the methods and the practicability of the graphical and logarithmic computations described.
Rapid indirect trajectory optimization on highly parallel computing architectures
NASA Astrophysics Data System (ADS)
Antony, Thomas
Trajectory optimization is a field which can benefit greatly from the advantages offered by parallel computing. The current state-of-the-art in trajectory optimization focuses on the use of direct optimization methods, such as the pseudo-spectral method. These methods are favored due to their ease of implementation and large convergence regions while indirect methods have largely been ignored in the literature in the past decade except for specific applications in astrodynamics. It has been shown that the shortcomings conventionally associated with indirect methods can be overcome by the use of a continuation method in which complex trajectory solutions are obtained by solving a sequence of progressively difficult optimization problems. High performance computing hardware is trending towards more parallel architectures as opposed to powerful single-core processors. Graphics Processing Units (GPU), which were originally developed for 3D graphics rendering have gained popularity in the past decade as high-performance, programmable parallel processors. The Compute Unified Device Architecture (CUDA) framework, a parallel computing architecture and programming model developed by NVIDIA, is one of the most widely used platforms in GPU computing. GPUs have been applied to a wide range of fields that require the solution of complex, computationally demanding problems. A GPU-accelerated indirect trajectory optimization methodology which uses the multiple shooting method and continuation is developed using the CUDA platform. The various algorithmic optimizations used to exploit the parallelism inherent in the indirect shooting method are described. The resulting rapid optimal control framework enables the construction of high quality optimal trajectories that satisfy problem-specific constraints and fully satisfy the necessary conditions of optimality. The benefits of the framework are highlighted by construction of maximum terminal velocity trajectories for a hypothetical long range weapon system. The techniques used to construct an initial guess from an analytic near-ballistic trajectory and the methods used to formulate the necessary conditions of optimality in a manner that is transparent to the designer are discussed. Various hypothetical mission scenarios that enforce different combinations of initial, terminal, interior point and path constraints demonstrate the rapid construction of complex trajectories without requiring any a-priori insight into the structure of the solutions. Trajectory problems of this kind were previously considered impractical to solve using indirect methods. The performance of the GPU-accelerated solver is found to be 2x--4x faster than MATLAB's bvp4c, even while running on GPU hardware that is five years behind the state-of-the-art.
GPU-based cloud service for Smith-Waterman algorithm using frequency distance filtration scheme.
Lee, Sheng-Ta; Lin, Chun-Yuan; Hung, Che Lun
2013-01-01
As the conventional means of analyzing the similarity between a query sequence and database sequences, the Smith-Waterman algorithm is feasible for a database search owing to its high sensitivity. However, this algorithm is still quite time consuming. CUDA programming can improve computations efficiently by using the computational power of massive computing hardware as graphics processing units (GPUs). This work presents a novel Smith-Waterman algorithm with a frequency-based filtration method on GPUs rather than merely accelerating the comparisons yet expending computational resources to handle such unnecessary comparisons. A user friendly interface is also designed for potential cloud server applications with GPUs. Additionally, two data sets, H1N1 protein sequences (query sequence set) and human protein database (database set), are selected, followed by a comparison of CUDA-SW and CUDA-SW with the filtration method, referred to herein as CUDA-SWf. Experimental results indicate that reducing unnecessary sequence alignments can improve the computational time by up to 41%. Importantly, by using CUDA-SWf as a cloud service, this application can be accessed from any computing environment of a device with an Internet connection without time constraints.
Nanoparticle assisted laser desorption/ionization mass spectrometry for small molecule analytes.
Abdelhamid, Hani Nasser
2018-03-01
Nanoparticle assisted laser desorption/ionization mass spectrometry (NPs-ALDI-MS) shows remarkable characteristics and has a promising future in terms of real sample analysis. The incorporation of NPs can advance several methods including surface assisted LDI-MS, and surface enhanced LDI-MS. These methods have advanced the detection of many thermally labile and nonvolatile biomolecules. Nanoparticles circumvent the drawbacks of conventional organic matrices for the analysis of small molecules. In most cases, NPs offer a clear background without interfering peaks, absence of fragmentation of thermally labile molecules, and allow the ionization of species with weak noncovalent interactions. Furthermore, an enhancement in sensitivity and selectivity can be achieved. NPs enable straightforward analysis of target species in a complex sample. This review (with 239 refs.) covers the progress made in laser-based mass spectrometry in combination with the use of metallic NPs (such as AuNPs, AgNPs, PtNPs, and PdNPs), NPs consisting of oxides and chalcogenides, silicon-based NPs, carbon-based nanomaterials, quantum dots, and metal-organic frameworks. Graphical abstract An overview is given on nanomaterials for use in surface-assisted laser desorption/ionization mass spectrometry of small molecules.
Adsorptive amputation of hazardous azo dye Congo red from wastewater: a critical review.
Raval, Nirav P; Shah, Prapti U; Shah, Nisha K
2016-08-01
Increasing amount of dyes in an ecosystem has propelled the search of various methods for dye removal. Amongst all the methods, adsorption occupies a prominent place in dye removal. Keeping this in mind, many adsorbents used for the removal of hazardous anionic azo dye Congo red (CR) from aqueous medium were reviewed by the authors. The main objectives behind this review article are to assemble the information on scattered adsorbents and enlighten the wide range of potentially effective adsorbents for CR removal. Thus, CR sorption by various adsorbents such as activated carbon, non-conventional low-cost materials, nanomaterials, composites and nanocomposites are surveyed and critically reviewed as well as their sorption capacities are also compared. This review also explores the grey areas of the adsorption performance of various adsorbents with reference to the effects of pH, contact time, initial dye concentration and adsorbent dosage. The equilibrium adsorption isotherm, kinetic and thermodynamic data of different adsorbents used for CR removal were also analysed. It is evident from a literature survey of more than 290 published papers that nanoparticle and nanocomposite adsorbents have demonstrated outstanding adsorption capabilities for CR. Graphical abstract ᅟ.
Wang, Liping; Duan, Haotian; Jiang, Jiebing; Long, Jiakun; Yu, Yingjia; Chen, Guiliang; Duan, Gengli
2017-09-01
A new, simple, and fast infrared-assisted self enzymolysis extraction (IRASEE) approach for the extraction of total flavonoid aglycones (TFA) mainly including baicalein, wogonin, and oroxylin A from Scutellariae Radix is presented to enhance extraction yield. Extraction enzymolysis temperature, enzymolysis liquid-to-solid ratio, enzymolysis pH, enzymolysis time and infrared power, the factors affecting IRASEE procedure, were investigated in a newly designed, temperature-controlled infrared-assisted extraction (TC-IRAE) system to acquire the optimum analysis conditions. The results illustrated that IRASEE possessed great advantages in terms of efficiency and time compared with other conventional extraction techniques. Furthermore, the mechanism of IRASEE was preliminarily explored by observing the microscopic change of the samples surface structures, studying the main chemical compositions change of the samples before and after extraction and investigating the kinetics and thermodynamics at three temperature levels during the IRASEE process. These findings revealed that IRASEE can destroy the surface microstructures to accelerate the mass transfer and reduce the activation energy to intensify the chemical process. This integrative study presents a simple, rapid, efficient, and environmental IRASEE method for TFA extraction which has promising prospects for other similar herbal medicines. Graphical Abstract ᅟ.
High Performance Molecular Visualization: In-Situ and Parallel Rendering with EGL.
Stone, John E; Messmer, Peter; Sisneros, Robert; Schulten, Klaus
2016-05-01
Large scale molecular dynamics simulations produce terabytes of data that is impractical to transfer to remote facilities. It is therefore necessary to perform visualization tasks in-situ as the data are generated, or by running interactive remote visualization sessions and batch analyses co-located with direct access to high performance storage systems. A significant challenge for deploying visualization software within clouds, clusters, and supercomputers involves the operating system software required to initialize and manage graphics acceleration hardware. Recently, it has become possible for applications to use the Embedded-system Graphics Library (EGL) to eliminate the requirement for windowing system software on compute nodes, thereby eliminating a significant obstacle to broader use of high performance visualization applications. We outline the potential benefits of this approach in the context of visualization applications used in the cloud, on commodity clusters, and supercomputers. We discuss the implementation of EGL support in VMD, a widely used molecular visualization application, and we outline benefits of the approach for molecular visualization tasks on petascale computers, clouds, and remote visualization servers. We then provide a brief evaluation of the use of EGL in VMD, with tests using developmental graphics drivers on conventional workstations and on Amazon EC2 G2 GPU-accelerated cloud instance types. We expect that the techniques described here will be of broad benefit to many other visualization applications.
High Performance Molecular Visualization: In-Situ and Parallel Rendering with EGL
Stone, John E.; Messmer, Peter; Sisneros, Robert; Schulten, Klaus
2016-01-01
Large scale molecular dynamics simulations produce terabytes of data that is impractical to transfer to remote facilities. It is therefore necessary to perform visualization tasks in-situ as the data are generated, or by running interactive remote visualization sessions and batch analyses co-located with direct access to high performance storage systems. A significant challenge for deploying visualization software within clouds, clusters, and supercomputers involves the operating system software required to initialize and manage graphics acceleration hardware. Recently, it has become possible for applications to use the Embedded-system Graphics Library (EGL) to eliminate the requirement for windowing system software on compute nodes, thereby eliminating a significant obstacle to broader use of high performance visualization applications. We outline the potential benefits of this approach in the context of visualization applications used in the cloud, on commodity clusters, and supercomputers. We discuss the implementation of EGL support in VMD, a widely used molecular visualization application, and we outline benefits of the approach for molecular visualization tasks on petascale computers, clouds, and remote visualization servers. We then provide a brief evaluation of the use of EGL in VMD, with tests using developmental graphics drivers on conventional workstations and on Amazon EC2 G2 GPU-accelerated cloud instance types. We expect that the techniques described here will be of broad benefit to many other visualization applications. PMID:27747137
ELATE: an open-source online application for analysis and visualization of elastic tensors
NASA Astrophysics Data System (ADS)
Gaillac, Romain; Pullumbi, Pluton; Coudert, François-Xavier
2016-07-01
We report on the implementation of a tool for the analysis of second-order elastic stiffness tensors, provided with both an open-source Python module and a standalone online application allowing the visualization of anisotropic mechanical properties. After describing the software features, how we compute the conventional elastic constants and how we represent them graphically, we explain our technical choices for the implementation. In particular, we focus on why a Python module is used to generate the HTML web page with embedded Javascript for dynamical plots.
NASA Technical Reports Server (NTRS)
Kindall, S. M.
1980-01-01
The computer code for the trajectory processor (#TRAJ) of the high fidelity relative motion program is described. The #TRAJ processor is a 12-degrees-of-freedom trajectory integrator (6 degrees of freedom for each of two vehicles) which can be used to generate digital and graphical data describing the relative motion of the Space Shuttle Orbiter and a free-flying cylindrical payload. A listing of the code, coding standards and conventions, detailed flow charts, and discussions of the computational logic are included.
Occlusal force discrimination by denture patients.
Pacer, R J; Bowman, D C
1975-06-01
A study was conducted on subjects with conventional dentures and with overlay dentures to compare their abilities to discriminate between occlusal forces. Perpendicular forces were applied to the dynamic center of the occlusal table of the mandibular denture. Each subject's ability to distinguish differences in values of force was observed and recorded. All subjects with dentures showed sensory threshold values close to those reported for natural teeth. A graphic plotting showed that the responses of subjects with overlay-type dentures were more closely correlated with the psychophysical law as expressed by Stevens as a power function. Since this phenomenon holds true for natural teeth, the overlay denture more closely resembles natural teeth in this type of sensory function than does the conventional denture. In addition to recognized advantages, such as preservation of the ridge and improved retention and stability, the overlay denture provides more typical sensory function than is provided by the conventional denture. This advantage should further motivate dentists and patients to consider the retention and utilization of at least two suitable teeth in an overlay-type denture service.
Graphical methods for the sensitivity analysis in discriminant analysis
Kim, Youngil; Anderson-Cook, Christine M.; Dae-Heung, Jang
2015-09-30
Similar to regression, many measures to detect influential data points in discriminant analysis have been developed. Many follow similar principles as the diagnostic measures used in linear regression in the context of discriminant analysis. Here we focus on the impact on the predicted classification posterior probability when a data point is omitted. The new method is intuitive and easily interpretative compared to existing methods. We also propose a graphical display to show the individual movement of the posterior probability of other data points when a specific data point is omitted. This enables the summaries to capture the overall pattern ofmore » the change.« less
Application of interactive computer graphics in wind-tunnel dynamic model testing
NASA Technical Reports Server (NTRS)
Doggett, R. V., Jr.; Hammond, C. E.
1975-01-01
The computer-controlled data-acquisition system recently installed for use with a transonic dynamics tunnel was described. This includes a discussion of the hardware/software features of the system. A subcritical response damping technique, called the combined randomdec/moving-block method, for use in windtunnel-model flutter testing, that has been implemented on the data-acquisition system, is described in some detail. Some results using the method are presented and the importance of using interactive graphics in applying the technique in near real time during wind-tunnel test operations is discussed.
Analysis of Flexural Fatigue Strength of Self Compacting Fibre Reinforced Concrete Beams
NASA Astrophysics Data System (ADS)
Murali, G.; Sudar Celestina, J. P. Arul; Subhashini, N.; Vigneshwari, M.
2017-07-01
This study presents the extensive statistical investigation ofvariations in flexural fatigue life of self-compacting Fibrous Concrete (FC) beams. For this purpose, the experimental data of earlier researchers were examined by two parameter Weibull distribution.Two methods namely Graphical and moment wereused to analyse the variations in experimental data and the results have been presented in the form of probability of survival. The Weibull parameters values obtained from graphical and method of moments are precise. At 0.7 stress level, the fatigue life shows 59861 cyclesfor areliability of 90%.
NASA Technical Reports Server (NTRS)
Bratanow, T.; Ecer, A.
1973-01-01
A general computational method for analyzing unsteady flow around pitching and plunging airfoils was developed. The finite element method was applied in developing an efficient numerical procedure for the solution of equations describing the flow around airfoils. The numerical results were employed in conjunction with computer graphics techniques to produce visualization of the flow. The investigation involved mathematical model studies of flow in two phases: (1) analysis of a potential flow formulation and (2) analysis of an incompressible, unsteady, viscous flow from Navier-Stokes equations.
NASA Astrophysics Data System (ADS)
Kang, Sungil; Roh, Annah; Nam, Bodam; Hong, Hyunki
2011-12-01
This paper presents a novel vision system for people detection using an omnidirectional camera mounted on a mobile robot. In order to determine regions of interest (ROI), we compute a dense optical flow map using graphics processing units, which enable us to examine compliance with the ego-motion of the robot in a dynamic environment. Shape-based classification algorithms are employed to sort ROIs into human beings and nonhumans. The experimental results show that the proposed system detects people more precisely than previous methods.
TU-D-209-03: Alignment of the Patient Graphic Model Using Fluoroscopic Images for Skin Dose Mapping
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oines, A; Oines, A; Kilian-Meneghin, J
2016-06-15
Purpose: The Dose Tracking System (DTS) was developed to provide realtime feedback of skin dose and dose rate during interventional fluoroscopic procedures. A color map on a 3D graphic of the patient represents the cumulative dose distribution on the skin. Automated image correlation algorithms are described which use the fluoroscopic procedure images to align and scale the patient graphic for more accurate dose mapping. Methods: Currently, the DTS employs manual patient graphic selection and alignment. To improve the accuracy of dose mapping and automate the software, various methods are explored to extract information about the beam location and patient morphologymore » from the procedure images. To match patient anatomy with a reference projection image, preprocessing is first used, including edge enhancement, edge detection, and contour detection. Template matching algorithms from OpenCV are then employed to find the location of the beam. Once a match is found, the reference graphic is scaled and rotated to fit the patient, using image registration correlation functions in Matlab. The algorithm runs correlation functions for all points and maps all correlation confidences to a surface map. The highest point of correlation is used for alignment and scaling. The transformation data is saved for later model scaling. Results: Anatomic recognition is used to find matching features between model and image and image registration correlation provides for alignment and scaling at any rotation angle with less than onesecond runtime, and at noise levels in excess of 150% of those found in normal procedures. Conclusion: The algorithm provides the necessary scaling and alignment tools to improve the accuracy of dose distribution mapping on the patient graphic with the DTS. Partial support from NIH Grant R01-EB002873 and Toshiba Medical Systems Corp.« less
Ahrens, T S; Schallom, L
2001-01-01
Techniques to measure pulmonary artery (PA) pressure waveforms include digital measurement, graphic measurement, and freeze-cursor measurement. Previous studies reported the inaccuracy of digital and freeze-cursor measurements. However, many of the previous studies were small and did not thoroughly examine the circumstances of when digital measurements might be inaccurate. To compare digital measurements and graphic measurements of PA and central venous pressure (CVP) waveforms in patients with a variety of respiratory patterns, and to compare digital measurements and graphic measurements of CVPs in patients with abnormal or right ventricular waveforms. A total of 928 patients were enrolled in this study. Waveforms from the PA and CVP were collected from each patient. The monitor pressure value (digital measurement) printed on the recorded waveform was compared with the pressure value obtained by a graphic strip recording and measured by one of the primary investigators (graphic measurement). Digital measurements were found to be inaccurate in measuring waveforms in all respiratory categories and in measuring right ventricular waveforms. PA diastolic values and CVP values were the most inaccurately measured waveforms. Digital errors of more than 4 mm Hg were common. There were instances in which the monitor's digital measurement was substantially different from the graphically measured value. This difference has the potential to mislead interpretation of clinical situations. The monitor's ability to occasionally give digital measurement values similar to the graphic measurements may lead to a false sense of security in clinicians. Because the accuracy of the monitor is inconsistent, the bedside clinician should interpret waveforms through use of a graphic recording rather than rely on the digital measurement on the monitor.
2014-01-01
Introduction: In 2011, the Food and Drug Administration published a final rule requiring cigarette packages and advertisements to include graphic health warning labels (HWLs) with new warning statements. Implementation of this rule has been stalled by legal challenge. This study assessed correlates of smoking-related intentions related to graphic HWLs among current cigarette smokers and nonsmokers in a national sample of U.S. young adults aged 18–34. Methods: Data were collected from 4,236 participants aged 18–34 using an online panel in January 2012 for the Legacy Young Adult Cohort Study. Analyses were weighted to provide nationally representative estimates. Our main outcome was assessed with a single item: “Do you think that new warning labels with graphic pictures would make you think about not smoking?” Results: Twenty-two percent of the young adults were current cigarette smokers. Fifty-three percent endorsed that new graphic HWLs would make them think about not smoking (40% among current smokers compared with 56% among nonsmokers). Among nonsmokers, those aged 18–24, females, Hispanics, and those who were aware of graphic cigarette HWLs were more likely to report intention to not smoke related to graphic HWLs. Among current smokers, intending to quit within the next 6 months was correlated with intention resulting from graphic HWLs. Hispanic ethnicity and intention to quit within 30 days were strong correlates of intention in light, nondaily, and self-identified social/occasional smokers. Conclusions: This study supports previous findings that graphic HWLs play an important role in preventing smoking, in addition to encouraging cessation in young adults. PMID:24212476
Reactions to FDA-Proposed Graphic Warning Labels Affixed to U.S. Smokers’ Cigarette Packs
Kreuter, Matthew W.; Boyum, Sonia; Thompson, Vetta S.; Caburnay, Charlene A.; Waters, Erika A.; Kaphingst, Kimberly A.; Rath, Suchitra; Fu, Qiang
2015-01-01
Introduction: Graphic warning labels have been shown to be more effective than text-only labels in increasing attention and perceived health risks, but most U.S. studies have involved single exposures in laboratory or Internet settings. Methods: We recruited a convenience sample (N = 202) of U.S. adult smokers from population subgroups with higher rates of smoking and smoking-related deaths who had participated in a larger survey about graphic warning labels. Participants were randomized to get 1 of 9 graphic + text labels or a text-only label. Research staff affixed a warning label sticker to participants’ cigarette pack(s) at enrollment. Color graphic labels covered slightly more than the lower half of packs. Black and white labels of current U.S. text-only warnings covered the existing side warning to prompt attention to the label (i.e., attention control). Participants received extra stickers of the same label for subsequent packs, and completed 3 telephone interviews in 1 week. Results: Participants reported low avoidance (<34%) and consistent use of the stickers (91%). Smokers consistently paid more attention to graphic than text-only labels. Only 5 of the 9 graphic warning labels were significantly associated with greater thoughts of health risks. Thinking about quitting and stopping smoking did not differ by label. Qualitative data illustrated differences in the “stickiness,” self-referencing, and counterarguments of graphic warning labels. Conclusions: U.S. smokers’ reactions to graphic warning labels on their own packs were similar to other, more controlled studies. Qualitative findings underscore the need for warning labels that encourage self-referential processing without increasing defensive reactions. PMID:25589676
Yeung, Ka Yee
2016-01-01
Reproducibility is vital in science. For complex computational methods, it is often necessary, not just to recreate the code, but also the software and hardware environment to reproduce results. Virtual machines, and container software such as Docker, make it possible to reproduce the exact environment regardless of the underlying hardware and operating system. However, workflows that use Graphical User Interfaces (GUIs) remain difficult to replicate on different host systems as there is no high level graphical software layer common to all platforms. GUIdock allows for the facile distribution of a systems biology application along with its graphics environment. Complex graphics based workflows, ubiquitous in systems biology, can now be easily exported and reproduced on many different platforms. GUIdock uses Docker, an open source project that provides a container with only the absolutely necessary software dependencies and configures a common X Windows (X11) graphic interface on Linux, Macintosh and Windows platforms. As proof of concept, we present a Docker package that contains a Bioconductor application written in R and C++ called networkBMA for gene network inference. Our package also includes Cytoscape, a java-based platform with a graphical user interface for visualizing and analyzing gene networks, and the CyNetworkBMA app, a Cytoscape app that allows the use of networkBMA via the user-friendly Cytoscape interface. PMID:27045593
Hung, Ling-Hong; Kristiyanto, Daniel; Lee, Sung Bong; Yeung, Ka Yee
2016-01-01
Reproducibility is vital in science. For complex computational methods, it is often necessary, not just to recreate the code, but also the software and hardware environment to reproduce results. Virtual machines, and container software such as Docker, make it possible to reproduce the exact environment regardless of the underlying hardware and operating system. However, workflows that use Graphical User Interfaces (GUIs) remain difficult to replicate on different host systems as there is no high level graphical software layer common to all platforms. GUIdock allows for the facile distribution of a systems biology application along with its graphics environment. Complex graphics based workflows, ubiquitous in systems biology, can now be easily exported and reproduced on many different platforms. GUIdock uses Docker, an open source project that provides a container with only the absolutely necessary software dependencies and configures a common X Windows (X11) graphic interface on Linux, Macintosh and Windows platforms. As proof of concept, we present a Docker package that contains a Bioconductor application written in R and C++ called networkBMA for gene network inference. Our package also includes Cytoscape, a java-based platform with a graphical user interface for visualizing and analyzing gene networks, and the CyNetworkBMA app, a Cytoscape app that allows the use of networkBMA via the user-friendly Cytoscape interface.
ROCOPT: A user friendly interactive code to optimize rocket structural components
NASA Technical Reports Server (NTRS)
Rule, William K.
1989-01-01
ROCOPT is a user-friendly, graphically-interfaced, microcomputer-based computer program (IBM compatible) that optimizes rocket components by minimizing the structural weight. The rocket components considered are ring stiffened truncated cones and cylinders. The applied loading is static, and can consist of any combination of internal or external pressure, axial force, bending moment, and torque. Stress margins are calculated by means of simple closed form strength of material type equations. Stability margins are determined by approximate, orthotropic-shell, closed-form equations. A modified form of Powell's method, in conjunction with a modified form of the external penalty method, is used to determine the minimum weight of the structure subject to stress and stability margin constraints, as well as user input constraints on the structural dimensions. The graphical interface guides the user through the required data prompts, explains program options and graphically displays results for easy interpretation.
Detection of text strings from mixed text/graphics images
NASA Astrophysics Data System (ADS)
Tsai, Chien-Hua; Papachristou, Christos A.
2000-12-01
A robust system for text strings separation from mixed text/graphics images is presented. Based on a union-find (region growing) strategy the algorithm is thus able to classify the text from graphics and adapts to changes in document type, language category (e.g., English, Chinese and Japanese), text font style and size, and text string orientation within digital images. In addition, it allows for a document skew that usually occurs in documents, without skew correction prior to discrimination while these proposed methods such a projection profile or run length coding are not always suitable for the condition. The method has been tested with a variety of printed documents from different origins with one common set of parameters, and the experimental results of the performance of the algorithm in terms of computational efficiency are demonstrated by using several tested images from the evaluation.
A semiparametric graphical modelling approach for large-scale equity selection
Liu, Han; Mulvey, John; Zhao, Tianqi
2016-01-01
We propose a new stock selection strategy that exploits rebalancing returns and improves portfolio performance. To effectively harvest rebalancing gains, we apply ideas from elliptical-copula graphical modelling and stability inference to select stocks that are as independent as possible. The proposed elliptical-copula graphical model has a latent Gaussian representation; its structure can be effectively inferred using the regularized rank-based estimators. The resulting algorithm is computationally efficient and scales to large data-sets. To show the efficacy of the proposed method, we apply it to conduct equity selection based on a 16-year health care stock data-set and a large 34-year stock data-set. Empirical tests show that the proposed method is superior to alternative strategies including a principal component analysis-based approach and the classical Markowitz strategy based on the traditional buy-and-hold assumption. PMID:28316507
ERIC Educational Resources Information Center
Lu, Hui-Ping; Chen, Jun-Hong; Lee, Chang-Franw
2016-01-01
Inspiration is the primary element of good design. Designers, however, also risk not being able to find inspiration. Novice designers commonly find themselves to be depressed during the conceptual design phase when they fail to find inspiration and the information to be creative. Accordingly, under the graphic design parameter, we have developed…
Graphics processing unit-assisted lossless decompression
Loughry, Thomas A.
2016-04-12
Systems and methods for decompressing compressed data that has been compressed by way of a lossless compression algorithm are described herein. In a general embodiment, a graphics processing unit (GPU) is programmed to receive compressed data packets and decompress such packets in parallel. The compressed data packets are compressed representations of an image, and the lossless compression algorithm is a Rice compression algorithm.
ERIC Educational Resources Information Center
Ivanov, Anisoara; Neacsu, Andrei
2011-01-01
This study describes the possibility and advantages of utilizing simple computer codes to complement the teaching techniques for high school physics. The authors have begun working on a collection of open source programs which allow students to compare the results and graphics from classroom exercises with the correct solutions and further more to…
Moving beyond the Bar Plot and the Line Graph to Create Informative and Attractive Graphics
ERIC Educational Resources Information Center
Larson-Hall, Jenifer
2017-01-01
Graphics are often mistaken for a mere frill in the methodological arsenal of data analysis when in fact they can be one of the simplest and at the same time most powerful methods of communicating statistical information (Tufte, 2001). The first section of the article argues for the statistical necessity of graphs, echoing and amplifying similar…
[Graphic reconstruction of anatomic surfaces].
Ciobanu, O
2004-01-01
The paper deals with the graphic reconstruction of anatomic surfaces in a virtual 3D setting. Scanning technologies and soft provides a greater flexibility in the digitization of surfaces and a higher resolution and accuracy. An alternative cheap method for the reconstruction of 3D anatomic surfaces is presented in connection with some studies and international projects developed by Medical Design research team.
ERIC Educational Resources Information Center
Goodson-Espy, Tracy; Lynch-Davis, Kathleen; Schram, Pamela; Quickenton, Art
2010-01-01
This paper describes the genesis and purpose of our geometry methods course, focusing on a geometry-teaching technology we created using NVIDIA[R] Chameleon demonstration. This article presents examples from a sequence of lessons centered about a 3D computer graphics demonstration of the chameleon and its geometry. In addition, we present data…
Monte Carlo MP2 on Many Graphical Processing Units.
Doran, Alexander E; Hirata, So
2016-10-11
In the Monte Carlo second-order many-body perturbation (MC-MP2) method, the long sum-of-product matrix expression of the MP2 energy, whose literal evaluation may be poorly scalable, is recast into a single high-dimensional integral of functions of electron pair coordinates, which is evaluated by the scalable method of Monte Carlo integration. The sampling efficiency is further accelerated by the redundant-walker algorithm, which allows a maximal reuse of electron pairs. Here, a multitude of graphical processing units (GPUs) offers a uniquely ideal platform to expose multilevel parallelism: fine-grain data-parallelism for the redundant-walker algorithm in which millions of threads compute and share orbital amplitudes on each GPU; coarse-grain instruction-parallelism for near-independent Monte Carlo integrations on many GPUs with few and infrequent interprocessor communications. While the efficiency boost by the redundant-walker algorithm on central processing units (CPUs) grows linearly with the number of electron pairs and tends to saturate when the latter exceeds the number of orbitals, on a GPU it grows quadratically before it increases linearly and then eventually saturates at a much larger number of pairs. This is because the orbital constructions are nearly perfectly parallelized on a GPU and thus completed in a near-constant time regardless of the number of pairs. In consequence, an MC-MP2/cc-pVDZ calculation of a benzene dimer is 2700 times faster on 256 GPUs (using 2048 electron pairs) than on two CPUs, each with 8 cores (which can use only up to 256 pairs effectively). We also numerically determine that the cost to achieve a given relative statistical uncertainty in an MC-MP2 energy increases as O(n 3 ) or better with system size n, which may be compared with the O(n 5 ) scaling of the conventional implementation of deterministic MP2. We thus establish the scalability of MC-MP2 with both system and computer sizes.
O'Callaghan, Sean; De Souza, David P; Isaac, Andrew; Wang, Qiao; Hodkinson, Luke; Olshansky, Moshe; Erwin, Tim; Appelbe, Bill; Tull, Dedreia L; Roessner, Ute; Bacic, Antony; McConville, Malcolm J; Likić, Vladimir A
2012-05-30
Gas chromatography-mass spectrometry (GC-MS) is a technique frequently used in targeted and non-targeted measurements of metabolites. Most existing software tools for processing of raw instrument GC-MS data tightly integrate data processing methods with graphical user interface facilitating interactive data processing. While interactive processing remains critically important in GC-MS applications, high-throughput studies increasingly dictate the need for command line tools, suitable for scripting of high-throughput, customized processing pipelines. PyMS comprises a library of functions for processing of instrument GC-MS data developed in Python. PyMS currently provides a complete set of GC-MS processing functions, including reading of standard data formats (ANDI- MS/NetCDF and JCAMP-DX), noise smoothing, baseline correction, peak detection, peak deconvolution, peak integration, and peak alignment by dynamic programming. A novel common ion single quantitation algorithm allows automated, accurate quantitation of GC-MS electron impact (EI) fragmentation spectra when a large number of experiments are being analyzed. PyMS implements parallel processing for by-row and by-column data processing tasks based on Message Passing Interface (MPI), allowing processing to scale on multiple CPUs in distributed computing environments. A set of specifically designed experiments was performed in-house and used to comparatively evaluate the performance of PyMS and three widely used software packages for GC-MS data processing (AMDIS, AnalyzerPro, and XCMS). PyMS is a novel software package for the processing of raw GC-MS data, particularly suitable for scripting of customized processing pipelines and for data processing in batch mode. PyMS provides limited graphical capabilities and can be used both for routine data processing and interactive/exploratory data analysis. In real-life GC-MS data processing scenarios PyMS performs as well or better than leading software packages. We demonstrate data processing scenarios simple to implement in PyMS, yet difficult to achieve with many conventional GC-MS data processing software. Automated sample processing and quantitation with PyMS can provide substantial time savings compared to more traditional interactive software systems that tightly integrate data processing with the graphical user interface.
Patterson Silver Wolf, David A; Tovar, Molly; Thompson, Kellie; Ishcomer, Jamie; Kreuter, Matthew W; Caburnay, Charlene; Boyum, Sonia
2016-01-01
Objective This study is the first to explore the impact of graphic cigarette labels with physical harm images on members of American Indian/Alaska Native (AI/AN) communities. The aim of this article is to investigate how AI/AN respond to particular graphic warning labels. Methods The parent study recruited smokers, at-risk smokers and non-smokers from three different age groups (youths aged 13–17 years, young adults aged 18–24 years and adults aged 25+ years) and five population subgroups with high smoking prevalence or smoking risk. Using nine graphic labels, this study collected participant data in the field via an iPad-administered survey and card sorting of graphic warning labels. This paper reports on findings for AI/AN participants. Results After viewing graphic warning labels, participants rated their likelihood of talking about smoking risks to friends, parents and siblings higher than their likelihood of talking to teachers and doctors. Further, this study found that certain labels (eg, the label of the toddler in the smoke cloud) made them think about their friends and family who smoke. Conclusions Given the influence of community social networks on health beliefs and attitudes, health communication using graphic warning labels could effect change in the smoking habits of AI/AN community members. Study findings suggest that graphic labels could serve as stimuli for conversations about the risks of smoking among AI/AN community members, and could be an important element of a peer-to-peer smoking cessation effort. PMID:27009143
NASA Technical Reports Server (NTRS)
Farassat, F.
1984-01-01
A simple mathematical model of a stationary source distribution for the supersonic-propeller noise-prediction formula of Farassat (1983) is developed to test the validity of the formula solutions. The conventional thickness source term is used in place of the Isom thickness formula; the relative importance of the line and surface integrals in the solutions is evaluated; and the numerical results are compared with those obtained with a conventional retarded-time solution in tables. Good agreement is obtained over elevation angles from 10 to 90 deg, and the line-integral contribution is found to be significant at all elevation angles and of the same order of magnitude as the surface-integral contribution at angles less than 30 deg. The amplitude-normalized directivity patterns for the four cases computed (x = 1.5 or 10; k = 5.0 or 50) are presented graphically.
Heckeroth, J; Boywitt, C D
2017-06-01
Considering the increasing relevance of handwritten electronically captured signatures, we evaluated the ability of forensic handwriting examiners (FHEs) to distinguish between authentic and simulated electronic signatures. Sixty-six professional FHEs examined the authenticity of electronic signatures captured with software by signotec on a smartphone Galaxy Note 4 by Samsung and signatures made with a ballpoint pen on paper (conventional signatures). In addition, we experimentally varied the name ("J. König" vs. "A. Zaiser") and the status (authentic vs. simulated) of the signatures in question. FHEs' conclusions about the authenticity did not show a statistically significant general difference between electronic and conventional signatures. Furthermore, no significant discrepancies between electronic and conventional signatures were found with regard to other important aspects of the authenticity examination such as questioned signatures' graphic information content, the suitability of the provided sample signatures, the necessity of further examinations and the levels of difficulty of the cases under examination. Thus, this study did not reveal any indications that electronic signatures captured with software by signotec on a Galaxy Note 4 are less well suited than conventional signatures for the examination of authenticity, precluding potential technical problems concerning the integrity of electronic signatures. Copyright © 2017 Elsevier B.V. All rights reserved.
The Traffic Management Advisor
NASA Technical Reports Server (NTRS)
Nedell, William; Erzberger, Heinz; Neuman, Frank
1990-01-01
The traffic management advisor (TMA) is comprised of algorithms, a graphical interface, and interactive tools for controlling the flow of air traffic into the terminal area. The primary algorithm incorporated in it is a real-time scheduler which generates efficient landing sequences and landing times for arrivals within about 200 n.m. from touchdown. A unique feature of the TMA is its graphical interface that allows the traffic manager to modify the computer-generated schedules for specific aircraft while allowing the automatic scheduler to continue generating schedules for all other aircraft. The graphical interface also provides convenient methods for monitoring the traffic flow and changing scheduling parameters during real-time operation.
Modeling crime events by d-separation method
NASA Astrophysics Data System (ADS)
Aarthee, R.; Ezhilmaran, D.
2017-11-01
Problematic legal cases have recently called for a scientifically founded method of dealing with the qualitative and quantitative roles of evidence in a case [1].To deal with quantitative, we proposed a d-separation method for modeling the crime events. A d-separation is a graphical criterion for identifying independence in a directed acyclic graph. By developing a d-separation method, we aim to lay the foundations for the development of a software support tool that can deal with the evidential reasoning in legal cases. Such a tool is meant to be used by a judge or juror, in alliance with various experts who can provide information about the details. This will hopefully improve the communication between judges or jurors and experts. The proposed method used to uncover more valid independencies than any other graphical criterion.
Review of Thawing Time Prediction Models Depending on Process Conditions and Product Characteristics
Kluza, Franciszek; Spiess, Walter E. L.; Kozłowicz, Katarzyna
2016-01-01
Summary Determining thawing times of frozen foods is a challenging problem as the thermophysical properties of the product change during thawing. A number of calculation models and solutions have been developed. The proposed solutions range from relatively simple analytical equations based on a number of assumptions to a group of empirical approaches that sometimes require complex calculations. In this paper analytical, empirical and graphical models are presented and critically reviewed. The conditions of solution, limitations and possible applications of the models are discussed. The graphical and semi--graphical models are derived from numerical methods. Using the numerical methods is not always possible as running calculations takes time, whereas the specialized software and equipment are not always cheap. For these reasons, the application of analytical-empirical models is more useful for engineering. It is demonstrated that there is no simple, accurate and feasible analytical method for thawing time prediction. Consequently, simplified methods are needed for thawing time estimation of agricultural and food products. The review reveals the need for further improvement of the existing solutions or development of new ones that will enable accurate determination of thawing time within a wide range of practical conditions of heat transfer during processing. PMID:27904387
Highway noise study : final report.
DOT National Transportation Integrated Search
1974-05-01
Three noise level measurement systems have been investigated and made operational. They are: : 1. Graphic Level Recording Method : 2. Periodic Sampling Method : 3. Environmental Noise Classifier Method : The relative accuracy of the three systems wer...
Alexander, Nathan; Woetzel, Nils; Meiler, Jens
2011-02-01
Clustering algorithms are used as data analysis tools in a wide variety of applications in Biology. Clustering has become especially important in protein structure prediction and virtual high throughput screening methods. In protein structure prediction, clustering is used to structure the conformational space of thousands of protein models. In virtual high throughput screening, databases with millions of drug-like molecules are organized by structural similarity, e.g. common scaffolds. The tree-like dendrogram structure obtained from hierarchical clustering can provide a qualitative overview of the results, which is important for focusing detailed analysis. However, in practice it is difficult to relate specific components of the dendrogram directly back to the objects of which it is comprised and to display all desired information within the two dimensions of the dendrogram. The current work presents a hierarchical agglomerative clustering method termed bcl::Cluster. bcl::Cluster utilizes the Pymol Molecular Graphics System to graphically depict dendrograms in three dimensions. This allows simultaneous display of relevant biological molecules as well as additional information about the clusters and the members comprising them.
Estimation of High-Dimensional Graphical Models Using Regularized Score Matching
Lin, Lina; Drton, Mathias; Shojaie, Ali
2017-01-01
Graphical models are widely used to model stochastic dependences among large collections of variables. We introduce a new method of estimating undirected conditional independence graphs based on the score matching loss, introduced by Hyvärinen (2005), and subsequently extended in Hyvärinen (2007). The regularized score matching method we propose applies to settings with continuous observations and allows for computationally efficient treatment of possibly non-Gaussian exponential family models. In the well-explored Gaussian setting, regularized score matching avoids issues of asymmetry that arise when applying the technique of neighborhood selection, and compared to existing methods that directly yield symmetric estimates, the score matching approach has the advantage that the considered loss is quadratic and gives piecewise linear solution paths under ℓ1 regularization. Under suitable irrepresentability conditions, we show that ℓ1-regularized score matching is consistent for graph estimation in sparse high-dimensional settings. Through numerical experiments and an application to RNAseq data, we confirm that regularized score matching achieves state-of-the-art performance in the Gaussian case and provides a valuable tool for computationally efficient estimation in non-Gaussian graphical models. PMID:28638498
Zhou, Yun; Sojkova, Jitka; Resnick, Susan M.; Wong, Dean F.
2012-01-01
Both the standardized uptake value ratio (SUVR) and the Logan plot result in biased distribution volume ratios (DVR) in ligand-receptor dynamic PET studies. The objective of this study is to use a recently developed relative equilibrium-based graphical plot (RE plot) method to improve and simplify the two commonly used methods for quantification of [11C]PiB PET. Methods The overestimation of DVR in SUVR was analyzed theoretically using the Logan and the RE plots. A bias-corrected SUVR (bcSUVR) was derived from the RE plot. Seventy-eight [11C]PiB dynamic PET scans (66 from controls and 12 from mildly cognitively impaired participants (MCI) from the Baltimore Longitudinal Study of Aging (BLSA)) were acquired over 90 minutes. Regions of interest (ROIs) were defined on coregistered MRIs. Both the ROI and pixelwise time activity curves (TACs) were used to evaluate the estimates of DVR. DVRs obtained using the Logan plot applied to ROI TACs were used as a reference for comparison of DVR estimates. Results Results from the theoretical analysis were confirmed by human studies. ROI estimates from the RE plot and the bcSUVR were nearly identical to those from the Logan plot with ROI TACs. In contrast, ROI estimates from DVR images in frontal, temporal, parietal, cingulate regions, and the striatum were underestimated by the Logan plot (controls 4 – 12%; MCI 9 – 16%) and overestimated by the SUVR (controls 8 – 16%; MCI 16 – 24%). This bias was higher in the MCI group than in controls (p < 0.01) but was not present when data were analyzed using either the RE plot or the bcSUVR. Conclusion The RE plot improves pixel-wise quantification of [11C]PiB dynamic PET compared to the conventional Logan plot. The bcSUVR results in lower bias and higher consistency of DVR estimates compared to SUVR. The RE plot and the bcSUVR are practical quantitative approaches that improve the analysis of [11C]PiB studies. PMID:22414634
NASA Astrophysics Data System (ADS)
Yozgatligil, Ceylan; Aslan, Sipan; Iyigun, Cem; Batmaz, Inci
2013-04-01
This study aims to compare several imputation methods to complete the missing values of spatio-temporal meteorological time series. To this end, six imputation methods are assessed with respect to various criteria including accuracy, robustness, precision, and efficiency for artificially created missing data in monthly total precipitation and mean temperature series obtained from the Turkish State Meteorological Service. Of these methods, simple arithmetic average, normal ratio (NR), and NR weighted with correlations comprise the simple ones, whereas multilayer perceptron type neural network and multiple imputation strategy adopted by Monte Carlo Markov Chain based on expectation-maximization (EM-MCMC) are computationally intensive ones. In addition, we propose a modification on the EM-MCMC method. Besides using a conventional accuracy measure based on squared errors, we also suggest the correlation dimension (CD) technique of nonlinear dynamic time series analysis which takes spatio-temporal dependencies into account for evaluating imputation performances. Depending on the detailed graphical and quantitative analysis, it can be said that although computational methods, particularly EM-MCMC method, are computationally inefficient, they seem favorable for imputation of meteorological time series with respect to different missingness periods considering both measures and both series studied. To conclude, using the EM-MCMC algorithm for imputing missing values before conducting any statistical analyses of meteorological data will definitely decrease the amount of uncertainty and give more robust results. Moreover, the CD measure can be suggested for the performance evaluation of missing data imputation particularly with computational methods since it gives more precise results in meteorological time series.
Learning with Interactive Computer Graphics in the Undergraduate Neuroscience Classroom
Pani, John R.; Chariker, Julia H.; Naaz, Farah; Mattingly, William; Roberts, Joshua; Sephton, Sandra E.
2014-01-01
Instruction of neuroanatomy depends on graphical representation and extended self-study. As a consequence, computer-based learning environments that incorporate interactive graphics should facilitate instruction in this area. The present study evaluated such a system in the undergraduate neuroscience classroom. The system used the method of adaptive exploration, in which exploration in a high fidelity graphical environment is integrated with immediate testing and feedback in repeated cycles of learning. The results of this study were that students considered the graphical learning environment to be superior to typical classroom materials used for learning neuroanatomy. Students managed the frequency and duration of study, test, and feedback in an efficient and adaptive manner. For example, the number of tests taken before reaching a minimum test performance of 90% correct closely approximated the values seen in more regimented experimental studies. There was a wide range of student opinion regarding the choice between a simpler and a more graphically compelling program for learning sectional anatomy. Course outcomes were predicted by individual differences in the use of the software that reflected general work habits of the students, such as the amount of time committed to testing. The results of this introduction into the classroom are highly encouraging for development of computer-based instruction in biomedical disciplines. PMID:24449123
Investigation of ODE integrators using interactive graphics. [Ordinary Differential Equations
NASA Technical Reports Server (NTRS)
Brown, R. L.
1978-01-01
Two FORTRAN programs using an interactive graphic terminal to generate accuracy and stability plots for given multistep ordinary differential equation (ODE) integrators are described. The first treats the fixed stepsize linear case with complex variable solutions, and generates plots to show accuracy and error response to step driving function of a numerical solution, as well as the linear stability region. The second generates an analog to the stability region for classes of non-linear ODE's as well as accuracy plots. Both systems can compute method coefficients from a simple specification of the method. Example plots are given.
Metaplot: A Novel Stata Graph for Assessing Heterogeneity at a Glance
Poorolajal, J; Mahmoodi, M; Majdzadeh, R; Fotouhi, A
2010-01-01
Background: Heterogeneity is usually a major concern in meta-analysis. Although there are some statistical approaches for assessing variability across studies, here we present a new approach to heterogeneity using “MetaPlot” that investigate the influence of a single study on the overall heterogeneity. Methods: MetaPlot is a two-way (x, y) graph, which can be considered as a complementary graphical approach for testing heterogeneity. This method shows graphically as well as numerically the results of an influence analysis, in which Higgins’ I2 statistic with 95% (Confidence interval) CI are computed omitting one study in each turn and then are plotted against reciprocal of standard error (1/SE) or “precision”. In this graph, “1/SE” lies on x axis and “I2 results” lies on y axe. Results: Having a first glance at MetaPlot, one can predict to what extent omission of a single study may influence the overall heterogeneity. The precision on x-axis enables us to distinguish the size of each trial. The graph describes I2 statistic with 95% CI graphically as well as numerically in one view for prompt comparison. It is possible to implement MetaPlot for meta-analysis of different types of outcome data and summary measures. Conclusion: This method presents a simple graphical approach to identify an outlier and its effect on overall heterogeneity at a glance. We wish to suggest MetaPlot to Stata experts to prepare its module for the software. PMID:23113013
ERIC Educational Resources Information Center
Erlichson, Herman
2000-01-01
Describes an experiment that extends Newton's instantaneous-impulse method of orbital analysis to a graphical method of orbit determination. Discusses the experiment's usefulness for teaching both horizontal projectile motion and instantaneous impulse. (WRM)
Kinematic synthesis of adjustable robotic mechanisms
NASA Astrophysics Data System (ADS)
Chuenchom, Thatchai
1993-01-01
Conventional hard automation, such as a linkage-based or a cam-driven system, provides high speed capability and repeatability but not the flexibility required in many industrial applications. The conventional mechanisms, that are typically single-degree-of-freedom systems, are being increasingly replaced by multi-degree-of-freedom multi-actuators driven by logic controllers. Although this new trend in sophistication provides greatly enhanced flexibility, there are many instances where the flexibility needs are exaggerated and the associated complexity is unnecessary. Traditional mechanism-based hard automation, on the other hand, neither can fulfill multi-task requirements nor are cost-effective mainly due to lack of methods and tools to design-in flexibility. This dissertation attempts to bridge this technological gap by developing Adjustable Robotic Mechanisms (ARM's) or 'programmable mechanisms' as a middle ground between high speed hard automation and expensive serial jointed-arm robots. This research introduces the concept of adjustable robotic mechanisms towards cost-effective manufacturing automation. A generalized analytical synthesis technique has been developed to support the computational design of ARM's that lays the theoretical foundation for synthesis of adjustable mechanisms. The synthesis method developed in this dissertation, called generalized adjustable dyad and triad synthesis, advances the well-known Burmester theory in kinematics to a new level. While this method provides planar solutions, a novel patented scheme is utilized for converting prescribed three-dimensional motion specifications into sets of planar projections. This provides an analytical and a computational tool for designing adjustable mechanisms that satisfy multiple sets of three-dimensional motion specifications. Several design issues were addressed, including adjustable parameter identification, branching defect, and mechanical errors. An efficient mathematical scheme for identification of adjustable member was also developed. The analytical synthesis techniques developed in this dissertation were successfully implemented in a graphic-intensive user-friendly computer program. A physical prototype of a general purpose adjustable robotic mechanism has been constructed to serve as a proof-of-concept model.
permGPU: Using graphics processing units in RNA microarray association studies.
Shterev, Ivo D; Jung, Sin-Ho; George, Stephen L; Owzar, Kouros
2010-06-16
Many analyses of microarray association studies involve permutation, bootstrap resampling and cross-validation, that are ideally formulated as embarrassingly parallel computing problems. Given that these analyses are computationally intensive, scalable approaches that can take advantage of multi-core processor systems need to be developed. We have developed a CUDA based implementation, permGPU, that employs graphics processing units in microarray association studies. We illustrate the performance and applicability of permGPU within the context of permutation resampling for a number of test statistics. An extensive simulation study demonstrates a dramatic increase in performance when using permGPU on an NVIDIA GTX 280 card compared to an optimized C/C++ solution running on a conventional Linux server. permGPU is available as an open-source stand-alone application and as an extension package for the R statistical environment. It provides a dramatic increase in performance for permutation resampling analysis in the context of microarray association studies. The current version offers six test statistics for carrying out permutation resampling analyses for binary, quantitative and censored time-to-event traits.
Compress compound images in H.264/MPGE-4 AVC by exploiting spatial correlation.
Lan, Cuiling; Shi, Guangming; Wu, Feng
2010-04-01
Compound images are a combination of text, graphics and natural image. They present strong anisotropic features, especially on the text and graphics parts. These anisotropic features often render conventional compression inefficient. Thus, this paper proposes a novel coding scheme from the H.264 intraframe coding. In the scheme, two new intramodes are developed to better exploit spatial correlation in compound images. The first is the residual scalar quantization (RSQ) mode, where intrapredicted residues are directly quantized and coded without transform. The second is the base colors and index map (BCIM) mode that can be viewed as an adaptive color quantization. In this mode, an image block is represented by several representative colors, referred to as base colors, and an index map to compress. Every block selects its coding mode from two new modes and the previous intramodes in H.264 by rate-distortion optimization (RDO). Experimental results show that the proposed scheme improves the coding efficiency even more than 10 dB at most bit rates for compound images and keeps a comparable efficient performance to H.264 for natural images.
Graphical User Interface in Art
NASA Astrophysics Data System (ADS)
Gwilt, Ian
This essay discusses the use of the Graphical User Interface (GUI) as a site of creative practice. By creatively repositioning the GUI as a work of art it is possible to challenge our understanding and expectations of the conventional computer interface wherein the icons and navigational architecture of the GUI no longer function as a technological tool. These artistic recontextualizations are often used to question our engagement with technology and to highlight the pivotal place that the domestic computer has taken in our everyday social, cultural and (increasingly), creative domains. Through these works the media specificity of the screen-based GUI can broken by dramatic changes in scale, form and configuration. This can be seen through the work of new media artists who have re-imagined the GUI in a number of creative forms both, within the digital, as image, animation, net and interactive art, and in the analogue, as print, painting, sculpture, installation and performative event. Furthermore as a creative work, the GUI can also be utilized as a visual way-finder to explore the relationship between the dynamic potentials of the digital and the concretized qualities of the material artifact.
Sarrafzadegan, Nizal; Kelishad, Roya; Rabiei, Katayoun; Abedi, Heidarali; Mohaseli, Khadijeh Fereydoun; Masooleh, Hasan Azaripour; Alavi, Mousa; Heidari, Gholamreza; Ghaffari, Mostafa; O’Loughlin, Jennifer
2012-01-01
Background: Iran is one of the countries that has ratified the World Health Organization Framework Convention of Tobacco Control (WHO-FCTC), and has implemented a series of tobacco control interventions including the Comprehensive Tobacco Control Law. Enforcement of this legislation and assessment of its outcome requires a dedicated evaluation system. This study aimed to develop a generic model to evaluate the implementation of the Comprehensive Tobacco Control Law in Iran that was provided based on WHO-FCTC articles. Materials and Methods: Using a grounded theory approach, qualitative data were collected from 265 subjects in individual interviews and focus group discussions with policymakers who designed the legislation, key stakeholders, and members of the target community. In addition, field observations data in supermarkets/shops, restaurants, teahouses and coffee shops were collected. Data were analyzed in two stages through conceptual theoretical coding. Findings: Overall, 617 open codes were extracted from the data into tables; 72 level-3 codes were retained from the level-2 code series. Using a Model Met paradigm, the relationships between the components of each paradigm were depicted graphically. The evaluation model entailed three levels, namely: short-term results, process evaluation and long-term results. Conclusions: Central concept of the process of evaluation is that enforcing the law influences a variety of internal and environmental factors including legislative changes. These factors will be examined during the process evaluation and context evaluation. The current model can be applicable for providing FCTC evaluation tools across other jurisdictions. PMID:23833621
Is complex allometry in field metabolic rates of mammals a statistical artifact?
Packard, Gary C
2017-01-01
Recent reports indicate that field metabolic rates (FMRs) of mammals conform to a pattern of complex allometry in which the exponent in a simple, two-parameter power equation increases steadily as a dependent function of body mass. The reports were based, however, on indirect analyses performed on logarithmic transformations of the original data. I re-examined values for FMR and body mass for 114 species of mammal by the conventional approach to allometric analysis (to illustrate why the approach is unreliable) and by linear and nonlinear regression on untransformed variables (to illustrate the power and versatility of newer analytical methods). The best of the regression models fitted directly to untransformed observations is a three-parameter power equation with multiplicative, lognormal, heteroscedastic error and an allometric exponent of 0.82. The mean function is a good fit to data in graphical display. The significant intercept in the model may simply have gone undetected in prior analyses because conventional allometry assumes implicitly that the intercept is zero; or the intercept may be a spurious finding resulting from bias introduced by the haphazard sampling that underlies "exploratory" analyses like the one reported here. The aforementioned issues can be resolved only by gathering new data specifically intended to address the question of scaling of FMR with body mass in mammals. However, there is no support for the concept of complex allometry in the relationship between FMR and body size in mammals. Copyright © 2016 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Zakirov, Andrey; Belousov, Sergei; Valuev, Ilya; Levchenko, Vadim; Perepelkina, Anastasia; Zempo, Yasunari
2017-10-01
We demonstrate an efficient approach to numerical modeling of optical properties of large-scale structures with typical dimensions much greater than the wavelength of light. For this purpose, we use the finite-difference time-domain (FDTD) method enhanced with a memory efficient Locally Recursive non-Locally Asynchronous (LRnLA) algorithm called DiamondTorre and implemented for General Purpose Graphical Processing Units (GPGPU) architecture. We apply our approach to simulation of optical properties of organic light emitting diodes (OLEDs), which is an essential step in the process of designing OLEDs with improved efficiency. Specifically, we consider a problem of excitation and propagation of surface plasmon polaritons (SPPs) in a typical OLED, which is a challenging task given that SPP decay length can be about two orders of magnitude greater than the wavelength of excitation. We show that with our approach it is possible to extend the simulated volume size sufficiently so that SPP decay dynamics is accounted for. We further consider an OLED with periodically corrugated metallic cathode and show how the SPP decay length can be greatly reduced due to scattering off the corrugation. Ultimately, we compare the performance of our algorithm to the conventional FDTD and demonstrate that our approach can efficiently be used for large-scale FDTD simulations with the use of only a single GPGPU-powered workstation, which is not practically feasible with the conventional FDTD.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sugano, Yasutaka; Mizuta, Masahiro; Takao, Seishin
Purpose: Radiotherapy of solid tumors has been performed with various fractionation regimens such as multi- and hypofractionations. However, the ability to optimize the fractionation regimen considering the physical dose distribution remains insufficient. This study aims to optimize the fractionation regimen, in which the authors propose a graphical method for selecting the optimal number of fractions (n) and dose per fraction (d) based on dose–volume histograms for tumor and normal tissues of organs around the tumor. Methods: Modified linear-quadratic models were employed to estimate the radiation effects on the tumor and an organ at risk (OAR), where the repopulation of themore » tumor cells and the linearity of the dose-response curve in the high dose range of the surviving fraction were considered. The minimization problem for the damage effect on the OAR was solved under the constraint that the radiation effect on the tumor is fixed by a graphical method. Here, the damage effect on the OAR was estimated based on the dose–volume histogram. Results: It was found that the optimization of fractionation scheme incorporating the dose–volume histogram is possible by employing appropriate cell surviving models. The graphical method considering the repopulation of tumor cells and a rectilinear response in the high dose range enables them to derive the optimal number of fractions and dose per fraction. For example, in the treatment of prostate cancer, the optimal fractionation was suggested to lie in the range of 8–32 fractions with a daily dose of 2.2–6.3 Gy. Conclusions: It is possible to optimize the number of fractions and dose per fraction based on the physical dose distribution (i.e., dose–volume histogram) by the graphical method considering the effects on tumor and OARs around the tumor. This method may stipulate a new guideline to optimize the fractionation regimen for physics-guided fractionation.« less
Serang, Oliver; Noble, William Stafford
2012-01-01
The problem of identifying the proteins in a complex mixture using tandem mass spectrometry can be framed as an inference problem on a graph that connects peptides to proteins. Several existing protein identification methods make use of statistical inference methods for graphical models, including expectation maximization, Markov chain Monte Carlo, and full marginalization coupled with approximation heuristics. We show that, for this problem, the majority of the cost of inference usually comes from a few highly connected subgraphs. Furthermore, we evaluate three different statistical inference methods using a common graphical model, and we demonstrate that junction tree inference substantially improves rates of convergence compared to existing methods. The python code used for this paper is available at http://noble.gs.washington.edu/proj/fido. PMID:22331862
Human-computer interfaces applied to numerical solution of the Plateau problem
NASA Astrophysics Data System (ADS)
Elias Fabris, Antonio; Soares Bandeira, Ivana; Ramos Batista, Valério
2015-09-01
In this work we present a code in Matlab to solve the Problem of Plateau numerically, and the code will include human-computer interface. The Problem of Plateau has applications in areas of knowledge like, for instance, Computer Graphics. The solution method will be the same one of the Surface Evolver, but the difference will be a complete graphical interface with the user. This will enable us to implement other kinds of interface like ocular mouse, voice, touch, etc. To date, Evolver does not include any graphical interface, which restricts its use by the scientific community. Specially, its use is practically impossible for most of the Physically Challenged People.
ERIC Educational Resources Information Center
Rajabzadeh, Massy
2012-01-01
In this experiment, students learn how to find the unknown concentration of sodium acetate using both the graphical treatment of standard addition and the standard addition equation. In the graphical treatment of standard addition, the peak area of the methyl peak in each of the sodium acetate standard solutions is found by integration using…
Graphical correlation of gaging-station records
Searcy, James K.
1960-01-01
A gaging-station record is a sample of the rate of flow of a stream at a given site. This sample can be used to estimate the magnitude and distribution of future flows if the record is long enough to be representative of the long-term flow of the stream. The reliability of a short-term record for estimating future flow characteristics can be improved through correlation with a long-term record. Correlation can be either numerical or graphical, but graphical correlation of gaging-station records has several advantages. The graphical correlation method is described in a step-by-step procedure with an illustrative problem of simple correlation, illustrative problems of three examples of multiple correlation--removing seasonal effect--and two examples of correlation of one record with two other records. Except in the problem on removal of seasonal effect, the same group of stations is used in the illustrative problems. The purpose of the problems is to illustrate the method--not to show the improvement that can result from multiple correlation as compared with simple correlation. Hydrologic factors determine whether a usable relation exists between gaging-station records. Statistics is only a tool for evaluating and using an existing relation, and the investigator must be guided by a knowledge of hydrology.
NASA Astrophysics Data System (ADS)
Sewell, Stephen
This thesis introduces a software framework that effectively utilizes low-cost commercially available Graphic Processing Units (GPUs) to simulate complex scientific plasma phenomena that are modeled using the Particle-In-Cell (PIC) paradigm. The software framework that was developed conforms to the Compute Unified Device Architecture (CUDA), a standard for general purpose graphic processing that was introduced by NVIDIA Corporation. This framework has been verified for correctness and applied to advance the state of understanding of the electromagnetic aspects of the development of the Aurora Borealis and Aurora Australis. For each phase of the PIC methodology, this research has identified one or more methods to exploit the problem's natural parallelism and effectively map it for execution on the graphic processing unit and its host processor. The sources of overhead that can reduce the effectiveness of parallelization for each of these methods have also been identified. One of the novel aspects of this research was the utilization of particle sorting during the grid interpolation phase. The final representation resulted in simulations that executed about 38 times faster than simulations that were run on a single-core general-purpose processing system. The scalability of this framework to larger problem sizes and future generation systems has also been investigated.
Fundamentals of satellite navigation
NASA Astrophysics Data System (ADS)
Stiller, A. H.
The basic operating principles and capabilities of conventional and satellite-based navigation systems for air, sea, and land vehicles are reviewed and illustrated with diagrams. Consideration is given to autonomous onboard systems; systems based on visible or radio beacons; the Transit, Cicada, Navstar-GPS, and Glonass satellite systems; the physical laws and parameters of satellite motion; the definition of time in satellite systems; and the content of the demodulated GPS data signal. The GPS and Glonass data format frames are presented graphically, and tables listing the GPS and Glonass satellites, their technical characteristics, and the (past or scheduled) launch dates are provided.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2004-08-24
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2007-01-02
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2006-03-21
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2002-01-01
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2000-01-01
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
Pitch-based carbon foam heat sink with phase change material
Klett, James W.; Burchell, Timothy D.
2007-01-23
A process for producing a carbon foam heat sink is disclosed which obviates the need for conventional oxidative stabilization. The process employs mesophase or isotropic pitch and a simplified process using a single mold. The foam has a relatively uniform distribution of pore sizes and a highly aligned graphic structure in the struts. The foam material can be made into a composite which is useful in high temperature sandwich panels for both thermal and structural applications. The foam is encased and filled with a phase change material to provide a very efficient heat sink device.
White, Adrian; White, Leon
2003-06-01
Well prepared posters are an effective means to communicate a simple message and stimulate discussion. A good poster requires considerable effort in identifying the vital ingredients and rejecting any superfluous material. The conventional structure for papers and abstracts is a suitable basis for posters on many subjects, with modification if necessary. Suitable topics include clinical trials, surveys, qualitative studies and case reports. Suggestions are made for contents that should be considered for each section. Careful planning of size, shape, flow and content will save time in preparation, and several technical graphical points are made, which may improve the attractiveness and readability of the poster.
NASA Technical Reports Server (NTRS)
1993-01-01
Using chordic technology, a data entry operator can finger key combinations for text or graphics input. Because only one hand is needed, a disabled person may use it. Strain and fatigue are less than when using a conventional keyboard; input is faster, and the system can be learned in about an hour. Infogrip, Inc. developed chordic input technology with Stennis Space Center (SSC). (NASA is interested in potentially faster human/computer interaction on spacecraft as well as a low cost tactile/visual training system for the handicapped.) The company is now marketing the BAT as an improved system for both disabled and non-disabled computer operators.
VA's Integrated Imaging System on three platforms.
Dayhoff, R E; Maloney, D L; Majurski, W J
1992-01-01
The DHCP Integrated Imaging System provides users with integrated patient data including text, image and graphics data. This system has been transferred from its original two screen DOS-based MUMPS platform to an X window workstation and a Microsoft Windows-based workstation. There are differences between these various platforms that impact on software design and on software development strategy. Data structures and conventions were used to isolate hardware, operating system, imaging software, and user-interface differences between platforms in the implementation of functionality for text and image display and interaction. The use of an object-oriented approach greatly increased system portability.
VA's Integrated Imaging System on three platforms.
Dayhoff, R. E.; Maloney, D. L.; Majurski, W. J.
1992-01-01
The DHCP Integrated Imaging System provides users with integrated patient data including text, image and graphics data. This system has been transferred from its original two screen DOS-based MUMPS platform to an X window workstation and a Microsoft Windows-based workstation. There are differences between these various platforms that impact on software design and on software development strategy. Data structures and conventions were used to isolate hardware, operating system, imaging software, and user-interface differences between platforms in the implementation of functionality for text and image display and interaction. The use of an object-oriented approach greatly increased system portability. PMID:1482983
GPU acceleration of Runge Kutta-Fehlberg and its comparison with Dormand-Prince method
NASA Astrophysics Data System (ADS)
Seen, Wo Mei; Gobithaasan, R. U.; Miura, Kenjiro T.
2014-07-01
There is a significant reduction of processing time and speedup of performance in computer graphics with the emergence of Graphic Processing Units (GPUs). GPUs have been developed to surpass Central Processing Unit (CPU) in terms of performance and processing speed. This evolution has opened up a new area in computing and researches where highly parallel GPU has been used for non-graphical algorithms. Physical or phenomenal simulations and modelling can be accelerated through General Purpose Graphic Processing Units (GPGPU) and Compute Unified Device Architecture (CUDA) implementations. These phenomena can be represented with mathematical models in the form of Ordinary Differential Equations (ODEs) which encompasses the gist of change rate between independent and dependent variables. ODEs are numerically integrated over time in order to simulate these behaviours. The classical Runge-Kutta (RK) scheme is the common method used to numerically solve ODEs. The Runge Kutta Fehlberg (RKF) scheme has been specially developed to provide an estimate of the principal local truncation error at each step, known as embedding estimate technique. This paper delves into the implementation of RKF scheme for GPU devices and compares its result with Dorman Prince method. A pseudo code is developed to show the implementation in detail. Hence, practitioners will be able to understand the data allocation in GPU, formation of RKF kernels and the flow of data to/from GPU-CPU upon RKF kernel evaluation. The pseudo code is then written in C Language and two ODE models are executed to show the achievable speedup as compared to CPU implementation. The accuracy and efficiency of the proposed implementation method is discussed in the final section of this paper.
An interactive graphics program for manipulation and display of panel method geometry
NASA Technical Reports Server (NTRS)
Hall, J. F.; Neuhart, D. H.; Walkley, K. B.
1983-01-01
Modern aerodynamic panel methods that handle large, complex geometries have made evident the need to interactively manipulate, modify, and view such configurations. With this purpose in mind, the GEOM program was developed. It is a menu driven, interactive program that uses the Tektronix PLOT 10 graphics software to display geometry configurations which are characterized by an abutting set of networks. These networks are composed of quadrilateral panels which are described by the coordinates of their corners. GEOM is divided into fourteen executive controlled functions. These functions are used to build configurations, scale and rotate networks, transpose networks defining M and N lines, graphically display selected networks, join and split networks, create wake networks, produce symmetric images of networks, repanel and rename networks, display configuration cross sections, and output network geometry in two formats. A data base management system is used to facilitate data transfers in this program. A sample session illustrating various capabilities of the code is included as a guide to program operation.
A Fast MHD Code for Gravitationally Stratified Media using Graphical Processing Units: SMAUG
NASA Astrophysics Data System (ADS)
Griffiths, M. K.; Fedun, V.; Erdélyi, R.
2015-03-01
Parallelization techniques have been exploited most successfully by the gaming/graphics industry with the adoption of graphical processing units (GPUs), possessing hundreds of processor cores. The opportunity has been recognized by the computational sciences and engineering communities, who have recently harnessed successfully the numerical performance of GPUs. For example, parallel magnetohydrodynamic (MHD) algorithms are important for numerical modelling of highly inhomogeneous solar, astrophysical and geophysical plasmas. Here, we describe the implementation of SMAUG, the Sheffield Magnetohydrodynamics Algorithm Using GPUs. SMAUG is a 1-3D MHD code capable of modelling magnetized and gravitationally stratified plasma. The objective of this paper is to present the numerical methods and techniques used for porting the code to this novel and highly parallel compute architecture. The methods employed are justified by the performance benchmarks and validation results demonstrating that the code successfully simulates the physics for a range of test scenarios including a full 3D realistic model of wave propagation in the solar atmosphere.
NASA Astrophysics Data System (ADS)
Unemi, Tatsuo
This chapter describes a basic framework of simulated breeding, a type of interactive evolutionary computing to breed artifacts, whose origin is Blind Watchmaker by Dawkins. These methods make it easy for humans to design a complex object adapted to his/her subjective criteria, just similarly to agricultural products we have been developing over thousands of years. Starting from randomly initialized genome, the solution candidates are improved through several generations with artificial selection. The graphical user interface helps the process of breeding with techniques of multifield user interface and partial breeding. The former improves the diversity of individuals that prevents being trapped at local optimum. The latter makes it possible for the user to fix features he/she already satisfied. These methods were examined through artistic applications by the author: SBART for graphics art and SBEAT for music. Combining with a direct genome editor and exportation to another graphical or musical tool on the computer, they can be powerful tools for artistic creation. These systems may contribute to the creation of a type of new culture.
Imagining a Stata / Python Combination
NASA Technical Reports Server (NTRS)
Fiedler, James
2012-01-01
There are occasions when a task is difficult in Stata, but fairly easy in a more general programming language. Python is a popular language for a range of uses. It is easy to use, has many high ]quality packages, and programs can be written relatively quickly. Is there any advantage in combining Stata and Python within a single interface? Stata already offers support for user-written programs, which allow extensive control over calculations, but somewhat less control over graphics. Also, except for specifying output, the user has minimal programmatic control over the user interface. Python can be used in a way that allows more control over the interface and graphics, and in so doing provide a roundabout method for satisfying some user requests (e.g., transparency levels in graphics and the ability to clear the results window). My talk will explore these ideas, present a possible method for combining Stata and Python, and give examples to demonstrate how this combination might be useful.
Geometric Detection Algorithms for Cavities on Protein Surfaces in Molecular Graphics: A Survey
Simões, Tiago; Lopes, Daniel; Dias, Sérgio; Fernandes, Francisco; Pereira, João; Jorge, Joaquim; Bajaj, Chandrajit; Gomes, Abel
2017-01-01
Detecting and analyzing protein cavities provides significant information about active sites for biological processes (e.g., protein-protein or protein-ligand binding) in molecular graphics and modeling. Using the three-dimensional structure of a given protein (i.e., atom types and their locations in 3D) as retrieved from a PDB (Protein Data Bank) file, it is now computationally viable to determine a description of these cavities. Such cavities correspond to pockets, clefts, invaginations, voids, tunnels, channels, and grooves on the surface of a given protein. In this work, we survey the literature on protein cavity computation and classify algorithmic approaches into three categories: evolution-based, energy-based, and geometry-based. Our survey focuses on geometric algorithms, whose taxonomy is extended to include not only sphere-, grid-, and tessellation-based methods, but also surface-based, hybrid geometric, consensus, and time-varying methods. Finally, we detail those techniques that have been customized for GPU (Graphics Processing Unit) computing. PMID:29520122
Johnson, Jason K.; Oyen, Diane Adele; Chertkov, Michael; ...
2016-12-01
Inference and learning of graphical models are both well-studied problems in statistics and machine learning that have found many applications in science and engineering. However, exact inference is intractable in general graphical models, which suggests the problem of seeking the best approximation to a collection of random variables within some tractable family of graphical models. In this paper, we focus on the class of planar Ising models, for which exact inference is tractable using techniques of statistical physics. Based on these techniques and recent methods for planarity testing and planar embedding, we propose a greedy algorithm for learning the bestmore » planar Ising model to approximate an arbitrary collection of binary random variables (possibly from sample data). Given the set of all pairwise correlations among variables, we select a planar graph and optimal planar Ising model defined on this graph to best approximate that set of correlations. Finally, we demonstrate our method in simulations and for two applications: modeling senate voting records and identifying geo-chemical depth trends from Mars rover data.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnson, Jason K.; Oyen, Diane Adele; Chertkov, Michael
Inference and learning of graphical models are both well-studied problems in statistics and machine learning that have found many applications in science and engineering. However, exact inference is intractable in general graphical models, which suggests the problem of seeking the best approximation to a collection of random variables within some tractable family of graphical models. In this paper, we focus on the class of planar Ising models, for which exact inference is tractable using techniques of statistical physics. Based on these techniques and recent methods for planarity testing and planar embedding, we propose a greedy algorithm for learning the bestmore » planar Ising model to approximate an arbitrary collection of binary random variables (possibly from sample data). Given the set of all pairwise correlations among variables, we select a planar graph and optimal planar Ising model defined on this graph to best approximate that set of correlations. Finally, we demonstrate our method in simulations and for two applications: modeling senate voting records and identifying geo-chemical depth trends from Mars rover data.« less
Development of MCAERO wing design panel method with interactive graphics module
NASA Technical Reports Server (NTRS)
Hawk, J. D.; Bristow, D. R.
1984-01-01
A reliable and efficient iterative method has been developed for designing wing section contours corresponding to a prescribed subcritical pressure distribution. The design process is initialized by using MCAERO (MCAIR 3-D Subsonic Potential Flow Analysis Code) to analyze a baseline configuration. A second program DMCAERO is then used to calculate a matrix containing the partial derivative of potential at each control point with respect to each unknown geometry parameter by applying a first-order expansion to the baseline equations in MCAERO. This matrix is calculated only once but is used in each iteration cycle to calculate the geometry perturbation and to analyze the perturbed geometry. The potential on the new geometry is calculated by linear extrapolation from the baseline solution. This extrapolated potential is converted to velocity by numerical differentiation, and velocity is converted to pressure by using Bernoulli's equation. There is an interactive graphics option which allows the user to graphically display the results of the design process and to interactively change either the geometry or the prescribed pressure distribution.
Zeng, Yun; Liu, Gang; Ma, Ying; Chen, Xiaoyuan; Ito, Yoichiro
2012-01-01
A new series of organic-high ionic strength aqueous two-phase solvents systems was designed for separation of highly polar compounds by spiral high-speed counter-current chromatography. A total of 21 solvent systems composed of 1-butanol-ethanol-saturated ammonium sulfate-water at various volume ratios are arranged according to an increasing order of polarity. Selection of the two-phase solvent system for a single compound or a multiple sample mixture can be achieved by two steps of partition coefficient measurements using a graphic method. The capability of the method is demonstrated by optimization of partition coefficient for seven highly polar samples including tartrazine (K=0.77), tryptophan (K=1.00), methyl green (K= 0.93), tyrosine (0.81), metanephrine (K=0.89), tyramine (K=0.98), and normetanephrine (K=0.96). Three sulfonic acid components in D&C Green No. 8 were successfully separated by HSCCC using the graphic selection of the two-phase solvent system. PMID:23467197
Stochastic DT-MRI connectivity mapping on the GPU.
McGraw, Tim; Nadar, Mariappan
2007-01-01
We present a method for stochastic fiber tract mapping from diffusion tensor MRI (DT-MRI) implemented on graphics hardware. From the simulated fibers we compute a connectivity map that gives an indication of the probability that two points in the dataset are connected by a neuronal fiber path. A Bayesian formulation of the fiber model is given and it is shown that the inversion method can be used to construct plausible connectivity. An implementation of this fiber model on the graphics processing unit (GPU) is presented. Since the fiber paths can be stochastically generated independently of one another, the algorithm is highly parallelizable. This allows us to exploit the data-parallel nature of the GPU fragment processors. We also present a framework for the connectivity computation on the GPU. Our implementation allows the user to interactively select regions of interest and observe the evolving connectivity results during computation. Results are presented from the stochastic generation of over 250,000 fiber steps per iteration at interactive frame rates on consumer-grade graphics hardware.
System and method for creating expert systems
NASA Technical Reports Server (NTRS)
Hughes, Peter M. (Inventor); Luczak, Edward C. (Inventor)
1998-01-01
A system and method provides for the creation of a highly graphical expert system without the need for programming in code. An expert system is created by initially building a data interface, defining appropriate Mission, User-Defined, Inferred, and externally-generated GenSAA (EGG) data variables whose data values will be updated and input into the expert system. Next, rules of the expert system are created by building appropriate conditions of the rules which must be satisfied and then by building appropriate actions of rules which are to be executed upon corresponding conditions being satisfied. Finally, an appropriate user interface is built which can be highly graphical in nature and which can include appropriate message display and/or modification of display characteristics of a graphical display object, to visually alert a user of the expert system of varying data values, upon conditions of a created rule being satisfied. The data interface building, rule building, and user interface building are done in an efficient manner and can be created without the need for programming in code.
FPGA Implementation of the Coupled Filtering Method and the Affine Warping Method.
Zhang, Chen; Liang, Tianzhu; Mok, Philip K T; Yu, Weichuan
2017-07-01
In ultrasound image analysis, the speckle tracking methods are widely applied to study the elasticity of body tissue. However, "feature-motion decorrelation" still remains as a challenge for the speckle tracking methods. Recently, a coupled filtering method and an affine warping method were proposed to accurately estimate strain values, when the tissue deformation is large. The major drawback of these methods is the high computational complexity. Even the graphics processing unit (GPU)-based program requires a long time to finish the analysis. In this paper, we propose field-programmable gate array (FPGA)-based implementations of both methods for further acceleration. The capability of FPGAs on handling different image processing components in these methods is discussed. A fast and memory-saving image warping approach is proposed. The algorithms are reformulated to build a highly efficient pipeline on FPGA. The final implementations on a Xilinx Virtex-7 FPGA are at least 13 times faster than the GPU implementation on the NVIDIA graphic card (GeForce GTX 580).
Evans, Abigail T.; Peters, Ellen; Strasser, Andrew A.; Emery, Lydia F.; Sheerin, Kaitlin M.; Romer, Daniel
2015-01-01
Objective Observational research suggests that placing graphic images on cigarette warning labels can reduce smoking rates, but field studies lack experimental control. Our primary objective was to determine the psychological processes set in motion by naturalistic exposure to graphic vs. text-only warnings in a randomized clinical trial involving exposure to modified cigarette packs over a 4-week period. Theories of graphic-warning impact were tested by examining affect toward smoking, credibility of warning information, risk perceptions, quit intentions, warning label memory, and smoking risk knowledge. Methods Adults who smoked between 5 and 40 cigarettes daily (N = 293; mean age = 33.7), did not have a contra-indicated medical condition, and did not intend to quit were recruited from Philadelphia, PA and Columbus, OH. Smokers were randomly assigned to receive their own brand of cigarettes for four weeks in one of three warning conditions: text only, graphic images plus text, or graphic images with elaborated text. Results Data from 244 participants who completed the trial were analyzed in structural-equation models. The presence of graphic images (compared to text-only) caused more negative affect toward smoking, a process that indirectly influenced risk perceptions and quit intentions (e.g., image->negative affect->risk perception->quit intention). Negative affect from graphic images also enhanced warning credibility including through increased scrutiny of the warnings, a process that also indirectly affected risk perceptions and quit intentions (e.g., image->negative affect->risk scrutiny->warning credibility->risk perception->quit intention). Unexpectedly, elaborated text reduced warning credibility. Finally, graphic warnings increased warning-information recall and indirectly increased smoking-risk knowledge at the end of the trial and one month later. Conclusions In the first naturalistic clinical trial conducted, graphic warning labels are more effective than text-only warnings in encouraging smokers to consider quitting and in educating them about smoking’s risks. Negative affective reactions to smoking, thinking about risks, and perceptions of credibility are mediators of their impact. Trial Registration Clinicaltrials.gov NCT01782053 PMID:26672982
NASA Astrophysics Data System (ADS)
Jackson, Chester W.; Alexander, Clark R.; Bush, David M.
2012-04-01
The AMBUR (Analyzing Moving Boundaries Using R) package for the R software environment provides a collection of functions for assisting with analyzing and visualizing historical shoreline change. The package allows import and export of geospatial data in ESRI shapefile format, which is compatible with most commercial and open-source GIS software. The "baseline and transect" method is the primary technique used to quantify distances and rates of shoreline movement, and to detect classification changes across time. Along with the traditional "perpendicular" transect method, two new transect methods, "near" and "filtered," assist with quantifying changes along curved shorelines that are problematic for perpendicular transect methods. Output from the analyses includes data tables, graphics, and geospatial data, which are useful in rapidly assessing trends and potential errors in the dataset. A forecasting function also allows the user to estimate the future location of the shoreline and store the results in a shapefile. Other utilities and tools provided in the package assist with preparing and manipulating geospatial data, error checking, and generating supporting graphics and shapefiles. The package can be customized to perform additional statistical, graphical, and geospatial functions, and, it is capable of analyzing the movement of any boundary (e.g., shorelines, glacier terminus, fire edge, and marine and terrestrial ecozones).
Innovative methods for knowledge transfer
NASA Astrophysics Data System (ADS)
Goanta, A. M.
2015-11-01
At this time, learning takes place, either with classic books on paper support or using books scanned or drawn and further converted into PDF or PPT files that are printed on type support CD / DVD. The latter modern means of learning, the study adds live on the Internet using search engines and not least e-learning method, which allows the study of bibliographic related materials in PDF or PPT, stacked and grouped on the basis of a curriculum imposed which can be accessed on a website via a user name and password. Innovative methods come to successfully use other file types than those mentioned above. The graphics in teaching technical subjects such as descriptive geometry can be achieved using animated PowerPoint files, allowing for visualization of steps to be taken, in the case of solving by drawing a descriptive geometry. Another innovative method relies on the use of HTML files, inspired by related sites help design software packages that can be used when teaching descriptive geometry that the technical design. Through this work, the author has proposed to present a new innovative method, which is inspired by the methods listed above, but involves using AVI files to teaching of computer-assisted type graphics or info graphics. In general this new author's method lends itself particularly well to the teaching of the use of software packages because the student actually see the place from where the delivered command is accessed and contextual options of right button of the mouse. These laboratory courses or mentoring can be freestanding cards support CD / DVD or can be posted on websites with restricted access based on user name and password. Practically paper presents the methodology of creating courses and tutorials in AVI format, and how to monitor the degree of accessing the website, on which there were posted mentioned teaching materials, using the tools offered by Google Analytics. The results consist of graphical work, about the degree of access to educational materials, made on the basis of various criteria, such as the: time of access, access location, age of those who accessed, chapters accessed, etc. Finally the author presents his findings on the benefits of this method.
Method of characteristics for three-dimensional axially symmetrical supersonic flows.
NASA Technical Reports Server (NTRS)
Sauer, R
1947-01-01
An approximation method for three-dimensional axially symmetrical supersonic flows is developed; it is based on the characteristics theory (represented partly graphically, partly analytically). Thereafter this method is applied to the construction of rotationally symmetrical nozzles. (author)
Macy, Jonathan T.; Chassin, Laurie; Presson, Clark C.; Yeung, Ellen
2015-01-01
Objective Test the effect of exposure to the U.S. Food and Drug Administration’s proposed graphic images with text warning statements for cigarette packages on implicit and explicit attitudes toward smoking. Design and methods A two-session web-based study was conducted with 2192 young adults 18–25 years old. During session one, demographics, smoking behavior, and baseline implicit and explicit attitudes were assessed. Session two, completed on average 18 days later, contained random assignment to viewing one of three sets of cigarette packages, graphic images with text warnings, text warnings only, or current U.S Surgeon General’s text warnings. Participants then completed post-exposure measures of implicit and explicit attitudes. ANCOVAs tested the effect of condition on the outcomes, controlling for baseline attitudes. Results Smokers who viewed packages with graphic images plus text warnings demonstrated more negative implicit attitudes compared to smokers in the other conditions (p=.004). For the entire sample, explicit attitudes were more negative for those who viewed graphic images plus text warnings compared to those who viewed current U.S. Surgeon General’s text warnings (p=.014), but there was no difference compared to those who viewed text-only warnings. Conclusion Graphic health warnings on cigarette packages can influence young adult smokers’ implicit attitudes toward smoking. PMID:26442992
Soft bilateral filtering volumetric shadows using cube shadow maps
Ali, Hatam H.; Sunar, Mohd Shahrizal; Kolivand, Hoshang
2017-01-01
Volumetric shadows often increase the realism of rendered scenes in computer graphics. Typical volumetric shadows techniques do not provide a smooth transition effect in real-time with conservation on crispness of boundaries. This research presents a new technique for generating high quality volumetric shadows by sampling and interpolation. Contrary to conventional ray marching method, which requires extensive time, this proposed technique adopts downsampling in calculating ray marching. Furthermore, light scattering is computed in High Dynamic Range buffer to generate tone mapping. The bilateral interpolation is used along a view rays to smooth transition of volumetric shadows with respect to preserving-edges. In addition, this technique applied a cube shadow map to create multiple shadows. The contribution of this technique isreducing the number of sample points in evaluating light scattering and then introducing bilateral interpolation to improve volumetric shadows. This contribution is done by removing the inherent deficiencies significantly in shadow maps. This technique allows obtaining soft marvelous volumetric shadows, having a good performance and high quality, which show its potential for interactive applications. PMID:28632740
An object-oriented data reduction system in Fortran
NASA Technical Reports Server (NTRS)
Bailey, J.
1992-01-01
A data reduction system for the AAO two-degree field project is being developed using an object-oriented approach. Rather than use an object-oriented language (such as C++) the system is written in Fortran and makes extensive use of existing subroutine libraries provided by the UK Starlink project. Objects are created using the extensible N-dimensional Data Format (NDF) which itself is based on the Hierarchical Data System (HDS). The software consists of a class library, with each class corresponding to a Fortran subroutine with a standard calling sequence. The methods of the classes provide operations on NDF objects at a similar level of functionality to the applications of conventional data reduction systems. However, because they are provided as callable subroutines, they can be used as building blocks for more specialist applications. The class library is not dependent on a particular software environment thought it can be used effectively in ADAM applications. It can also be used from standalone Fortran programs. It is intended to develop a graphical user interface for use with the class library to form the 2dF data reduction system.
SimGraph: A Flight Simulation Data Visualization Workstation
NASA Technical Reports Server (NTRS)
Kaplan, Joseph A.; Kenney, Patrick S.
1997-01-01
Today's modern flight simulation research produces vast amounts of time sensitive data, making a qualitative analysis of the data difficult while it remains in a numerical representation. Therefore, a method of merging related data together and presenting it to the user in a more comprehensible format is necessary. Simulation Graphics (SimGraph) is an object-oriented data visualization software package that presents simulation data in animated graphical displays for easy interpretation. Data produced from a flight simulation is presented by SimGraph in several different formats, including: 3-Dimensional Views, Cockpit Control Views, Heads-Up Displays, Strip Charts, and Status Indicators. SimGraph can accommodate the addition of new graphical displays to allow the software to be customized to each user s particular environment. A new display can be developed and added to SimGraph without having to design a new application, allowing the graphics programmer to focus on the development of the graphical display. The SimGraph framework can be reused for a wide variety of visualization tasks. Although it was created for the flight simulation facilities at NASA Langley Research Center, SimGraph can be reconfigured to almost any data visualization environment. This paper describes the capabilities and operations of SimGraph.
Cartographic symbol library considering symbol relations based on anti-aliasing graphic library
NASA Astrophysics Data System (ADS)
Mei, Yang; Li, Lin
2007-06-01
Cartographic visualization represents geographic information with a map form, which enables us retrieve useful geospatial information. In digital environment, cartographic symbol library is the base of cartographic visualization and is an essential component of Geographic Information System as well. Existing cartographic symbol libraries have two flaws. One is the display quality and the other one is relations adjusting. Statistic data presented in this paper indicate that the aliasing problem is a major factor on the symbol display quality on graphic display devices. So, effective graphic anti-aliasing methods based on a new anti-aliasing algorithm are presented and encapsulated in an anti-aliasing graphic library with the form of Component Object Model. Furthermore, cartographic visualization should represent feature relation in the way of correctly adjusting symbol relations besides displaying an individual feature. But current cartographic symbol libraries don't have this capability. This paper creates a cartographic symbol design model to implement symbol relations adjusting. Consequently the cartographic symbol library based on this design model can provide cartographic visualization with relations adjusting capability. The anti-aliasing graphic library and the cartographic symbol library are sampled and the results prove that the two libraries both have better efficiency and effect.
TIGER: Turbomachinery interactive grid generation
NASA Technical Reports Server (NTRS)
Soni, Bharat K.; Shih, Ming-Hsin; Janus, J. Mark
1992-01-01
A three dimensional, interactive grid generation code, TIGER, is being developed for analysis of flows around ducted or unducted propellers. TIGER is a customized grid generator that combines new technology with methods from general grid generation codes. The code generates multiple block, structured grids around multiple blade rows with a hub and shroud for either C grid or H grid topologies. The code is intended for use with a Euler/Navier-Stokes solver also being developed, but is general enough for use with other flow solvers. TIGER features a silicon graphics interactive graphics environment that displays a pop-up window, graphics window, and text window. The geometry is read as a discrete set of points with options for several industrial standard formats and NASA standard formats. Various splines are available for defining the surface geometries. Grid generation is done either interactively or through a batch mode operation using history files from a previously generated grid. The batch mode operation can be done either with a graphical display of the interactive session or with no graphics so that the code can be run on another computer system. Run time can be significantly reduced by running on a Cray-YMP.
A comparison of two approaches for representing AAC vocabulary for young children.
Worah, Smita; McNaughton, David; Light, Janice; Benedek-Wood, Elizabeth
2015-01-01
Young children with complex communication needs often experience difficulty in using currently available graphic symbol systems as a method of augmentative and alternative communication (AAC). Information on young children's performance with graphic representations based on this population's conceptualizations of these vocabulary items may assist in the development of more effective AAC systems. This study developed Developmentally Appropriate Symbols (DAS) for 10 early emerging vocabulary concepts using procedures designed to address both conceptual and appeal issues for graphic representations for young children. Using a post-test only, between-subjects comparison group design, 40 typically-developing 2.5-3.5-year-old children were randomly assigned to receive a brief training in either of two different types of graphic symbol sets: (a) DAS or (b) Picture Communication Symbols (PCS), a, commercially available graphic symbol system. Results of a two sample independent t-test provide evidence that children in the DAS condition correctly identified more symbols than children trained with the PCS symbols. There was no evidence of a preference between the symbol sets. The results provide support for careful consideration of children's use and understanding of language in developing AAC systems for young children.
NASA Astrophysics Data System (ADS)
Alshakova, E. L.
2017-01-01
The program in the AutoLISP language allows automatically to form parametrical drawings during the work in the AutoCAD software product. Students study development of programs on AutoLISP language with the use of the methodical complex containing methodical instructions in which real examples of creation of images and drawings are realized. Methodical instructions contain reference information necessary for the performance of the offered tasks. The method of step-by-step development of the program is the basis for training in programming on AutoLISP language: the program draws elements of the drawing of a detail by means of definitely created function which values of arguments register in that sequence in which AutoCAD gives out inquiries when performing the corresponding command in the editor. The process of the program design is reduced to the process of step-by-step formation of functions and sequence of their calls. The author considers the development of the AutoLISP program for the creation of parametrical drawings of details, the defined design, the user enters the dimensions of elements of details. These programs generate variants of tasks of the graphic works performed in educational process of "Engineering graphics", "Engineering and computer graphics" disciplines. Individual tasks allow to develop at students skills of independent work in reading and creation of drawings, as well as 3D modeling.
Display system for imaging scientific telemetric information
NASA Technical Reports Server (NTRS)
Zabiyakin, G. I.; Rykovanov, S. N.
1979-01-01
A system for imaging scientific telemetric information, based on the M-6000 minicomputer and the SIGD graphic display, is described. Two dimensional graphic display of telemetric information and interaction with the computer, in analysis and processing of telemetric parameters displayed on the screen is provided. The running parameter information output method is presented. User capabilities in the analysis and processing of telemetric information imaged on the display screen and the user language are discussed and illustrated.
Graphics applications utilizing parallel processing
NASA Technical Reports Server (NTRS)
Rice, John R.
1990-01-01
The results are presented of research conducted to develop a parallel graphic application algorithm to depict the numerical solution of the 1-D wave equation, the vibrating string. The research was conducted on a Flexible Flex/32 multiprocessor and a Sequent Balance 21000 multiprocessor. The wave equation is implemented using the finite difference method. The synchronization issues that arose from the parallel implementation and the strategies used to alleviate the effects of the synchronization overhead are discussed.
Storyboard method of end-user programming with natural language configuration
Bouchard, Ann M; Osbourn, Gordon C
2013-11-19
A technique for end-user programming includes populating a template with graphically illustrated actions and then invoking a command to generate a screen element based on the template. The screen element is rendered within a computing environment and provides a mechanism for triggering execution of a sequence of user actions. The sequence of user actions is based at least in part on the graphically illustrated actions populated into the template.
Contaminants | Hydrogen and Fuel Cells | NREL
-Derived Contaminants Overview Materials Methods Data Tool Partners Publications System Contaminants using several screening methods. The materials are from different manufacturers, comprise different Characterization Methods A flowchart graphic that shows the experimental methods used in the system contaminants
Rodríguez-Arias, Miquel Angel; Rodó, Xavier
2004-03-01
Here we describe a practical, step-by-step primer to scale-dependent correlation (SDC) analysis. The analysis of transitory processes is an important but often neglected topic in ecological studies because only a few statistical techniques appear to detect temporary features accurately enough. We introduce here the SDC analysis, a statistical and graphical method to study transitory processes at any temporal or spatial scale. SDC analysis, thanks to the combination of conventional procedures and simple well-known statistical techniques, becomes an improved time-domain analogue of wavelet analysis. We use several simple synthetic series to describe the method, a more complex example, full of transitory features, to compare SDC and wavelet analysis, and finally we analyze some selected ecological series to illustrate the methodology. The SDC analysis of time series of copepod abundances in the North Sea indicates that ENSO primarily is the main climatic driver of short-term changes in population dynamics. SDC also uncovers some long-term, unexpected features in the population. Similarly, the SDC analysis of Nicholson's blowflies data locates where the proposed models fail and provides new insights about the mechanism that drives the apparent vanishing of the population cycle during the second half of the series.
Domingos Alves, Renata; Romero-González, Roberto; López-Ruiz, Rosalía; Jiménez-Medina, M L; Garrido Frenich, Antonia
2016-11-01
An analytical method based on a modified QuPPe (quick polar pesticide) extraction procedure coupled with liquid chromatography-tandem mass spectrometry (LC-MS/MS) was evaluated for the determination of four polar compounds (chlorate, fosetyl-Al, maleic hydrazide, and perchlorate) in nutraceutical products obtained from soy. Experimental conditions including extraction such as solvent, acidification, time, and clean-up sorbents were varied. Acidified acetonitrile (1 % formic acid, v/v) was used as extraction solvent instead of methanol (conventional QuPPe), which provides a doughy mixture which cannot be injected into the LC. Clean-up or derivatization steps were avoided. For analysis, several stationary phases were evaluated and Hypercarb (porous graphitic carbon) provided the best results. The optimized method was validated and recoveries ranged between 46 and 119 %, and correction factors can be used for quantification purposes bearing in mind that inter-day precision was equal to or lower than 17 %. Limits of quantification (LOQs) ranged from 4 to 100 μg kg -1 . Soy-based nutraceutical products were analyzed and chlorate was detected in five samples at concentrations between 63 and 1642 μg kg -1 . Graphical Abstract Analysis of polar compounds in soy-based nutraceutical products.
A practical implementation of multi-frequency widefield frequency-domain FLIM
Chen, Hongtao
2013-01-01
Widefield frequency-domain fluorescence lifetime imaging microscopy (FD-FLIM) is a fast and accurate method to measure the fluorescence lifetime, especially in kinetic studies in biomedical researches. However, the small range of modulation frequencies available in commercial instruments makes this technique limited in its applications. Here we describe a practical implementation of multi-frequency widefield FD-FLIM using a pulsed supercontinuum laser and a direct digital synthesizer. In this instrument we use a pulse to modulate the image intensifier rather than the more conventional sine wave modulation. This allows parallel multi-frequency FLIM measurement using the Fast Fourier Transform and the cross-correlation technique, which permits precise and simultaneous isolation of individual frequencies. In addition, the pulse modulation at the cathode of image intensifier restored the loss of optical resolution caused by the defocusing effect when the voltage at the cathode is sinusoidally modulated. Furthermore, in our implementation of this technique, data can be graphically analyzed by the phasor method while data are acquired, which allows easy fit-free lifetime analysis of FLIM images. Here our measurements of standard fluorescent samples and a Föster resonance energy transfer pair demonstrate that the widefield multi-frequency FLIM system is a valuable and simple tool in fluorescence imaging studies. PMID:23296945
Geological modelling of mineral deposits for prediction in mining
NASA Astrophysics Data System (ADS)
Sides, E. J.
Accurate prediction of the shape, location, size and properties of the solid rock materials to be extracted during mining is essential for reliable technical and financial planning. This is achieved through geological modelling of the three-dimensional (3D) shape and properties of the materials present in mineral deposits, and the presentation of results in a form which is accessible to mine planning engineers. In recent years the application of interactive graphics software, offering 3D database handling, modelling and visualisation, has greatly enhanced the options available for predicting the subsurface limits and characteristics of mineral deposits. A review of conventional 3D geological interpretation methods, and the model struc- tures and modelling methods used in reserve estimation and mine planning software packages, illustrates the importance of such approaches in the modern mining industry. Despite the widespread introduction and acceptance of computer hardware and software in mining applications, in recent years, there has been little fundamental change in the way in which geology is used in orebody modelling for predictive purposes. Selected areas of current research, aimed at tackling issues such as the use of orientation data, quantification of morphological differences, incorporation of geological age relationships, multi-resolution models and the application of virtual reality hardware and software, are discussed.
Saka, Takashi
2016-05-01
The dynamical theory for perfect crystals in the Laue case was reformulated using the Riemann surface, as used in complex analysis. In the two-beam approximation, each branch of the dispersion surface is specified by one sheet of the Riemann surface. The characteristic features of the dispersion surface are analytically revealed using four parameters, which are the real and imaginary parts of two quantities specifying the degree of departure from the exact Bragg condition and the reflection strength. By representing these parameters on complex planes, these characteristics can be graphically depicted on the Riemann surface. In the conventional case, the absorption is small and the real part of the reflection strength is large, so the formulation is the same as the traditional analysis. However, when the real part of the reflection strength is small or zero, the two branches of the dispersion surface cross, and the dispersion relationship becomes similar to that of the Bragg case. This is because the geometrical relationships among the parameters are similar in both cases. The present analytical method is generally applicable, irrespective of the magnitudes of the parameters. Furthermore, the present method analytically revealed many characteristic features of the dispersion surface and will be quite instructive for further numerical calculations of rocking curves.
Graphical evaluation of complexometric titration curves.
Guinon, J L
1985-04-01
A graphical method, based on logarithmic concentration diagrams, for construction, without any calculations, of complexometric titration curves is examined. The titration curves obtained for different kinds of unidentate, bidentate and quadridentate ligands clearly show why only chelating ligands are usually used in titrimetric analysis. The method has also been applied to two practical cases where unidentate ligands are used: (a) the complexometric determination of mercury(II) with halides and (b) the determination of cyanide with silver, which involves both a complexation and a precipitation system; for this purpose construction of the diagrams for the HgCl(2)/HgCl(+)/Hg(2+) and Ag(CN)(2)(-)/AgCN/CN(-) systems is considered in detail.
The multifacet graphically contracted function method. I. Formulation and implementation
NASA Astrophysics Data System (ADS)
Shepard, Ron; Gidofalvi, Gergely; Brozell, Scott R.
2014-08-01
The basic formulation for the multifacet generalization of the graphically contracted function (MFGCF) electronic structure method is presented. The analysis includes the discussion of linear dependency and redundancy of the arc factor parameters, the computation of reduced density matrices, Hamiltonian matrix construction, spin-density matrix construction, the computation of optimization gradients for single-state and state-averaged calculations, graphical wave function analysis, and the efficient computation of configuration state function and Slater determinant expansion coefficients. Timings are given for Hamiltonian matrix element and analytic optimization gradient computations for a range of model problems for full-CI Shavitt graphs, and it is observed that both the energy and the gradient computation scale as O(N2n4) for N electrons and n orbitals. The important arithmetic operations are within dense matrix-matrix product computational kernels, resulting in a computationally efficient procedure. An initial implementation of the method is used to present applications to several challenging chemical systems, including N2 dissociation, cubic H8 dissociation, the symmetric dissociation of H2O, and the insertion of Be into H2. The results are compared to the exact full-CI values and also to those of the previous single-facet GCF expansion form.
The multifacet graphically contracted function method. I. Formulation and implementation.
Shepard, Ron; Gidofalvi, Gergely; Brozell, Scott R
2014-08-14
The basic formulation for the multifacet generalization of the graphically contracted function (MFGCF) electronic structure method is presented. The analysis includes the discussion of linear dependency and redundancy of the arc factor parameters, the computation of reduced density matrices, Hamiltonian matrix construction, spin-density matrix construction, the computation of optimization gradients for single-state and state-averaged calculations, graphical wave function analysis, and the efficient computation of configuration state function and Slater determinant expansion coefficients. Timings are given for Hamiltonian matrix element and analytic optimization gradient computations for a range of model problems for full-CI Shavitt graphs, and it is observed that both the energy and the gradient computation scale as O(N(2)n(4)) for N electrons and n orbitals. The important arithmetic operations are within dense matrix-matrix product computational kernels, resulting in a computationally efficient procedure. An initial implementation of the method is used to present applications to several challenging chemical systems, including N2 dissociation, cubic H8 dissociation, the symmetric dissociation of H2O, and the insertion of Be into H2. The results are compared to the exact full-CI values and also to those of the previous single-facet GCF expansion form.
González-García, Estefanía; Maly, Marek; de la Mata, Francisco Javier; Gómez, Rafael; Marina, María Luisa; García, María Concepción
2016-11-01
Protein sample preparation is a critical and an unsustainable step since it involves the use of tedious methods that usually require high amount of solvents. The development of new materials offers additional opportunities in protein sample preparation. This work explores, for the first time, the potential application of carboxylate-terminated carbosilane dendrimers to the purification/enrichment of proteins. Studies on dendrimer binding to proteins, based on protein fluorescence intensity and emission wavelengths measurements, demonstrated the interaction between carboxylate-terminated carbosilane dendrimers and proteins at all tested pH levels. Interactions were greatly affected by the protein itself, pH, and dendrimer concentration and generation. Especially interesting was the interaction at acidic pH since it resulted in a significant protein precipitation. Dendrimer-protein interactions were modeled observing stable complexes for all proteins. Carboxylate-terminated carbosilane dendrimers at acidic pH were successfully used in the purification/enrichment of proteins extracted from a complex sample. Graphical Abstract Images showing the growing turbidity of solutions containing a mixture of proteins (lysozyme, myoglobin, and BSA) at different protein:dendrimer ratios (1:0, 1:1, 1:8, and 1:20) at acidic pH and SDS-PAGE profiles of the corresponsing supernatants. Comparison of SDS-PAGE profiles for the pellets obtained during the purification of proteins present in a complex sample using a conventional "no-clean" method based on acetone precipitation and the proposed "greener" method using carboxylate-terminated carbosilane dendrimer at a 1:20 protein:dendrimer ratio.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oh-ishi, Katsuyoshi, E-mail: oh-ishi@kc.chuo-u.ac.jp; Nagumo, Kenta; Tateishi, Kazuya
Mo-Re-C compounds containing Mo{sub 7}Re{sub 13}C with the β-Mn structure were synthesized with high-melting-temperature metals Mo, Re, and C powders using a conventional solid state method with a planetary ball milling machine instead of the arc melting method. Use of the ball milling machine was necessary to obtain Mo{sub 7}Re{sub 13}C with the β-Mn structure using the solid state method. Almost single-phase Mo{sub 7}Re{sub 13}C with a trace of impurity were obtained using the synthesis method. By XRF and lattice parameter measurements on the samples, Fe element existed in the compound synthesized using the planetary ball milling machine with amore » pot and balls made of steel, though Fe element was not detected in the compound synthesized using a pot and balls made of tungsten carbide. The former compound containg the Fe atom did not show superconductivity but the latter compound without the Fe atom showed superconductivity at 6.1 K. - Graphical abstract: Temperature dependence of the magnetic susceptibility measured under 10 Oe for the superconducting PBM-T samples without Fe element and non-superconducting PBM-S with Fe element. The inset is the enlarged view of the data for the PBM-S sample.« less
NASA Astrophysics Data System (ADS)
Szafranko, Elżbieta
2017-10-01
Assessment of variant solutions developed for a building investment project needs to be made at the stage of planning. While considering alternative solutions, the investor defines various criteria, but a direct evaluation of the degree of their fulfilment by developed variant solutions can be very difficult. In practice, there are different methods which enable the user to include a large number of parameters into an analysis, but their implementation can be challenging. Some methods require advanced mathematical computations, preceded by complicating input data processing, and the generated results may not lend themselves easily to interpretation. Hence, during her research, the author has developed a systemic approach, which involves several methods and whose goal is to compare their outcome. The final stage of the proposed method consists of graphic interpretation of results. The method has been tested on a variety of building and development projects.
Miller, C L; Hill, D J; Quester, P G; Hiller, J E
2009-01-01
Background: In March 2006, Australia introduced graphic pictorial warnings on cigarette packets. For the first time, packs include the Quitline number. Objective: To measure the combined effect of graphic cigarette pack warnings and printing the Quitline number on packs on calls to the Australian Quitline service. Methods: Calls to the Australian Quitline were monitored over 4 years, 2 years before and after the new packets were introduced. Results: There were twice as many calls to the Quitline in 2006 (the year of introduction), as there were in each of the preceding 2 years. The observed increase in calls exceeds that explained by the accompanying television advertising alone. While call volume tapered back in 2007, it remained at a level higher than before the introduction of new packets. No change was observed in the proportion of first time callers. Conclusion: Introducing graphic cigarette packet warnings and the Quitline number on cigarette packets boosts demand for Quitline services, with likely flow on effects to cessation. PMID:19211613
AnthropMMD: An R package with a graphical user interface for the mean measure of divergence.
Santos, Frédéric
2018-01-01
The mean measure of divergence is a dissimilarity measure between groups of individuals described by dichotomous variables. It is well suited to datasets with many missing values, and it is generally used to compute distance matrices and represent phenograms. Although often used in biological anthropology and archaeozoology, this method suffers from a lack of implementation in common statistical software. A package for the R statistical software, AnthropMMD, is presented here. Offering a dynamic graphical user interface, it is the first one dedicated to Smith's mean measure of divergence. The package also provides facilities for graphical representations and the crucial step of trait selection, so that the entire analysis can be performed through the graphical user interface. Its use is demonstrated using an artificial dataset, and the impact of trait selection is discussed. Finally, AnthropMMD is compared to three other free tools available for calculating the mean measure of divergence, and is proven to be consistent with them. © 2017 Wiley Periodicals, Inc.
SIMD Optimization of Linear Expressions for Programmable Graphics Hardware
Bajaj, Chandrajit; Ihm, Insung; Min, Jungki; Oh, Jinsang
2009-01-01
The increased programmability of graphics hardware allows efficient graphical processing unit (GPU) implementations of a wide range of general computations on commodity PCs. An important factor in such implementations is how to fully exploit the SIMD computing capacities offered by modern graphics processors. Linear expressions in the form of ȳ = Ax̄ + b̄, where A is a matrix, and x̄, ȳ and b̄ are vectors, constitute one of the most basic operations in many scientific computations. In this paper, we propose a SIMD code optimization technique that enables efficient shader codes to be generated for evaluating linear expressions. It is shown that performance can be improved considerably by efficiently packing arithmetic operations into four-wide SIMD instructions through reordering of the operations in linear expressions. We demonstrate that the presented technique can be used effectively for programming both vertex and pixel shaders for a variety of mathematical applications, including integrating differential equations and solving a sparse linear system of equations using iterative methods. PMID:19946569
NASA Astrophysics Data System (ADS)
Lanzagorta, Marco O.; Gomez, Richard B.; Uhlmann, Jeffrey K.
2003-08-01
In recent years, computer graphics has emerged as a critical component of the scientific and engineering process, and it is recognized as an important computer science research area. Computer graphics are extensively used for a variety of aerospace and defense training systems and by Hollywood's special effects companies. All these applications require the computer graphics systems to produce high quality renderings of extremely large data sets in short periods of time. Much research has been done in "classical computing" toward the development of efficient methods and techniques to reduce the rendering time required for large datasets. Quantum Computing's unique algorithmic features offer the possibility of speeding up some of the known rendering algorithms currently used in computer graphics. In this paper we discuss possible implementations of quantum rendering algorithms. In particular, we concentrate on the implementation of Grover's quantum search algorithm for Z-buffering, ray-tracing, radiosity, and scene management techniques. We also compare the theoretical performance between the classical and quantum versions of the algorithms.
Analysis of the upper massif of the craniofacial with the radial method – practical use
Lepich, Tomasz; Dąbek, Józefa; Stompel, Daniel; Gielecki, Jerzy S.
2011-01-01
Introduction The analysis of the upper massif of the craniofacial (UMC) is widely used in many fields of science. The aim of the study was to create a high resolution computer system based on a digital information record and on vector graphics, that could enable dimension measuring and evaluation of craniofacial shape using the radial method. Material and methods The study was carried out on 184 skulls, in a good state of preservation, from the early middle ages. The examined skulls were fixed into Molisson's craniostat in the author's own modification. They were directed in space towards the Frankfurt plane and photographed in frontal norm with a digital camera. The parameters describing the plane and dimensional structure of the UMC and orbits were obtained thanks to the computer analysis of the function recordings picturing the craniofacial structures and using software combining raster graphics with vector graphics. Results It was compared mean values of both orbits separately for male and female groups. In female skulls the comparison of the left and right side did not show statistically significant differences. In male group, higher values were observed for the right side. Only the circularity index presented higher values for the left side. Conclusions Computer graphics with the software used for analysing digital pictures of UMC and orbits increase the precision of measurements as well as the calculation possibilities. Recognition of the face in the post mortem examination is crucial for those working on identification in anthropology and criminology laboratories. PMID:22291834