Remote information service access system based on a client-server-service model
Konrad, Allan M.
1996-01-01
A local host computing system, a remote host computing system as connected by a network, and service functionalities: a human interface service functionality, a starter service functionality, and a desired utility service functionality, and a Client-Server-Service (CSS) model is imposed on each service functionality. In one embodiment, this results in nine logical components and three physical components (a local host, a remote host, and an intervening network), where two of the logical components are integrated into one Remote Object Client component, and that Remote Object Client component and the other seven logical components are deployed among the local host and remote host in a manner which eases compatibility and upgrade problems, and provides an illusion to a user that a desired utility service supported on a remote host resides locally on the user's local host, thereby providing ease of use and minimal software maintenance for users of that remote service.
Remote information service access system based on a client-server-service model
Konrad, A.M.
1997-12-09
A local host computing system, a remote host computing system as connected by a network, and service functionalities: a human interface service functionality, a starter service functionality, and a desired utility service functionality, and a Client-Server-Service (CSS) model is imposed on each service functionality. In one embodiment, this results in nine logical components and three physical components (a local host, a remote host, and an intervening network), where two of the logical components are integrated into one Remote Object Client component, and that Remote Object Client component and the other seven logical components are deployed among the local host and remote host in a manner which eases compatibility and upgrade problems, and provides an illusion to a user that a desired utility service supported on a remote host resides locally on the user`s local host, thereby providing ease of use and minimal software maintenance for users of that remote service. 16 figs.
Remote information service access system based on a client-server-service model
Konrad, Allan M.
1999-01-01
A local host computing system, a remote host computing system as connected by a network, and service functionalities: a human interface service functionality, a starter service functionality, and a desired utility service functionality, and a Client-Server-Service (CSS) model is imposed on each service functionality. In one embodiment, this results in nine logical components and three physical components (a local host, a remote host, and an intervening network), where two of the logical components are integrated into one Remote Object Client component, and that Remote Object Client component and the other seven logical components are deployed among the local host and remote host in a manner which eases compatibility and upgrade problems, and provides an illusion to a user that a desired utility service supported on a remote host resides locally on the user's local host, thereby providing ease of use and minimal software maintenance for users of that remote service.
Remote information service access system based on a client-server-service model
Konrad, A.M.
1996-08-06
A local host computing system, a remote host computing system as connected by a network, and service functionalities: a human interface service functionality, a starter service functionality, and a desired utility service functionality, and a Client-Server-Service (CSS) model is imposed on each service functionality. In one embodiment, this results in nine logical components and three physical components (a local host, a remote host, and an intervening network), where two of the logical components are integrated into one Remote Object Client component, and that Remote Object Client component and the other seven logical components are deployed among the local host and remote host in a manner which eases compatibility and upgrade problems, and provides an illusion to a user that a desired utility service supported on a remote host resides locally on the user`s local host, thereby providing ease of use and minimal software maintenance for users of that remote service. 16 figs.
Remote information service access system based on a client-server-service model
Konrad, Allan M.
1997-01-01
A local host computing system, a remote host computing system as connected by a network, and service functionalities: a human interface service functionality, a starter service functionality, and a desired utility service functionality, and a Client-Server-Service (CSS) model is imposed on each service functionality. In one embodiment, this results in nine logical components and three physical components (a local host, a remote host, and an intervening network), where two of the logical components are integrated into one Remote Object Client component, and that Remote Object Client component and the other seven logical components are deployed among the local host and remote host in a manner which eases compatibility and upgrade problems, and provides an illusion to a user that a desired utility service supported on a remote host resides locally on the user's local host, thereby providing ease of use and minimal software maintenance for users of that remote service.
KNET - DISTRIBUTED COMPUTING AND/OR DATA TRANSFER PROGRAM
NASA Technical Reports Server (NTRS)
Hui, J.
1994-01-01
KNET facilitates distributed computing between a UNIX compatible local host and a remote host which may or may not be UNIX compatible. It is capable of automatic remote login. That is, it performs on the user's behalf the chore of handling host selection, user name, and password to the designated host. Once the login has been successfully completed, the user may interactively communicate with the remote host. Data output from the remote host may be directed to the local screen, to a local file, and/or to a local process. Conversely, data input from the keyboard, a local file, or a local process may be directed to the remote host. KNET takes advantage of the multitasking and terminal mode control features of the UNIX operating system. A parent process is used as the upper layer for interfacing with the local user. A child process is used for a lower layer for interfacing with the remote host computer, and optionally one or more child processes can be used for the remote data output. Output may be directed to the screen and/or to the local processes under the control of a data pipe switch. In order for KNET to operate, the local and remote hosts must observe a common communications protocol. KNET is written in ANSI standard C-language for computers running UNIX. It has been successfully implemented on several Sun series computers and a DECstation 3100 and used to run programs remotely on VAX VMS and UNIX based computers. It requires 100K of RAM under SunOS and 120K of RAM under DEC RISC ULTRIX. An electronic copy of the documentation is provided on the distribution medium. The standard distribution medium for KNET is a .25 inch streaming magnetic tape cartridge in UNIX tar format. It is also available on a 3.5 inch diskette in UNIX tar format. KNET was developed in 1991 and is a copyrighted work with all copyright vested in NASA. UNIX is a registered trademark of AT&T Bell Laboratories. Sun and SunOS are trademarks of Sun Microsystems, Inc. DECstation, VAX, VMS, and ULTRIX are trademarks of Digital Equipment Corporation.
General-Purpose Serial Interface For Remote Control
NASA Technical Reports Server (NTRS)
Busquets, Anthony M.; Gupton, Lawrence E.
1990-01-01
Computer controls remote television camera. General-purpose controller developed to serve as interface between host computer and pan/tilt/zoom/focus functions on series of automated video cameras. Interface port based on 8251 programmable communications-interface circuit configured for tristated outputs, and connects controller system to any host computer with RS-232 input/output (I/O) port. Accepts byte-coded data from host, compares them with prestored codes in read-only memory (ROM), and closes or opens appropriate switches. Six output ports control opening and closing of as many as 48 switches. Operator controls remote television camera by speaking commands, in system including general-purpose controller.
Monitoring system including an electronic sensor platform and an interrogation transceiver
Kinzel, Robert L.; Sheets, Larry R.
2003-09-23
A wireless monitoring system suitable for a wide range of remote data collection applications. The system includes at least one Electronic Sensor Platform (ESP), an Interrogator Transceiver (IT) and a general purpose host computer. The ESP functions as a remote data collector from a number of digital and analog sensors located therein. The host computer provides for data logging, testing, demonstration, installation checkout, and troubleshooting of the system. The IT transmits signals from one or more ESP's to the host computer to the ESP's. The IT host computer may be powered by a common power supply, and each ESP is individually powered by a battery. This monitoring system has an extremely low power consumption which allows remote operation of the ESP for long periods; provides authenticated message traffic over a wireless network; utilizes state-of-health and tamper sensors to ensure that the ESP is secure and undamaged; has robust housing of the ESP suitable for use in radiation environments; and is low in cost. With one base station (host computer and interrogator transceiver), multiple ESP's may be controlled at a single monitoring site.
Microcomputer software development facilities
NASA Technical Reports Server (NTRS)
Gorman, J. S.; Mathiasen, C.
1980-01-01
A more efficient and cost effective method for developing microcomputer software is to utilize a host computer with high-speed peripheral support. Application programs such as cross assemblers, loaders, and simulators are implemented in the host computer for each of the microcomputers for which software development is a requirement. The host computer is configured to operate in a time share mode for multiusers. The remote terminals, printers, and down loading capabilities provided are based on user requirements. With this configuration a user, either local or remote, can use the host computer for microcomputer software development. Once the software is developed (through the code and modular debug stage) it can be downloaded to the development system or emulator in a test area where hardware/software integration functions can proceed. The microcomputer software program sources reside in the host computer and can be edited, assembled, loaded, and then downloaded as required until the software development project has been completed.
Virtual Computing Laboratories: A Case Study with Comparisons to Physical Computing Laboratories
ERIC Educational Resources Information Center
Burd, Stephen D.; Seazzu, Alessandro F.; Conway, Christopher
2009-01-01
Current technology enables schools to provide remote or virtual computing labs that can be implemented in multiple ways ranging from remote access to banks of dedicated workstations to sophisticated access to large-scale servers hosting virtualized workstations. This paper reports on the implementation of a specific lab using remote access to…
Design of Remote GPRS-based Gas Data Monitoring System
NASA Astrophysics Data System (ADS)
Yan, Xiyue; Yang, Jianhua; Lu, Wei
2018-01-01
In order to solve the problem of remote data transmission of gas flowmeter, and realize unattended operation on the spot, an unattended remote monitoring system based on GPRS for gas data is designed in this paper. The slave computer of this system adopts embedded microprocessor to read data of gas flowmeter through rs-232 bus and transfers it to the host computer through DTU. In the host computer, the VB program dynamically binds the Winsock control to receive and parse data. By using dynamic data exchange, the Kingview configuration software realizes history trend curve, real time trend curve, alarm, print, web browsing and other functions.
The Remote Analysis Station (RAS) as an instructional system
NASA Technical Reports Server (NTRS)
Rogers, R. H.; Wilson, C. L.; Dye, R. H.; Jaworski, E.
1981-01-01
"Hands-on" training in LANDSAT data analysis techniques can be obtained using a desk-top, interactive remote analysis station (RAS) which consists of a color CRT imagery display, with alphanumeric overwrite and keyboard, as well as a cursor controller and modem. This portable station can communicate via modem and dial-up telephone with a host computer at 1200 baud or it can be hardwired to a host computer at 9600 baud. A Z80 microcomputer controls the display refresh memory and remote station processing. LANDSAT data is displayed as three-band false-color imagery, one-band color-sliced imagery, or color-coded processed imagery. Although the display memory routinely operates at 256 x 256 picture elements, a display resolution of 128 x 128 can be selected to fill the display faster. In the false color mode the computer packs the data into one 8-bit character. When the host is not sending pictorial information the characters sent are in ordinary ASCII code. System capabilities are described.
NASA Technical Reports Server (NTRS)
Peri, Frank, Jr.
1992-01-01
A flight digital data acquisition system that uses the MIL-STD-1553B bus for transmission of data to a host computer for control law processing is described. The instrument, the Remote Interface Unit (RIU), can accommodate up to 16 input channels and eight output channels. The RIU employs a digital signal processor to perform local digital filtering before sending data to the host. The system allows flexible sensor and actuator data organization to facilitate quick control law computations on the host computer. The instrument can also run simple control laws autonomously without host intervention. The RIU and host computer together have replaced a similar larger, ground minicomputer system with favorable results.
Design of Remote Monitoring System of Irrigation based on GSM and ZigBee Technology
NASA Astrophysics Data System (ADS)
Xiao xi, Zheng; Fang, Zhao; Shuaifei, Shao
2018-03-01
To solve the problems of low level of irrigation and waste of water resources, a remote monitoring system for farmland irrigation based on GSM communication technology and ZigBee technology was designed. The system is composed of sensors, GSM communication module, ZigBee module, host computer, valve and so on. The system detects and closes the pump and the electromagnetic valve according to the need of the system, and transmits the monitoring information to the host computer or the user’s Mobile phone through the GSM communication network. Experiments show that the system has low power consumption, friendly man-machine interface, convenient and simple. It can monitor agricultural environment remotely and control related irrigation equipment at any time and place, and can better meet the needs of remote monitoring of farmland irrigation.
Serial network simplifies the design of multiple microcomputer systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Folkes, D.
1981-01-01
Recently there has been a lot of interest in developing network communication schemes for carrying digital data between locally distributed computing stations. Many of these schemes have focused on distributed networking techniques for data processing applications. These applications suggest the use of a serial, multipoint bus, where a number of remote intelligent units act as slaves to a central or host computer. Each slave would be serially addressable from the host and would perform required operations upon being addressed by the host. Based on an MK3873 single-chip microcomputer, the SCU 20 is designed to be such a remote slave device.more » The capabilities of the SCU 20 and its use in systems applications are examined.« less
Remote observing with NASA's Deep Space Network
NASA Astrophysics Data System (ADS)
Kuiper, T. B. H.; Majid, W. A.; Martinez, S.; Garcia-Miro, C.; Rizzo, J. R.
2012-09-01
The Deep Space Network (DSN) communicates with spacecraft as far away as the boundary between the Solar System and the interstellar medium. To make this possible, large sensitive antennas at Canberra, Australia, Goldstone, California, and Madrid, Spain, provide for constant communication with interplanetary missions. We describe the procedures for radioastronomical observations using this network. Remote access to science monitor and control computers by authorized observers is provided by two-factor authentication through a gateway at the Jet Propulsion Laboratory (JPL) in Pasadena. To make such observations practical, we have devised schemes based on SSH tunnels and distributed computing. At the very minimum, one can use SSH tunnels and VNC (Virtual Network Computing, a remote desktop software suite) to control the science hosts within the DSN Flight Operations network. In this way we have controlled up to three telescopes simultaneously. However, X-window updates can be slow and there are issues involving incompatible screen sizes and multi-screen displays. Consequently, we are now developing SSH tunnel-based schemes in which instrument control and monitoring, and intense data processing, are done on-site by the remote DSN hosts while data manipulation and graphical display are done at the observer's host. We describe our approaches to various challenges, our experience with what worked well and lessons learned, and directions for future development.
Definition and maintenance of a telemetry database dictionary
NASA Technical Reports Server (NTRS)
Knopf, William P. (Inventor)
2007-01-01
A telemetry dictionary database includes a component for receiving spreadsheet workbooks of telemetry data over a web-based interface from other computer devices. Another component routes the spreadsheet workbooks to a specified directory on the host processing device. A process then checks the received spreadsheet workbooks for errors, and if no errors are detected the spreadsheet workbooks are routed to another directory to await initiation of a remote database loading process. The loading process first converts the spreadsheet workbooks to comma separated value (CSV) files. Next, a network connection with the computer system that hosts the telemetry dictionary database is established and the CSV files are ported to the computer system that hosts the telemetry dictionary database. This is followed by a remote initiation of a database loading program. Upon completion of loading a flatfile generation program is manually initiated to generate a flatfile to be used in a mission operations environment by the core ground system.
Okamoto, E; Shimanaka, M; Suzuki, S; Baba, K; Mitamura, Y
1999-01-01
The usefulness of a remote monitoring system that uses a personal handy phone for artificial heart implanted patients was investigated. The type of handy phone used in this study was a personal handy phone system (PHS), which is a system developed in Japan that uses the NTT (Nippon Telephone and Telegraph, Inc.) telephone network service. The PHS has several advantages: high-speed data transmission, low power output, little electromagnetic interference with medical devices, and easy locating of patients. In our system, patients have a mobile computer (Toshiba, Libretto 50, Kawasaki, Japan) for data transmission control between an implanted controller and a host computer (NEC, PC-9821V16) in the hospital. Information on the motor rotational angle (8 bits) and motor current (8 bits) of the implanted motor driven heart is fed into the mobile computer from the implanted controller (Hitachi, H8/532, Yokohama, Japan) according to 32-bit command codes from the host computer. Motor current and motor rotational angle data from inside the body are framed together by a control code (frame number and parity) for data error checking and correcting at the receiving site, and the data are sent through the PHS connection to the mobile computer. The host computer calculates pump outflow and arterial pressure from the motor rotational angle and motor current values and displays the data in real-time waveforms. The results of this study showed that accurate data on motor rotational angle and current could be transmitted from the subjects while they were walking or driving a car to the host computer at a data transmission rate of 9600 bps. This system is useful for remote monitoring of patients with an implanted artificial heart.
Cooperative processing user interfaces for AdaNET
NASA Technical Reports Server (NTRS)
Gutzmann, Kurt M.
1991-01-01
A cooperative processing user interface (CUI) system shares the task of graphical display generation and presentation between the user's computer and a remote host. The communications link between the two computers is typically a modem or Ethernet. The two main purposes of a CUI are reduction of the amount of data transmitted between user and host machines, and provision of a graphical user interface system to make the system easier to use.
Quantum processing by remote quantum control
NASA Astrophysics Data System (ADS)
Qiang, Xiaogang; Zhou, Xiaoqi; Aungskunsiri, Kanin; Cable, Hugo; O'Brien, Jeremy L.
2017-12-01
Client-server models enable computations to be hosted remotely on quantum servers. We present a novel protocol for realizing this task, with practical advantages when using technology feasible in the near term. Client tasks are realized as linear combinations of operations implemented by the server, where the linear coefficients are hidden from the server. We report on an experimental demonstration of our protocol using linear optics, which realizes linear combination of two single-qubit operations by a remote single-qubit control. In addition, we explain when our protocol can remain efficient for larger computations, as well as some ways in which privacy can be maintained using our protocol.
Documentary of MFENET, a national computer network
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shuttleworth, B.O.
1977-06-01
The national Magnetic Fusion Energy Computer Network (MFENET) is a newly operational star network of geographically separated heterogeneous hosts and a communications subnetwork of PDP-11 processors. Host processors interfaced to the subnetwork currently include a CDC 7600 at the Central Computer Center (CCC) and several DECsystem-10's at User Service Centers (USC's). The network was funded by a U.S. government agency (ERDA) to provide in an economical manner the needed computational resources to magnetic confinement fusion researchers. Phase I operation of MFENET distributed the processing power of the CDC 7600 among the USC's through the provision of file transport between anymore » two hosts and remote job entry to the 7600. Extending the capabilities of Phase I, MFENET Phase II provided interactive terminal access to the CDC 7600 from the USC's. A file management system is maintained at the CCC for all network users. The history and development of MFENET are discussed, with emphasis on the protocols used to link the host computers and the USC software. Comparisons are made of MFENET versus ARPANET (Advanced Research Projects Agency Computer Network) and DECNET (Digital Distributed Network Architecture). DECNET and MFENET host-to host, host-to-CCP, and link protocols are discussed in detail. The USC--CCP interface is described briefly. 43 figures, 2 tables.« less
A software control system for the ACTS high-burst-rate link evaluation terminal
NASA Technical Reports Server (NTRS)
Reinhart, Richard C.; Daugherty, Elaine S.
1991-01-01
Control and performance monitoring of NASA's High Burst Rate Link Evaluation Terminal (HBR-LET) is accomplished by using several software control modules. Different software modules are responsible for controlling remote radio frequency (RF) instrumentation, supporting communication between a host and a remote computer, controlling the output power of the Link Evaluation Terminal and data display. Remote commanding of microwave RF instrumentation and the LET digital ground terminal allows computer control of various experiments, including bit error rate measurements. Computer communication allows system operators to transmit and receive from the Advanced Communications Technology Satellite (ACTS). Finally, the output power control software dynamically controls the uplink output power of the terminal to compensate for signal loss due to rain fade. Included is a discussion of each software module and its applications.
Robotic laboratory for distance education
NASA Astrophysics Data System (ADS)
Luciano, Sarah C.; Kost, Alan R.
2016-09-01
This project involves the construction of a remote-controlled laboratory experiment that can be accessed by online students. The project addresses a need to provide a laboratory experience for students who are taking online courses to be able to provide an in-class experience. The chosen task for the remote user is an optical engineering experiment, specifically aligning a spatial filter. We instrument the physical laboratory set up in Tucson, AZ at the University of Arizona. The hardware in the spatial filter experiment is augmented by motors and cameras to allow the user to remotely control the hardware. The user interacts with a software on their computer, which communicates with a server via Internet connection to the host computer in the Optics Laboratory at the University of Arizona. Our final overall system is comprised of several subsystems. These are the optical experiment set-up, which is a spatial filter experiment; the mechanical subsystem, which interfaces the motors with the micrometers to move the optical hardware; the electrical subsystem, which allows for the electrical communications from the remote computer to the host computer to the hardware; and finally the software subsystem, which is the means by which messages are communicated throughout the system. The goal of the project is to convey as much of an in-lab experience as possible by allowing the user to directly manipulate hardware and receive visual feedback in real-time. Thus, the remote user is able to learn important concepts from this particular experiment and is able to connect theory to the physical world by actually seeing the outcome of a procedure. The latter is a learning experience that is often lost with distance learning and is one that this project hopes to provide.
NASA Astrophysics Data System (ADS)
Li, J.; Zhang, T.; Huang, Q.; Liu, Q.
2014-12-01
Today's climate datasets are featured with large volume, high degree of spatiotemporal complexity and evolving fast overtime. As visualizing large volume distributed climate datasets is computationally intensive, traditional desktop based visualization applications fail to handle the computational intensity. Recently, scientists have developed remote visualization techniques to address the computational issue. Remote visualization techniques usually leverage server-side parallel computing capabilities to perform visualization tasks and deliver visualization results to clients through network. In this research, we aim to build a remote parallel visualization platform for visualizing and analyzing massive climate data. Our visualization platform was built based on Paraview, which is one of the most popular open source remote visualization and analysis applications. To further enhance the scalability and stability of the platform, we have employed cloud computing techniques to support the deployment of the platform. In this platform, all climate datasets are regular grid data which are stored in NetCDF format. Three types of data access methods are supported in the platform: accessing remote datasets provided by OpenDAP servers, accessing datasets hosted on the web visualization server and accessing local datasets. Despite different data access methods, all visualization tasks are completed at the server side to reduce the workload of clients. As a proof of concept, we have implemented a set of scientific visualization methods to show the feasibility of the platform. Preliminary results indicate that the framework can address the computation limitation of desktop based visualization applications.
Remote canopy hemispherical image collection system
NASA Astrophysics Data System (ADS)
Wan, Xuefen; Liu, Bingyu; Yang, Yi; Han, Fang; Cui, Jian
2016-11-01
Canopies are major part of plant photosynthesis and have distinct architectural elements such as tree crowns, whorls, branches, shoots, etc. By measuring canopy structural parameters, the solar radiation interception, photosynthesis effects and the spatio-temporal distribution of solar radiation under the canopy can be evaluated. Among canopy structure parameters, Leaf Area Index (LAI) is the key one. Leaf area index is a crucial variable in agronomic and environmental studies, because of its importance for estimating the amount of radiation intercepted by the canopy and the crop water requirements. The LAI can be achieved by hemispheric images which are obtained below the canopy with high accuracy and effectiveness. But existing hemispheric images canopy-LAI measurement technique is based on digital SLR camera with a fisheye lens. Users need to collect hemispheric image manually. The SLR camera with fisheye lens is not suit for long-term canopy-LAI outdoor measurement too. And the high cost of SLR limits its capacity. In recent years, with the development of embedded system and image processing technology, low cost remote canopy hemispheric image acquisition technology is becoming possible. In this paper, we present a remote hemispheric canopy image acquisition system with in-field/host configuration. In-field node based on imbed platform, low cost image sensor and fisheye lens is designed to achieve hemispherical image of plant canopy at distance with low cost. Solar radiation and temperature/humidity data, which are important for evaluating image data validation, are obtained for invalid hemispherical image elimination and node maintenance too. Host computer interacts with in-field node by 3G network. The hemispherical image calibration and super resolution are used to improve image quality in host computer. Results show that the remote canopy image collection system can make low cost remote canopy image acquisition for LAI effectively. It will be a potential technology candidate for low-cost remote canopy hemispherical image collection to measure canopy LAI.
Design of temperature monitoring system based on CAN bus
NASA Astrophysics Data System (ADS)
Zhang, Li
2017-10-01
The remote temperature monitoring system based on the Controller Area Network (CAN) bus is designed to collect the multi-node remote temperature. By using the STM32F103 as main controller and multiple DS18B20s as temperature sensors, the system achieves a master-slave node data acquisition and transmission based on the CAN bus protocol. And making use of the serial port communication technology to communicate with the host computer, the system achieves the function of remote temperature storage, historical data show and the temperature waveform display.
Communication-Gateway Software For NETEX, DECnet, And TCP/IP
NASA Technical Reports Server (NTRS)
Keith, B.; Ferry, D.; Fendler, E.
1990-01-01
Communications gateway software, GATEWAY, provides process-to-process communication between remote applications programs in different protocol domains. Communicating peer processes may be resident on any paired combination of NETEX, DECnet, or TCP/IP hosts. Provides necessary mapping from one protocol to another and facilitates practical intermachine communications in cost-effective manner by eliminating need to standardize on single protocol or to implement multiple protocols in host computers. Written in Ada.
Evaluating virtual hosted desktops for graphics-intensive astronomy
NASA Astrophysics Data System (ADS)
Meade, B. F.; Fluke, C. J.
2018-04-01
Visualisation of data is critical to understanding astronomical phenomena. Today, many instruments produce datasets that are too big to be downloaded to a local computer, yet many of the visualisation tools used by astronomers are deployed only on desktop computers. Cloud computing is increasingly used to provide a computation and simulation platform in astronomy, but it also offers great potential as a visualisation platform. Virtual hosted desktops, with graphics processing unit (GPU) acceleration, allow interactive, graphics-intensive desktop applications to operate co-located with astronomy datasets stored in remote data centres. By combining benchmarking and user experience testing, with a cohort of 20 astronomers, we investigate the viability of replacing physical desktop computers with virtual hosted desktops. In our work, we compare two Apple MacBook computers (one old and one new, representing hardware and opposite ends of the useful lifetime) with two virtual hosted desktops: one commercial (Amazon Web Services) and one in a private research cloud (the Australian NeCTAR Research Cloud). For two-dimensional image-based tasks and graphics-intensive three-dimensional operations - typical of astronomy visualisation workflows - we found that benchmarks do not necessarily provide the best indication of performance. When compared to typical laptop computers, virtual hosted desktops can provide a better user experience, even with lower performing graphics cards. We also found that virtual hosted desktops are equally simple to use, provide greater flexibility in choice of configuration, and may actually be a more cost-effective option for typical usage profiles.
Omaki, Elise; Rizzutti, Nicholas; Shields, Wendy; Zhu, Jeffrey; McDonald, Eileen; Stevens, Martha W; Gielen, Andrea
2017-04-01
The aims of this literature review are to (1) summarise how computer and mobile technology-based health behaviour change applications have been evaluated in unintentional injury prevention, (2) describe how these successes can be applied to injury-prevention programmes in the future and (3) identify research gaps. Studies included in this systematic review were education and behaviour change intervention trials and programme evaluations in which the intervention was delivered by either a computer or mobile technology and addressed an unintentional injury prevention topic. Articles were limited to those published in English and after 1990. Among the 44 technology-based injury-prevention studies included in this review, 16 studies evaluated locally hosted software programmes, 4 studies offered kiosk-based programmes, 11 evaluated remotely hosted internet programmes, 2 studies used mobile technology or portable devices and 11 studies evaluated virtual-reality interventions. Locally hosted software programmes and remotely hosted internet programmes consistently increased knowledge and behaviours. Kiosk programmes showed evidence of modest knowledge and behaviour gains. Both programmes using mobile technology improved behaviours. Virtual-reality programmes consistently improved behaviours, but there were little gains in knowledge. No studies evaluated text-messaging programmes dedicated to injury prevention. There is much potential for computer-based programmes to be used for injury-prevention behaviour change. The reviewed studies provide evidence that computer-based communication is effective in conveying information and influencing how participants think about an injury topic and adopt safety behaviours. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
An Interactive Web-Based Analysis Framework for Remote Sensing Cloud Computing
NASA Astrophysics Data System (ADS)
Wang, X. Z.; Zhang, H. M.; Zhao, J. H.; Lin, Q. H.; Zhou, Y. C.; Li, J. H.
2015-07-01
Spatiotemporal data, especially remote sensing data, are widely used in ecological, geographical, agriculture, and military research and applications. With the development of remote sensing technology, more and more remote sensing data are accumulated and stored in the cloud. An effective way for cloud users to access and analyse these massive spatiotemporal data in the web clients becomes an urgent issue. In this paper, we proposed a new scalable, interactive and web-based cloud computing solution for massive remote sensing data analysis. We build a spatiotemporal analysis platform to provide the end-user with a safe and convenient way to access massive remote sensing data stored in the cloud. The lightweight cloud storage system used to store public data and users' private data is constructed based on open source distributed file system. In it, massive remote sensing data are stored as public data, while the intermediate and input data are stored as private data. The elastic, scalable, and flexible cloud computing environment is built using Docker, which is a technology of open-source lightweight cloud computing container in the Linux operating system. In the Docker container, open-source software such as IPython, NumPy, GDAL, and Grass GIS etc., are deployed. Users can write scripts in the IPython Notebook web page through the web browser to process data, and the scripts will be submitted to IPython kernel to be executed. By comparing the performance of remote sensing data analysis tasks executed in Docker container, KVM virtual machines and physical machines respectively, we can conclude that the cloud computing environment built by Docker makes the greatest use of the host system resources, and can handle more concurrent spatial-temporal computing tasks. Docker technology provides resource isolation mechanism in aspects of IO, CPU, and memory etc., which offers security guarantee when processing remote sensing data in the IPython Notebook. Users can write complex data processing code on the web directly, so they can design their own data processing algorithm.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 47 Telecommunication 2 2011-10-01 2011-10-01 false Host/remote message Cable and Wire Facilities... TELECOMMUNICATIONS COMPANIES 1 Telecommunications Property Cable and Wire Facilities § 36.157 Host/remote message Cable and Wire Facilities (C&WF)—Category 4—apportionment procedures. (a) Host/Remote Message C&WF...
Code of Federal Regulations, 2010 CFR
2010-10-01
... 47 Telecommunication 2 2010-10-01 2010-10-01 false Host/remote message Cable and Wire Facilities... TELECOMMUNICATIONS COMPANIES 1 Telecommunications Property Cable and Wire Facilities § 36.157 Host/remote message Cable and Wire Facilities (C&WF)—Category 4—apportionment procedures. (a) Host/Remote Message C&WF...
The Defense Message System and the U.S. Coast Guard
1992-06-01
these mail services, the Internet also provides a File Transfer Protocol (FTP) and remote login between host computers (TELNET) capabilities. 17 [Ref...the Joint Maritime Intelligence Element (JMIE), Zincdust, and Emerald . [Ref. 27] 4. Secure Data Network The Coast Guard’s Secure Data Network (SDN
Video PATSEARCH: A Mixed-Media System.
ERIC Educational Resources Information Center
Schulman, Jacque-Lynne
1982-01-01
Describes a videodisc-based information display system in which a computer terminal is used to search the online PATSEARCH database from a remote host with local microcomputer control to select and display drawings from the retrieved records. System features and system components are discussed and criteria for system evaluation are presented.…
Autonomous mobile robot for radiologic surveys
Dudar, A.M.; Wagner, D.G.; Teese, G.D.
1994-06-28
An apparatus is described for conducting radiologic surveys. The apparatus comprises in the main a robot capable of following a preprogrammed path through an area, a radiation monitor adapted to receive input from a radiation detector assembly, ultrasonic transducers for navigation and collision avoidance, and an on-board computer system including an integrator for interfacing the radiation monitor and the robot. Front and rear bumpers are attached to the robot by bumper mounts. The robot may be equipped with memory boards for the collection and storage of radiation survey information. The on-board computer system is connected to a remote host computer via a UHF radio link. The apparatus is powered by a rechargeable 24-volt DC battery, and is stored at a docking station when not in use and/or for recharging. A remote host computer contains a stored database defining paths between points in the area where the robot is to operate, including but not limited to the locations of walls, doors, stationary furniture and equipment, and sonic markers if used. When a program consisting of a series of paths is downloaded to the on-board computer system, the robot conducts a floor survey autonomously at any preselected rate. When the radiation monitor detects contamination, the robot resurveys the area at reduced speed and resumes its preprogrammed path if the contamination is not confirmed. If the contamination is confirmed, the robot stops and sounds an alarm. 5 figures.
Autonomous mobile robot for radiologic surveys
Dudar, Aed M.; Wagner, David G.; Teese, Gregory D.
1994-01-01
An apparatus for conducting radiologic surveys. The apparatus comprises in the main a robot capable of following a preprogrammed path through an area, a radiation monitor adapted to receive input from a radiation detector assembly, ultrasonic transducers for navigation and collision avoidance, and an on-board computer system including an integrator for interfacing the radiation monitor and the robot. Front and rear bumpers are attached to the robot by bumper mounts. The robot may be equipped with memory boards for the collection and storage of radiation survey information. The on-board computer system is connected to a remote host computer via a UHF radio link. The apparatus is powered by a rechargeable 24-volt DC battery, and is stored at a docking station when not in use and/or for recharging. A remote host computer contains a stored database defining paths between points in the area where the robot is to operate, including but not limited to the locations of walls, doors, stationary furniture and equipment, and sonic markers if used. When a program consisting of a series of paths is downloaded to the on-board computer system, the robot conducts a floor survey autonomously at any preselected rate. When the radiation monitor detects contamination, the robot resurveys the area at reduced speed and resumes its preprogrammed path if the contamination is not confirmed. If the contamination is confirmed, the robot stops and sounds an alarm.
Design of cold chain logistics remote monitoring system based on ZigBee and GPS location
NASA Astrophysics Data System (ADS)
Zong, Xiaoping; Shao, Heling
2017-03-01
This paper designed a remote monitoring system based on Bee Zig wireless sensor network and GPS positioning, according to the characteristics of cold chain logistics. The system consisted of the ZigBee network, gateway and monitoring center. ZigBee network temperature acquisition modules and GPS positioning acquisition module were responsible for data collection, and then send the data to the host computer through the GPRS network and Internet to realize remote monitoring of vehicle with functions of login permissions, temperature display, latitude and longitude display, historical data, real-time alarm and so on. Experiments showed that the system is stable, reliable and effective to realize the real-time remote monitoring of the vehicle in the process of cold chain transport.
Remote Neural Pendants In A Welding-Control System
NASA Technical Reports Server (NTRS)
Venable, Richard A.; Bucher, Joseph H.
1995-01-01
Neural network integrated circuits enhance functionalities of both remote terminals (called "pendants") and communication links, without necessitating installation of additional wires in links. Makes possible to incorporate many features into pendant, including real-time display of critical welding parameters and other process information, capability for communication between technician at pendant and host computer or technician elsewhere in system, and switches and potentiometers through which technician at pendant exerts remote control over such critical aspects of welding process as current, voltage, rate of travel, flow of gas, starting, and stopping. Other potential manufacturing applications include control of spray coating and of curing of composite materials. Potential nonmanufacturing uses include remote control of heating, air conditioning, and lighting in electrically noisy and otherwise hostile environments.
A Tool for Parameter-space Explorations
NASA Astrophysics Data System (ADS)
Murase, Yohsuke; Uchitane, Takeshi; Ito, Nobuyasu
A software for managing simulation jobs and results, named "OACIS", is presented. It controls a large number of simulation jobs executed in various remote servers, keeps these results in an organized way, and manages the analyses on these results. The software has a web browser front end, and users can submit various jobs to appropriate remote hosts from a web browser easily. After these jobs are finished, all the result files are automatically downloaded from the computational hosts and stored in a traceable way together with the logs of the date, host, and elapsed time of the jobs. Some visualization functions are also provided so that users can easily grasp the overview of the results distributed in a high-dimensional parameter space. Thus, OACIS is especially beneficial for the complex simulation models having many parameters for which a lot of parameter searches are required. By using API of OACIS, it is easy to write a code that automates parameter selection depending on the previous simulation results. A few examples of the automated parameter selection are also demonstrated.
Watching elderly and disabled person's physical condition by remotely controlled monorail robot
NASA Astrophysics Data System (ADS)
Nagasaka, Yasunori; Matsumoto, Yoshinori; Fukaya, Yasutoshi; Takahashi, Tomoichi; Takeshita, Toru
2001-10-01
We are developing a nursing system using robots and cameras. The cameras are mounted on a remote controlled monorail robot which moves inside a room and watches the elderly. It is necessary to pay attention to the elderly at home or nursing homes all time. This requires staffs to pay attention to them at every time. The purpose of our system is to help those staffs. This study intends to improve such situation. A host computer controls a monorail robot to go in front of the elderly using the images taken by cameras on the ceiling. A CCD camera is mounted on the monorail robot to take pictures of their facial expression or movements. The robot sends the images to a host computer that checks them whether something unusual happens or not. We propose a simple calibration method for positioning the monorail robots to track the moves of the elderly for keeping their faces at center of camera view. We built a small experiment system, and evaluated our camera calibration method and image processing algorithm.
Software development to support sensor control of robot arc welding
NASA Technical Reports Server (NTRS)
Silas, F. R., Jr.
1986-01-01
The development of software for a Digital Equipment Corporation MINC-23 Laboratory Computer to provide functions of a workcell host computer for Space Shuttle Main Engine (SSME) robotic welding is documented. Routines were written to transfer robot programs between the MINC and an Advanced Robotic Cyro 750 welding robot. Other routines provide advanced program editing features while additional software allows communicatin with a remote computer aided design system. Access to special robot functions were provided to allow advanced control of weld seam tracking and process control for future development programs.
The potential and prospects of proximal remote sensing of arthropod pests.
Nansen, Christian
2016-04-01
Bench-top or proximal remote sensing applications are widely used as part of quality control and machine vision systems in commercial operations. In addition, these technologies are becoming increasingly important in insect systematics and studies of insect physiology and pest management. This paper provides a review and discussion of how proximal remote sensing may contribute valuable quantitative information regarding identification of species, assessment of insect responses to insecticides, insect host responses to parasitoids and performance of biological control agents. The future role of proximal remote sensing is discussed as an exciting path for novel paths of multidisciplinary research among entomologists and scientists from a wide range of other disciplines, including image processing engineers, medical engineers, research pharmacists and computer scientists. © 2015 Society of Chemical Industry. © 2015 Society of Chemical Industry.
Distributed run of a one-dimensional model in a regional application using SOAP-based web services
NASA Astrophysics Data System (ADS)
Smiatek, Gerhard
This article describes the setup of a distributed computing system in Perl. It facilitates the parallel run of a one-dimensional environmental model on a number of simple network PC hosts. The system uses Simple Object Access Protocol (SOAP) driven web services offering the model run on remote hosts and a multi-thread environment distributing the work and accessing the web services. Its application is demonstrated in a regional run of a process-oriented biogenic emission model for the area of Germany. Within a network consisting of up to seven web services implemented on Linux and MS-Windows hosts, a performance increase of approximately 400% has been reached compared to a model run on the fastest single host.
Fast, cheap and in control: spectral imaging with handheld devices
NASA Astrophysics Data System (ADS)
Gooding, Edward A.; Deutsch, Erik R.; Huehnerhoff, Joseph; Hajian, Arsen R.
2017-05-01
Remote sensing has moved out of the laboratory and into the real world. Instruments using reflection or Raman imaging modalities become faster, cheaper and more powerful annually. Enabling technologies include virtual slit spectrometer design, high power multimode diode lasers, fast open-loop scanning systems, low-noise IR-sensitive array detectors and low-cost computers with touchscreen interfaces. High-volume manufacturing assembles these components into inexpensive portable or handheld devices that make possible sophisticated decision-making based on robust data analytics. Examples include threat, hazmat and narcotics detection; remote gas sensing; biophotonic screening; environmental remediation and a host of other applications.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Junghyun; Gangwon, Jo; Jaehoon, Jung
Applications written solely in OpenCL or CUDA cannot execute on a cluster as a whole. Most previous approaches that extend these programming models to clusters are based on a common idea: designating a centralized host node and coordinating the other nodes with the host for computation. However, the centralized host node is a serious performance bottleneck when the number of nodes is large. In this paper, we propose a scalable and distributed OpenCL framework called SnuCL-D for large-scale clusters. SnuCL-D's remote device virtualization provides an OpenCL application with an illusion that all compute devices in a cluster are confined inmore » a single node. To reduce the amount of control-message and data communication between nodes, SnuCL-D replicates the OpenCL host program execution and data in each node. We also propose a new OpenCL host API function and a queueing optimization technique that significantly reduce the overhead incurred by the previous centralized approaches. To show the effectiveness of SnuCL-D, we evaluate SnuCL-D with a microbenchmark and eleven benchmark applications on a large-scale CPU cluster and a medium-scale GPU cluster.« less
Single-crystal diffraction instrument TriCS at SINQ
NASA Astrophysics Data System (ADS)
Schefer, J.; Könnecke, M.; Murasik, A.; Czopnik, A.; Strässle, Th; Keller, P.; Schlumpf, N.
2000-03-01
The single-crystal diffractometer TriCS at the Swiss Continuous Spallation Source (SINQ) is presently in the commissioning phase. A two-dimensional wire detector produced by EMBL was delivered in March 1999. The instrument is presently tested with a single detector. First measurements on magnetic structures have been performed. The instrument is remotely controlled using JAVA-based software and a UNIX DEC-α host computer.
Portable Microcomputer Utilization for On-Line Pulmonary Testing
Pugh, R.; Fourre, J.; Karetzky, M.
1981-01-01
A host-remote pulmonary function testing system is described that is flexible, non-dedicated, inexpensive, and readily upgradable. It is applicable for laboratories considering computerization as well as for those which have converted to one of the already available but restricted systems. The remote unit has an 8 slot bus for memory, input-output boards, and an A-D converter. It has its own terminal for manual input and display of computed and measured data which is transmitted via an acoustic modem to a larger microcomputer. The program modules are written in Pascal-Z and/or the supplied Z-80 macro assembler as external procedures.
24-channel dual microcontroller-based voltage controller for ion optics remote control
NASA Astrophysics Data System (ADS)
Bengtsson, L.
2018-05-01
The design of a 24-channel voltage control instrument for Wenzel Elektronik N1130 NIM modules is described. This instrument is remote controlled from a LabVIEW GUI on a host Windows computer and is intended for ion optics control in electron affinity measurements on negative ions at the CERN-ISOLDE facility. Each channel has a resolution of 12 bits and has a normally distributed noise with a standard deviation of <1 mV. The instrument is designed as a standard 2-unit NIM module where the electronic hardware consists of a printed circuit board with two asynchronously operating microcontrollers.
The Hico Image Processing System: A Web-Accessible Hyperspectral Remote Sensing Toolbox
NASA Astrophysics Data System (ADS)
Harris, A. T., III; Goodman, J.; Justice, B.
2014-12-01
As the quantity of Earth-observation data increases, the use-case for hosting analytical tools in geospatial data centers becomes increasingly attractive. To address this need, HySpeed Computing and Exelis VIS have developed the HICO Image Processing System, a prototype cloud computing system that provides online, on-demand, scalable remote sensing image processing capabilities. The system provides a mechanism for delivering sophisticated image processing analytics and data visualization tools into the hands of a global user community, who will only need a browser and internet connection to perform analysis. Functionality of the HICO Image Processing System is demonstrated using imagery from the Hyperspectral Imager for the Coastal Ocean (HICO), an imaging spectrometer located on the International Space Station (ISS) that is optimized for acquisition of aquatic targets. Example applications include a collection of coastal remote sensing algorithms that are directed at deriving critical information on water and habitat characteristics of our vulnerable coastal environment. The project leverages the ENVI Services Engine as the framework for all image processing tasks, and can readily accommodate the rapid integration of new algorithms, datasets and processing tools.
47 CFR 36.125 - Local switching equipment-Category 3.
Code of Federal Regulations, 2010 CFR
2010-10-01
... electronic analog or digital remote line locations. Equipment used for the identification, recording and... which has a common intermediate distributing frame, market group or other separately identifiable... composed of an electronic analog or digital host office and all of its remote locations. A host/remote...
Naver: a PC-cluster-based VR system
NASA Astrophysics Data System (ADS)
Park, ChangHoon; Ko, HeeDong; Kim, TaiYun
2003-04-01
In this paper, we present a new framework NAVER for virtual reality application. The NAVER is based on a cluster of low-cost personal computers. The goal of NAVER is to provide flexible, extensible, scalable and re-configurable framework for the virtual environments defined as the integration of 3D virtual space and external modules. External modules are various input or output devices and applications on the remote hosts. From the view of system, personal computers are divided into three servers according to its specific functions: Render Server, Device Server and Control Server. While Device Server contains external modules requiring event-based communication for the integration, Control Server contains external modules requiring synchronous communication every frame. And, the Render Server consists of 5 managers: Scenario Manager, Event Manager, Command Manager, Interaction Manager and Sync Manager. These managers support the declaration and operation of virtual environment and the integration with external modules on remote servers.
MEDLINE: the options for health professionals.
Wood, E H
1994-01-01
The bibliographic database MEDLINE, produced by the National Library of Medicine (NLM), is a computerized index to the world's biomedical literature. The database can be searched back to 1966 and contains 6.8 million records. The various means of access are divided, for the purposes of this article, into three categories: logging onto a remote host computer by telephone and modem or by the Internet; subscribing to part or all of the database on compact disc (CD-ROM); and leasing the data on a transport medium such as magnetic tape or CDs for loading on a local host computer. Decisions about which method is preferable in a given situation depend on cost, availability of hardware and software, local expertise, and the size of the intended user population. Trends include increased access to the Internet by health professionals, increased network speed, links from MEDLINE records to full-text databases or online journals, and integration of MEDLINE into wider health information systems.
Space physics analysis network node directory (The Yellow Pages): Fourth edition
NASA Technical Reports Server (NTRS)
Peters, David J.; Sisson, Patricia L.; Green, James L.; Thomas, Valerie L.
1989-01-01
The Space Physics Analysis Network (SPAN) is a component of the global DECnet Internet, which has over 17,000 host computers. The growth of SPAN from its implementation in 1981 to its present size of well over 2,500 registered SPAN host computers, has created a need for users to acquire timely information about the network through a central source. The SPAN Network Information Center (SPAN-NIC) an online facility managed by the National Space Science Data Center (NSSDC) was developed to meet this need for SPAN-wide information. The remote node descriptive information in this document is not currently contained in the SPAN-NIC database, but will be incorporated in the near future. Access to this information is also available to non-DECnet users over a variety of networks such as Telenet, the NASA Packet Switched System (NPSS), and the TCP/IP Internet. This publication serves as the Yellow Pages for SPAN node information. The document also provides key information concerning other computer networks connected to SPAN, nodes associated with each SPAN routing center, science discipline nodes, contacts for primary SPAN nodes, and SPAN reference information. A section on DECnet Internetworking discusses SPAN connections with other wide-area DECnet networks (many with thousands of nodes each). Another section lists node names and their disciplines, countries, and institutions in the SPAN Network Information Center Online Data Base System. All remote sites connected to US-SPAN and European-SPAN (E-SPAN) are indexed. Also provided is information on the SPAN tail circuits, i.e., those remote nodes connected directly to a SPAN routing center, which is the local point of contact for resolving SPAN-related problems. Reference material is included for those who wish to know more about SPAN. Because of the rapid growth of SPAN, the SPAN Yellow Pages is reissued periodically.
GATEWAY - COMMUNICATIONS GATEWAY SOFTWARE FOR NETEX, DECNET, AND TCP/IP
NASA Technical Reports Server (NTRS)
Keith, B.
1994-01-01
The Communications Gateway Software, GATEWAY, provides process-to-process communication between remote applications programs in different protocol domains. Communicating peer processes may be resident on any paired combination of NETEX, DECnet, or TCP/IP hosts. The gateway provides the necessary mapping from one protocol to another and will facilitate practical intermachine communications in a cost effective manner by eliminating the need to standardize on a single protocol or the need to implement multiple protocols in the host computers. The purpose of the gateway is to support data transfers between application programs on different host computers using different protocols. The gateway computer must be physically connected to both host computers and must contain the system software needed to use the communication protocols of both host computers. The communication process between application partners can be divided into three phases: session establishment, data transfer, and session termination. The communication protocols supported by GATEWAY (DECnet, NETEX, and TCP/IP) have addressing mechanisms that allow an application to identify itself and distinguish among other applications on the network. The exact form of the address varies depending on whether an application is passively offering (awaiting the receipt of a network connection from another network application) or actively connecting to another network. When the gateway is started, GATEWAY reads a file of address pairs. One of the address pairs is used by GATEWAY for passively offering on one network while the other address in the pair is used for actively connecting on the other network establishing the session. Now the two application partners can send and receive data in a manner appropriate to their home networks. GATEWAY accommodates full duplex transmissions. Thus, if the application partners are sophisticated enough, they can send and receive simultaneously. GATEWAY also keeps track of the number of bytes contained in each ransferred data packet. If GATEWAY detects an error during the data transfer, the sessions on both networks are terminated and the passive offer on the appropriate network is reissued. After performing the desired data transfer, one of the remote applications will send a network disconnect to the gateway to close its communication link. Upon detecting this network disconnect, GATEWAY replies with its own disconnect to ensure that the network connection has been fully terminated. Then, GATEWAY terminates its session with the other application by closing the communication link. GATEWAY has been implemented on a DEC VAX under VMS 4.7. It is written in ADA and has a central memory requirement of approximately 406K bytes. The communications protocols supported by GATEWAY are Network Systems Corporation's Network Executive (NETEX), Excelan's TCP/IP, and DECnet. GATEWAY was developed in 1988.
Expanding Your Laboratory by Accessing Collaboratory Resources
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoyt, David W.; Burton, Sarah D.; Peterson, Michael R.
2004-03-01
The Environmental Molecular Sciences Laboratory (EMSL) in Richland, Washington, is the home of a research facility setup by the United States Department of Energy (DOE). The facility is atypical because it houses over 100 cutting-edge research systems for the use of researchers all over the United States and the world. Access to the lab is requested through a peer-review proposal process and the scientists who use the facility are generally referred to as ‘users’. There are six main research facilities housed in EMSL, all of which host visiting researchers. Several of these facilities also participate in the EMSL Collaboratory, amore » remote access capability supported by EMSL operations funds. Of these, the High-Field Magnetic Resonance Facility (HFMRF) and Molecular Science Computing Facility (MSCF) have a significant number of their users performing remote work. The HFMRF in EMSL currently houses 12 NMR spectrometers that range in magnet field strength from 7.05T to 21.1T. Staff associated with the NMR facility offers scientific expertise in the areas of structural biology, solid-state materials/catalyst characterization, and magnetic resonance imaging (MRI) techniques. The way in which the HFMRF operates, with a high level of dedication to remote operation across the full suite of High-Field NMR spectrometers, has earned it the name “Virtual NMR Facility”. This review will focus on the operational aspects of remote research done in the High-Field Magnetic Resonance Facility and the computer tools that make remote experiments possible.« less
1990-01-01
ininsj^>ji-f\\Jinoj(vjM o ro -»T co m o •- IO M a m»*NO>M>fNNininNOmso(\\iOfininO’Om<-<if(MO- jio o^MONroN«-o»-Of-|ioinf-N>tO’-m>iin...will present interesting challanges for the SAFECOMP system. The Powell site remote location from Malmstrom AFB (Host Support Base) requires the
NASA Technical Reports Server (NTRS)
Millwater, Harry; Riha, David
1996-01-01
The NESSUS and NASTRAN computer codes were successfully integrated. The enhanced NESSUS code will use NASTRAN for the structural Analysis and NESSUS for the probabilistic analysis. Any quantities in the NASTRAN bulk data input can be random variables. Any NASTRAN result that is written to the output2 file can be returned to NESSUS as the finite element result. The interfacing between NESSUS and NASTRAN is handled automatically by NESSUS. NESSUS and NASTRAN can be run on different machines using the remote host option.
Multiple node remote messaging
Blumrich, Matthias A.; Chen, Dong; Gara, Alan G.; Giampapa, Mark E.; Heidelberger, Philip; Ohmacht, Martin; Salapura, Valentina; Steinmacher-Burow, Burkhard; Vranas, Pavlos
2010-08-31
A method for passing remote messages in a parallel computer system formed as a network of interconnected compute nodes includes that a first compute node (A) sends a single remote message to a remote second compute node (B) in order to control the remote second compute node (B) to send at least one remote message. The method includes various steps including controlling a DMA engine at first compute node (A) to prepare the single remote message to include a first message descriptor and at least one remote message descriptor for controlling the remote second compute node (B) to send at least one remote message, including putting the first message descriptor into an injection FIFO at the first compute node (A) and sending the single remote message and the at least one remote message descriptor to the second compute node (B).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bonior, Jason D; Hu, Zhen; Guo, Terry N.
This letter presents an experimental demonstration of software-defined-radio-based wireless tomography using computer-hosted radio devices called Universal Software Radio Peripheral (USRP). This experimental brief follows our vision and previous theoretical study of wireless tomography that combines wireless communication and RF tomography to provide a novel approach to remote sensing. Automatic data acquisition is performed inside an RF anechoic chamber. Semidefinite relaxation is used for phase retrieval, and the Born iterative method is utilized for imaging the target. Experimental results are presented, validating our vision of wireless tomography.
NASA Technical Reports Server (NTRS)
Mishchenko, Michael I.; Yang, Ping
2018-01-01
In this paper we make practical use of the recently developed first-principles approach to electromagnetic scattering by particles immersed in an unbounded absorbing host medium. Specifically, we introduce an actual computational tool for the calculation of pertinent far-field optical observables in the context of the classical Lorenzâ€"Mie theory. The paper summarizes the relevant theoretical formalism, explains various aspects of the corresponding numerical algorithm, specifies the input and output parameters of a FORTRAN program available at https://www.giss.nasa.gov/staff/mmishchenko/Lorenz-Mie.html, and tabulates benchmark results useful for testing purposes. This public-domain FORTRAN program enables one to solve the following two important problems: (i) simulate theoretically the reading of a remote well-collimated radiometer measuring electromagnetic scattering by an individual spherical particle or a small random group of spherical particles; and (ii) compute the single-scattering parameters that enter the vector radiative transfer equation derived directly from the Maxwell equations.
NASA Astrophysics Data System (ADS)
Mishchenko, Michael I.; Yang, Ping
2018-01-01
In this paper we make practical use of the recently developed first-principles approach to electromagnetic scattering by particles immersed in an unbounded absorbing host medium. Specifically, we introduce an actual computational tool for the calculation of pertinent far-field optical observables in the context of the classical Lorenz-Mie theory. The paper summarizes the relevant theoretical formalism, explains various aspects of the corresponding numerical algorithm, specifies the input and output parameters of a FORTRAN program available at https://www.giss.nasa.gov/staff/mmishchenko/Lorenz-Mie.html, and tabulates benchmark results useful for testing purposes. This public-domain FORTRAN program enables one to solve the following two important problems: (i) simulate theoretically the reading of a remote well-collimated radiometer measuring electromagnetic scattering by an individual spherical particle or a small random group of spherical particles; and (ii) compute the single-scattering parameters that enter the vector radiative transfer equation derived directly from the Maxwell equations.
World Data Center for Remote Sensing of the Atmosphere (WDC-RSAT)
NASA Astrophysics Data System (ADS)
Bittner, Michael; Höppner, Kathrin; Hildenbrand, Beate
Since 2003 the Applied Remote Sensing Cluster of DLR has hosted and operated the World Data Center for Remote Sensing of the Atmosphere (WDC-RSAT, http://wdc.dlr.de) under the non-governmental auspices of the International Council of Science (ICSU). WDC-RSAT offers scientists and the general public free access to a continuously growing collection of satellite-based atmosphere-related data sets and services. These data holdings range from raw data collected by remote sensors, to information products derived from the raw data ("value adding"). The current WDC-RSAT data holding contains data and information products addressing atmospheric trace gases, clouds, surface parameters, solar radiation, and special services as near-real time (NRT) information related to e.g. European air quality, UV radiation forecasts, global ozone level maps. In addition to archiving data sets, WDC-RSAT cooperates with other data centers and strives to provide additional services, which include data analysis and value adding, data summaries, campaign planning support, and data set validation and publication. In Germany, three current ICSU World Data Centers, namely WDC-Climate (hosted by the German Climate Computer Center, DKRZ), WDC-MARE (co-hosted by AWI and the University of Bremen), WDC-RSAT (hosted by DLR), and the pending WDC-Terra (to be hosted by GFZ) founded in 2004 the German WDC cluster for "Earth System Research", in order to promote Earth System Science and Research in Germany and abroad. This Cluster is actively pursuing a strategy, using information technology, to make data related to Earth Systems available to an as wide and as interdisciplinary possible audience. In early 2006 a scientific advisory committee for WDC-RSAT was established. External experts representing space agencies, weather services, atmospheric remote sensing technologies, and atmospheric science help to guide WDC-RSAT in setting and reaching its goals. In cooperation with the World Meteorological Organization (WMO), WDC-RSAT is aiming to become part of a data integration center to be created within the Global Atmosphere Watch (GAW)-program of the WMO. This center would concern itself with linking different GAW- relevant data sets both with each other and with models. In this context WDC-RSAT will also handle non-satellite based data which is relevant within the context of validation. WDC-RSAT is envisaged to play a major role within the recently established international and global Network for the Detection of Mesopause Change (NDMC) with the mission to promote international cooperation among research groups active in the mesopause region ( 80-100 km height) to enhance the suitability of airglow observations for the detection of long-term trends. WDC-RSAT will serve as a communication and data management platform for this worldwide network of ground-based measurements.
Kahnert, Michael; Nousiainen, Timo; Lindqvist, Hannakaisa
2013-04-08
Optical properties of light absorbing carbon (LAC) aggregates encapsulated in a shell of sulfate are computed for realistic model geometries based on field measurements. Computations are performed for wavelengths from the UV-C to the mid-IR. Both climate- and remote sensing-relevant optical properties are considered. The results are compared to commonly used simplified model geometries, none of which gives a realistic representation of the distribution of the LAC mass within the host material and, as a consequence, fail to predict the optical properties accurately. A new core-gray shell model is introduced, which accurately reproduces the size- and wavelength dependence of the integrated and differential optical properties.
NASA Astrophysics Data System (ADS)
Barberis, Stefano; Carminati, Leonardo; Leveraro, Franco; Mazza, Simone Michele; Perini, Laura; Perlz, Francesco; Rebatto, David; Tura, Ruggero; Vaccarossa, Luca; Villaplana, Miguel
2015-12-01
We present the approach of the University of Milan Physics Department and the local unit of INFN to allow and encourage the sharing among different research areas of computing, storage and networking resources (the largest ones being those composing the Milan WLCG Tier-2 centre and tailored to the needs of the ATLAS experiment). Computing resources are organised as independent HTCondor pools, with a global master in charge of monitoring them and optimising their usage. The configuration has to provide satisfactory throughput for both serial and parallel (multicore, MPI) jobs. A combination of local, remote and cloud storage options are available. The experience of users from different research areas operating on this shared infrastructure is discussed. The promising direction of improving scientific computing throughput by federating access to distributed computing and storage also seems to fit very well with the objectives listed in the European Horizon 2020 framework for research and development.
NASA Technical Reports Server (NTRS)
Davis, Bruce E.; Elliot, Gregory
1989-01-01
Jackson State University recently established the Center for Spatial Data Research and Applications, a Geographical Information System (GIS) and remote sensing laboratory. Taking advantage of new technologies and new directions in the spatial (geographic) sciences, JSU is building a Center of Excellence in Spatial Data Management. New opportunities for research, applications, and employment are emerging. GIS requires fundamental shifts and new demands in traditional computer science and geographic training. The Center is not merely another computer lab but is one setting the pace in a new applied frontier. GIS and its associated technologies are discussed. The Center's facilities are described. An ARC/INFO GIS runs on a Vax mainframe, with numerous workstations. Image processing packages include ELAS, LIPS, VICAR, and ERDAS. A host of hardware and software peripheral are used in support. Numerous projects are underway, such as the construction of a Gulf of Mexico environmental data base, development of AI in image processing, a land use dynamics study of metropolitan Jackson, and others. A new academic interdisciplinary program in Spatial Data Management is under development, combining courses in Geography and Computer Science. The broad range of JSU's GIS and remote sensing activities is addressed. The impacts on changing paradigms in the university and in the professional world conclude the discussion.
High-performance dual-speed CCD camera system for scientific imaging
NASA Astrophysics Data System (ADS)
Simpson, Raymond W.
1996-03-01
Traditionally, scientific camera systems were partitioned with a `camera head' containing the CCD and its support circuitry and a camera controller, which provided analog to digital conversion, timing, control, computer interfacing, and power. A new, unitized high performance scientific CCD camera with dual speed readout at 1 X 106 or 5 X 106 pixels per second, 12 bit digital gray scale, high performance thermoelectric cooling, and built in composite video output is described. This camera provides all digital, analog, and cooling functions in a single compact unit. The new system incorporates the A/C converter, timing, control and computer interfacing in the camera, with the power supply remaining a separate remote unit. A 100 Mbyte/second serial link transfers data over copper or fiber media to a variety of host computers, including Sun, SGI, SCSI, PCI, EISA, and Apple Macintosh. Having all the digital and analog functions in the camera made it possible to modify this system for the Woods Hole Oceanographic Institution for use on a remote controlled submersible vehicle. The oceanographic version achieves 16 bit dynamic range at 1.5 X 105 pixels/second, can be operated at depths of 3 kilometers, and transfers data to the surface via a real time fiber optic link.
REMOTE: Modem Communicator Program for the IBM personal computer
DOE Office of Scientific and Technical Information (OSTI.GOV)
McGirt, F.
1984-06-01
REMOTE, a Modem Communicator Program, was developed to provide full duplex serial communication with arbitrary remote computers via either dial-up telephone modems or direct lines. The latest version of REMOTE (documented in this report) was developed for the IBM Personal Computer.
Teleoperated control system for underground room and pillar mining
Mayercheck, William D.; Kwitowski, August J.; Brautigam, Albert L.; Mueller, Brian K.
1992-01-01
A teleoperated mining system is provided for remotely controlling the various machines involved with thin seam mining. A thin seam continuous miner located at a mining face includes a camera mounted thereon and a slave computer for controlling the miner and the camera. A plurality of sensors for relaying information about the miner and the face to the slave computer. A slave computer controlled ventilation sub-system which removes combustible material from the mining face. A haulage sub-system removes material mined by the continuous miner from the mining face to a collection site and is also controlled by the slave computer. A base station, which controls the supply of power and water to the continuous miner, haulage system, and ventilation systems, includes cable/hose handling module for winding or unwinding cables/hoses connected to the miner, an operator control module, and a hydraulic power and air compressor module for supplying air to the miner. An operator controlled host computer housed in the operator control module is connected to the slave computer via a two wire communications line.
Kuzovkov, Vladislav; Yanov, Yuri; Levin, Sergey; Bovo, Roberto; Rosignoli, Monica; Eskilsson, Gunnar; Willbas, Staffan
2014-07-01
Remote programming is safe and is well received by health-care professionals and cochlear implant (CI) users. It can be adopted into clinic routine as an alternative to face-to-face programming. Telemedicine allows a patient to be treated anywhere in the world. Although it is a growing field, little research has been published on its application to CI programming. We examined hearing professionals' and CI users' subjective reactions to the remote programming experience, including the quality of the programming and the use of the relevant technology. Remote CI programming was performed in Italy, Sweden, and Russia. Programming sessions had three participants: a CI user, a local host, and a remote expert. After the session, each CI user, local host, and remote expert each completed a questionnaire on their experience. In all, 33 remote programming sessions were carried out, resulting in 99 completed questionnaires. The overwhelming majority of study participants responded positively to all aspects of remote programming. CI users were satisfied with the results in 96.9% of the programming sessions; 100% of participants would use remote programming again. Although technical problems were encountered, they did not cause the sessions to be considerably longer than face-to-face sessions.
Remote hardware-reconfigurable robotic camera
NASA Astrophysics Data System (ADS)
Arias-Estrada, Miguel; Torres-Huitzil, Cesar; Maya-Rueda, Selene E.
2001-10-01
In this work, a camera with integrated image processing capabilities is discussed. The camera is based on an imager coupled to an FPGA device (Field Programmable Gate Array) which contains an architecture for real-time computer vision low-level processing. The architecture can be reprogrammed remotely for application specific purposes. The system is intended for rapid modification and adaptation for inspection and recognition applications, with the flexibility of hardware and software reprogrammability. FPGA reconfiguration allows the same ease of upgrade in hardware as a software upgrade process. The camera is composed of a digital imager coupled to an FPGA device, two memory banks, and a microcontroller. The microcontroller is used for communication tasks and FPGA programming. The system implements a software architecture to handle multiple FPGA architectures in the device, and the possibility to download a software/hardware object from the host computer into its internal context memory. System advantages are: small size, low power consumption, and a library of hardware/software functionalities that can be exchanged during run time. The system has been validated with an edge detection and a motion processing architecture, which will be presented in the paper. Applications targeted are in robotics, mobile robotics, and vision based quality control.
2016-07-15
AFRL-AFOSR-JP-TR-2016-0068 Multi-scale Computational Electromagnetics for Phenomenology and Saliency Characterization in Remote Sensing Hean-Teik...SUBTITLE Multi-scale Computational Electromagnetics for Phenomenology and Saliency Characterization in Remote Sensing 5a. CONTRACT NUMBER 5b. GRANT NUMBER... electromagnetics to the application in microwave remote sensing as well as extension of modelling capability with computational flexibility to study
2016-07-15
AFRL-AFOSR-JP-TR-2016-0068 Multi-scale Computational Electromagnetics for Phenomenology and Saliency Characterization in Remote Sensing Hean-Teik...SUBTITLE Multi-scale Computational Electromagnetics for Phenomenology and Saliency Characterization in Remote Sensing 5a. CONTRACT NUMBER 5b. GRANT NUMBER...electromagnetics to the application in microwave remote sensing as well as extension of modelling capability with computational flexibility to study
ERIC Educational Resources Information Center
Howe, Grant
2009-01-01
With money tight, more and more districts are considering remote access as a way to reduce expenses and budget information technology costs more effectively. Remote access allows staff members to work with a hosted software application from any school campus without being tied to a specific physical location. Each school can access critical…
Open architecture CMM motion controller
NASA Astrophysics Data System (ADS)
Chang, David; Spence, Allan D.; Bigg, Steve; Heslip, Joe; Peterson, John
2001-12-01
Although initially the only Coordinate Measuring Machine (CMM) sensor available was a touch trigger probe, technological advances in sensors and computing have greatly increased the variety of available inspection sensors. Non-contact laser digitizers and analog scanning touch probes require very well tuned CMM motion control, as well as an extensible, open architecture interface. This paper describes the implementation of a retrofit CMM motion controller designed for open architecture interface to a variety of sensors. The controller is based on an Intel Pentium microcomputer and a Servo To Go motion interface electronics card. Motor amplifiers, safety, and additional interface electronics are housed in a separate enclosure. Host Signal Processing (HSP) is used for the motion control algorithm. Compared to the usual host plus DSP architecture, single CPU HSP simplifies integration with the various sensors, and implementation of software geometric error compensation. Motion control tuning is accomplished using a remote computer via 100BaseTX Ethernet. A Graphical User Interface (GUI) is used to enter geometric error compensation data, and to optimize the motion control tuning parameters. It is shown that this architecture achieves the required real time motion control response, yet is much easier to extend to additional sensors.
Waggle: A Framework for Intelligent Attentive Sensing and Actuation
NASA Astrophysics Data System (ADS)
Sankaran, R.; Jacob, R. L.; Beckman, P. H.; Catlett, C. E.; Keahey, K.
2014-12-01
Advances in sensor-driven computation and computationally steered sensing will greatly enable future research in fields including environmental and atmospheric sciences. We will present "Waggle," an open-source hardware and software infrastructure developed with two goals: (1) reducing the separation and latency between sensing and computing and (2) improving the reliability and longevity of sensing-actuation platforms in challenging and costly deployments. Inspired by "deep-space probe" systems, the Waggle platform design includes features that can support longitudinal studies, deployments with varying communication links, and remote management capabilities. Waggle lowers the barrier for scientists to incorporate real-time data from their sensors into their computations and to manipulate the sensors or provide feedback through actuators. A standardized software and hardware design allows quick addition of new sensors/actuators and associated software in the nodes and enables them to be coupled with computational codes both insitu and on external compute infrastructure. The Waggle framework currently drives the deployment of two observational systems - a portable and self-sufficient weather platform for study of small-scale effects in Chicago's urban core and an open-ended distributed instrument in Chicago that aims to support several research pursuits across a broad range of disciplines including urban planning, microbiology and computer science. Built around open-source software, hardware, and Linux OS, the Waggle system comprises two components - the Waggle field-node and Waggle cloud-computing infrastructure. Waggle field-node affords a modular, scalable, fault-tolerant, secure, and extensible platform for hosting sensors and actuators in the field. It supports insitu computation and data storage, and integration with cloud-computing infrastructure. The Waggle cloud infrastructure is designed with the goal of scaling to several hundreds of thousands of Waggle nodes. It supports aggregating data from sensors hosted by the nodes, staging computation, relaying feedback to the nodes and serving data to end-users. We will discuss the Waggle design principles and their applicability to various observational research pursuits, and demonstrate its capabilities.
The remote sensing image segmentation mean shift algorithm parallel processing based on MapReduce
NASA Astrophysics Data System (ADS)
Chen, Xi; Zhou, Liqing
2015-12-01
With the development of satellite remote sensing technology and the remote sensing image data, traditional remote sensing image segmentation technology cannot meet the massive remote sensing image processing and storage requirements. This article put cloud computing and parallel computing technology in remote sensing image segmentation process, and build a cheap and efficient computer cluster system that uses parallel processing to achieve MeanShift algorithm of remote sensing image segmentation based on the MapReduce model, not only to ensure the quality of remote sensing image segmentation, improved split speed, and better meet the real-time requirements. The remote sensing image segmentation MeanShift algorithm parallel processing algorithm based on MapReduce shows certain significance and a realization of value.
NASA Astrophysics Data System (ADS)
Kong, D.; Donnellan, A.; Pierce, M. E.
2012-12-01
QuakeSim is an online computational framework focused on using remotely sensed geodetic imaging data to model and understand earthquakes. With the rise in online social networking over the last decade, many tools and concepts have been developed that are useful to research groups. In particular, QuakeSim is interested in the ability for researchers to post, share, and annotate files generated by modeling tools in order to facilitate collaboration. To accomplish this, features were added to the preexisting QuakeSim site that include single sign-on, automated saving of output from modeling tools, and a personal user space to manage sharing permissions on these saved files. These features implement OpenID and Lightweight Data Access Protocol (LDAP) technologies to manage files across several different servers, including a web server running Drupal and other servers hosting the computational tools themselves.
NASA Astrophysics Data System (ADS)
Gordon, M. K.; Showalter, M. R.; Ballard, L.; Tiscareno, M.; French, R. S.; Olson, D.
2017-06-01
The PDS RMS Node hosts OPUS - an accurate, comprehensive search tool for spacecraft remote sensing observations. OPUS supports Cassini: CIRS, ISS, UVIS, VIMS; New Horizons: LORRI, MVIC; Galileo SSI; Voyager ISS; and Hubble: ACS, STIS, WFC3, WFPC2.
Remote sensing - A new view for public health
NASA Technical Reports Server (NTRS)
Morrison, D. R.; Barnes, C. M.; Fuller, C. E.
1973-01-01
It is shown that the technology of remote sensing can be of great importance to the field of public health. This possibility is based on the deepened understanding of the biologies and ecologies of the vector/organism/host interelationships of arthropod-, soil-, and water-borne diseases to result from the information that remote sensing can provide.
NASA Astrophysics Data System (ADS)
Pressley, Jackson R.; Pabst, Donald; Sower, Gary D.; Nee, Larry; Green, Brian; Howard, Peter
2001-10-01
The United States Army has contracted EG&G Technical Services to build the GSTAMIDS EMD Block 0. This system autonomously detects and marks buried anti-tank land mines from an unmanned vehicle. It consists of a remotely operated host vehicle, standard teleoperation system (STS) control, mine detection system (MDS) and a control vehicle. Two complete systems are being fabricated, along with a third MDS. The host vehicle for Block 0 is the South African Meerkat that has overpass capability for anti-tank mines, as well as armor anti-mine blast protection and ballistic protection. It is operated via the STS radio link from within the control vehicle. The Main Computer System (MCS), located in the control vehicle, receives sensor data from the MDS via a high speed radio link, processes and fuses the data to make a decision of a mine detection, and sends the information back to the host vehicle for a mark to be placed on the mine location. The MCS also has the capability to interface into the FBCB2 system via SINGARS radio. The GSTAMIDS operator station and the control vehicle communications system also connect to the MCS. The MDS sensors are mounted on the host vehicle and include Ground Penetrating Radar (GPR), Pulsed Magnetic Induction (PMI) metal detector, and (as an option) long-wave infrared (LWIR). A distributed processing architecture is used so that pre-processing is performed on data at the sensor level before transmission to the MCS, minimizing required throughput. Nine (9) channels each of GPR and PMI are mounted underneath the meerkat to provide a three-meter detection swath. Two IR cameras are mounted on the upper sides of the Meerkat, providing a field of view of the required swath with overlap underneath the vehicle. Also included on the host vehicle are an Internal Navigation System (INS), Global Positioning System (GPS), and radio communications for remote control and data transmission. The GSTAMIDS Block 0 is designed as a modular, expandable system with sufficient bandwidth and processing capability for incorporation of additional sensor systems in future Blocks. It is also designed to operate in adverse weather conditions and to be transportable around the world.
Implementation of Multispectral Image Classification on a Remote Adaptive Computer
NASA Technical Reports Server (NTRS)
Figueiredo, Marco A.; Gloster, Clay S.; Stephens, Mark; Graves, Corey A.; Nakkar, Mouna
1999-01-01
As the demand for higher performance computers for the processing of remote sensing science algorithms increases, the need to investigate new computing paradigms its justified. Field Programmable Gate Arrays enable the implementation of algorithms at the hardware gate level, leading to orders of m a,gnitude performance increase over microprocessor based systems. The automatic classification of spaceborne multispectral images is an example of a computation intensive application, that, can benefit from implementation on an FPGA - based custom computing machine (adaptive or reconfigurable computer). A probabilistic neural network is used here to classify pixels of of a multispectral LANDSAT-2 image. The implementation described utilizes Java client/server application programs to access the adaptive computer from a remote site. Results verify that a remote hardware version of the algorithm (implemented on an adaptive computer) is significantly faster than a local software version of the same algorithm implemented on a typical general - purpose computer).
NASA Technical Reports Server (NTRS)
Glaese, John R.; Tobbe, Patrick A.
1986-01-01
The Space Station Mechanism Test Bed consists of a hydraulically driven, computer controlled six degree of freedom (DOF) motion system with which docking, berthing, and other mechanisms can be evaluated. Measured contact forces and moments are provided to the simulation host computer to enable representation of orbital contact dynamics. This report describes the development of a generalized math model which represents the relative motion between two rigid orbiting vehicles. The model allows motion in six DOF for each body, with no vehicle size limitation. The rotational and translational equations of motion are derived. The method used to transform the forces and moments from the sensor location to the vehicles' centers of mass is also explained. Two math models of docking mechanisms, a simple translational spring and the Remote Manipulator System end effector, are presented along with simulation results. The translational spring model is used in an attempt to verify the simulation with compensated hardware in the loop results.
Introduction to Cosmology, Proceedings of the Polish Astronomical Society volume 4
NASA Astrophysics Data System (ADS)
Biernacka, Monika; Bajan, Katarzyna; Stachowski, Grzegorz; Pollo, Agnieszka
2016-07-01
On 11-23 July 2016, Jan Kochanowski University in Kielce was the host of the Second Cosmological School "Introduction to Cosmology". The main purpose of the School was to provide an introduction to a selection of the most interesting topics in modern cosmology, both in theory and observations. The program included a series of mini-workshops on cosmological simulations, Virtual Observatory database and tools and Spectral Energy Distribution tting. The School was intended for undergraduate, MSc and PhD students, as well as young postdoctoral researchers. The School was co-organized by the Polish Astronomical Society, the Jan Kochanowski University in Kielce, the Jagiellonian University in Cracow, the Nuclear Centre for Nuclear Research and the N. Copernicus Astronomical Center in Warsaw. The Interdisciplinary Centre for Mathematical and Computational Modeling kindly provided us with the possibility to remotely use their computing facilities.
Introduction to Cosmology, Proceedings of the Polish Astronomical Society volume 4
NASA Astrophysics Data System (ADS)
Biernacka, Monika; Bajan, Katarzyna; Stachowski, Grzegorz; Pollo, Agnieszka
2017-08-01
On 11-23 July 2016, Jan Kochanowski University in Kielce was the host of the Second Cosmological School "Introduction to Cosmology". The main purpose of the School was to provide an introduction to a selection of the most interesting topics in modern cosmology, both in theory and observations. The program included a series of mini-workshops on cosmological simulations, Virtual Observatory database and tools and Spectral Energy Distribution tting. The School was intended for undergraduate, MSc and PhD students, as well as young postdoctoral researchers. The School was co-organized by the Polish Astronomical Society, the Jan Kochanowski University in Kielce, the Jagiellonian University in Cracow, the Nuclear Centre for Nuclear Research and the N. Copernicus Astronomical Center in Warsaw. The Interdisciplinary Centre for Mathematical and Computational Modeling kindly provided us with the possibility to remotely use their computing facilities.
NASA Technical Reports Server (NTRS)
1990-01-01
While a new technology called 'virtual reality' is still at the 'ground floor' level, one of its basic components, 3D computer graphics is already in wide commercial use and expanding. Other components that permit a human operator to 'virtually' explore an artificial environment and to interact with it are being demonstrated routinely at Ames and elsewhere. Virtual reality might be defined as an environment capable of being virtually entered - telepresence, it is called - or interacted with by a human. The Virtual Interface Environment Workstation (VIEW) is a head-mounted stereoscopic display system in which the display may be an artificial computer-generated environment or a real environment relayed from remote video cameras. Operator can 'step into' this environment and interact with it. The DataGlove has a series of fiber optic cables and sensors that detect any movement of the wearer's fingers and transmit the information to a host computer; a computer generated image of the hand will move exactly as the operator is moving his gloved hand. With appropriate software, the operator can use the glove to interact with the computer scene by grasping an object. The DataSuit is a sensor equipped full body garment that greatly increases the sphere of performance for virtual reality simulations.
NASA Astrophysics Data System (ADS)
Qiu, Yongfeng; Liu, Jinliang; Yang, Jianhua; Cheng, Xinbing; Yang, Xiao
2017-08-01
A compact control system based on Delphi and Field Programmable Gate Array(FPGA) is developed for a repetitive intense electron-beam accelerator(IEBA), whose output power is 10GW and pulse duration is 160ns. The system uses both hardware and software solutions. It comprises a host computer, a communication module and a main control unit. A device independent applications programming interface, devised using Delphi, is installed on the host computer. Stability theory of voltage in repetitive mode is analyzed and a detailed overview of the hardware and software configuration is presented. High voltage experiment showed that the control system fulfilled the requests of remote operation and data-acquisition. The control system based on a time-sequence control method is used to keep constant of the voltage of the primary capacitor in every shot, which ensured the stable and reliable operation of the electron beam accelerator in the repetitive mode during the experiment. Compared with the former control system based on Labview and PIC micro-controller developed in our laboratory, the present one is more compact, and with higher precision in the time dimension. It is particularly useful for automatic control of IEBA in the high power microwave effects research experiments where pulse-to-pulse reproducibility is required.
Contributing opportunistic resources to the grid with HTCondor-CE-Bosco
NASA Astrophysics Data System (ADS)
Weitzel, Derek; Bockelman, Brian
2017-10-01
The HTCondor-CE [1] is the primary Compute Element (CE) software for the Open Science Grid. While it offers many advantages for large sites, for smaller, WLCG Tier-3 sites or opportunistic clusters, it can be a difficult task to install, configure, and maintain the HTCondor-CE. Installing a CE typically involves understanding several pieces of software, installing hundreds of packages on a dedicated node, updating several configuration files, and implementing grid authentication mechanisms. On the other hand, accessing remote clusters from personal computers has been dramatically improved with Bosco: site admins only need to setup SSH public key authentication and appropriate accounts on a login host. In this paper, we take a new approach with the HTCondor-CE-Bosco, a CE which combines the flexibility and reliability of the HTCondor-CE with the easy-to-install Bosco. The administrators of the opportunistic resource are not required to install any software: only SSH access and a user account are required from the host site. The OSG can then run the grid-specific portions from a central location. This provides a new, more centralized, model for running grid services, which complements the traditional distributed model. We will show the architecture of a HTCondor-CE-Bosco enabled site, as well as feedback from multiple sites that have deployed it.
Design of MPPT Controller Monitoring Software Based on QT Framework
NASA Astrophysics Data System (ADS)
Meng, X. Z.; Lu, P. G.
2017-10-01
The MPPT controller was a hardware device for tracking the maximum power point of solar photovoltaic array. Multiple controllers could be working as networking mode by specific communicating protocol. In this article, based on C++ GUI programming with Qt frame, we designed one sort of desktop application for monitoring and analyzing operational parameter of MPPT controller. The type of communicating protocol for building network was Modbus protocol which using Remote Terminal Unit mode and The desktop application of host computer was connected with all the controllers in the network through RS485 communication or ZigBee wireless communication. Using this application, user could monitor the parameter of controller wherever they were by internet.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Link, Hamilton E.; Schroeppel, Richard Crabtree; Neumann, William Douglas
2004-10-01
If software is designed so that the software can issue functions that will move that software from one computing platform to another, then the software is said to be 'mobile'. There are two general areas of security problems associated with mobile code. The 'secure host' problem involves protecting the host from malicious mobile code. The 'secure mobile code' problem, on the other hand, involves protecting the code from malicious hosts. This report focuses on the latter problem. We have found three distinct camps of opinions regarding how to secure mobile code. There are those who believe special distributed hardware ismore » necessary, those who believe special distributed software is necessary, and those who believe neither is necessary. We examine all three camps, with a focus on the third. In the distributed software camp we examine some commonly proposed techniques including Java, D'Agents and Flask. For the specialized hardware camp, we propose a cryptographic technique for 'tamper-proofing' code over a large portion of the software/hardware life cycle by careful modification of current architectures. This method culminates by decrypting/authenticating each instruction within a physically protected CPU, thereby protecting against subversion by malicious code. Our main focus is on the camp that believes that neither specialized software nor hardware is necessary. We concentrate on methods of code obfuscation to render an entire program or a data segment on which a program depends incomprehensible. The hope is to prevent or at least slow down reverse engineering efforts and to prevent goal-oriented attacks on the software and execution. The field of obfuscation is still in a state of development with the central problem being the lack of a basis for evaluating the protection schemes. We give a brief introduction to some of the main ideas in the field, followed by an in depth analysis of a technique called 'white-boxing'. We put forth some new attacks and improvements on this method as well as demonstrating its implementation for various algorithms. We also examine cryptographic techniques to achieve obfuscation including encrypted functions and offer a new application to digital signature algorithms. To better understand the lack of security proofs for obfuscation techniques, we examine in detail general theoretical models of obfuscation. We explain the need for formal models in order to obtain provable security and the progress made in this direction thus far. Finally we tackle the problem of verifying remote execution. We introduce some methods of verifying remote exponentiation computations and some insight into generic computation checking.« less
A web portal for hydrodynamical, cosmological simulations
NASA Astrophysics Data System (ADS)
Ragagnin, A.; Dolag, K.; Biffi, V.; Cadolle Bel, M.; Hammer, N. J.; Krukau, A.; Petkova, M.; Steinborn, D.
2017-07-01
This article describes a data centre hosting a web portal for accessing and sharing the output of large, cosmological, hydro-dynamical simulations with a broad scientific community. It also allows users to receive related scientific data products by directly processing the raw simulation data on a remote computing cluster. The data centre has a multi-layer structure: a web portal, a job control layer, a computing cluster and a HPC storage system. The outer layer enables users to choose an object from the simulations. Objects can be selected by visually inspecting 2D maps of the simulation data, by performing highly compounded and elaborated queries or graphically by plotting arbitrary combinations of properties. The user can run analysis tools on a chosen object. These services allow users to run analysis tools on the raw simulation data. The job control layer is responsible for handling and performing the analysis jobs, which are executed on a computing cluster. The innermost layer is formed by a HPC storage system which hosts the large, raw simulation data. The following services are available for the users: (I) CLUSTERINSPECT visualizes properties of member galaxies of a selected galaxy cluster; (II) SIMCUT returns the raw data of a sub-volume around a selected object from a simulation, containing all the original, hydro-dynamical quantities; (III) SMAC creates idealized 2D maps of various, physical quantities and observables of a selected object; (IV) PHOX generates virtual X-ray observations with specifications of various current and upcoming instruments.
NASA Astrophysics Data System (ADS)
van Lew, Baldur; Botha, Charl P.; Milles, Julien R.; Vrooman, Henri A.; van de Giessen, Martijn; Lelieveldt, Boudewijn P. F.
2015-03-01
The cohort size required in epidemiological imaging genetics studies often mandates the pooling of data from multiple hospitals. Patient data, however, is subject to strict privacy protection regimes, and physical data storage may be legally restricted to a hospital network. To enable biomarker discovery, fast data access and interactive data exploration must be combined with high-performance computing resources, while respecting privacy regulations. We present a system using fast and inherently secure light-paths to access distributed data, thereby obviating the need for a central data repository. A secure private cloud computing framework facilitates interactive, computationally intensive exploration of this geographically distributed, privacy sensitive data. As a proof of concept, MRI brain imaging data hosted at two remote sites were processed in response to a user command at a third site. The system was able to automatically start virtual machines, run a selected processing pipeline and write results to a user accessible database, while keeping data locally stored in the hospitals. Individual tasks took approximately 50% longer compared to a locally hosted blade server but the cloud infrastructure reduced the total elapsed time by a factor of 40 using 70 virtual machines in the cloud. We demonstrated that the combination light-path and private cloud is a viable means of building an analysis infrastructure for secure data analysis. The system requires further work in the areas of error handling, load balancing and secure support of multiple users.
Computer applications in remote sensing education
NASA Technical Reports Server (NTRS)
Danielson, R. L.
1980-01-01
Computer applications to instruction in any field may be divided into two broad generic classes: computer-managed instruction and computer-assisted instruction. The division is based on how frequently the computer affects the instructional process and how active a role the computer affects the instructional process and how active a role the computer takes in actually providing instruction. There are no inherent characteristics of remote sensing education to preclude the use of one or both of these techniques, depending on the computer facilities available to the instructor. The characteristics of the two classes are summarized, potential applications to remote sensing education are discussed, and the advantages and disadvantages of computer applications to the instructional process are considered.
Peter T. Wolter; Philip A. Townsend; Brian R. Sturtevant; Clayton C. Kingdon
2008-01-01
Insects and disease affect large areas of forest in the U.S. and Canada. Understanding ecosystem impacts of such disturbances requires knowledge of host species distribution patterns on the landscape. In this study, we mapped the distribution and abundance of host species for the spruce budworm (Choristoneura fumiferana) to facilitate landscape scale...
Laptop Computer - Based Facial Recognition System Assessment
DOE Office of Scientific and Technical Information (OSTI.GOV)
R. A. Cain; G. B. Singleton
2001-03-01
The objective of this project was to assess the performance of the leading commercial-off-the-shelf (COTS) facial recognition software package when used as a laptop application. We performed the assessment to determine the system's usefulness for enrolling facial images in a database from remote locations and conducting real-time searches against a database of previously enrolled images. The assessment involved creating a database of 40 images and conducting 2 series of tests to determine the product's ability to recognize and match subject faces under varying conditions. This report describes the test results and includes a description of the factors affecting the results.more » After an extensive market survey, we selected Visionics' FaceIt{reg_sign} software package for evaluation and a review of the Facial Recognition Vendor Test 2000 (FRVT 2000). This test was co-sponsored by the US Department of Defense (DOD) Counterdrug Technology Development Program Office, the National Institute of Justice, and the Defense Advanced Research Projects Agency (DARPA). Administered in May-June 2000, the FRVT 2000 assessed the capabilities of facial recognition systems that were currently available for purchase on the US market. Our selection of this Visionics product does not indicate that it is the ''best'' facial recognition software package for all uses. It was the most appropriate package based on the specific applications and requirements for this specific application. In this assessment, the system configuration was evaluated for effectiveness in identifying individuals by searching for facial images captured from video displays against those stored in a facial image database. An additional criterion was that the system be capable of operating discretely. For this application, an operational facial recognition system would consist of one central computer hosting the master image database with multiple standalone systems configured with duplicates of the master operating in remote locations. Remote users could perform real-time searches where network connectivity is not available. As images are enrolled at the remote locations, periodic database synchronization is necessary.« less
Flexible distributed architecture for semiconductor process control and experimentation
NASA Astrophysics Data System (ADS)
Gower, Aaron E.; Boning, Duane S.; McIlrath, Michael B.
1997-01-01
Semiconductor fabrication requires an increasingly expensive and integrated set of tightly controlled processes, driving the need for a fabrication facility with fully computerized, networked processing equipment. We describe an integrated, open system architecture enabling distributed experimentation and process control for plasma etching. The system was developed at MIT's Microsystems Technology Laboratories and employs in-situ CCD interferometry based analysis in the sensor-feedback control of an Applied Materials Precision 5000 Plasma Etcher (AME5000). Our system supports accelerated, advanced research involving feedback control algorithms, and includes a distributed interface that utilizes the internet to make these fabrication capabilities available to remote users. The system architecture is both distributed and modular: specific implementation of any one task does not restrict the implementation of another. The low level architectural components include a host controller that communicates with the AME5000 equipment via SECS-II, and a host controller for the acquisition and analysis of the CCD sensor images. A cell controller (CC) manages communications between these equipment and sensor controllers. The CC is also responsible for process control decisions; algorithmic controllers may be integrated locally or via remote communications. Finally, a system server images connections from internet/intranet (web) based clients and uses a direct link with the CC to access the system. Each component communicates via a predefined set of TCP/IP socket based messages. This flexible architecture makes integration easier and more robust, and enables separate software components to run on the same or different computers independent of hardware or software platform.
NASA Technical Reports Server (NTRS)
Rediess, Herman A.; Hewett, M. D.
1991-01-01
The requirements are assessed for the use of remote computation to support HRV flight testing. First, remote computational requirements were developed to support functions that will eventually be performed onboard operational vehicles of this type. These functions which either cannot be performed onboard in the time frame of initial HRV flight test programs because the technology of airborne computers will not be sufficiently advanced to support the computational loads required, or it is not desirable to perform the functions onboard in the flight test program for other reasons. Second, remote computational support either required or highly desirable to conduct flight testing itself was addressed. The use is proposed of an Automated Flight Management System which is described in conceptual detail. Third, autonomous operations is discussed and finally, unmanned operations.
Flow Ambiguity: A Path Towards Classically Driven Blind Quantum Computation
NASA Astrophysics Data System (ADS)
Mantri, Atul; Demarie, Tommaso F.; Menicucci, Nicolas C.; Fitzsimons, Joseph F.
2017-07-01
Blind quantum computation protocols allow a user to delegate a computation to a remote quantum computer in such a way that the privacy of their computation is preserved, even from the device implementing the computation. To date, such protocols are only known for settings involving at least two quantum devices: either a user with some quantum capabilities and a remote quantum server or two or more entangled but noncommunicating servers. In this work, we take the first step towards the construction of a blind quantum computing protocol with a completely classical client and single quantum server. Specifically, we show how a classical client can exploit the ambiguity in the flow of information in measurement-based quantum computing to construct a protocol for hiding critical aspects of a computation delegated to a remote quantum computer. This ambiguity arises due to the fact that, for a fixed graph, there exist multiple choices of the input and output vertex sets that result in deterministic measurement patterns consistent with the same fixed total ordering of vertices. This allows a classical user, computing only measurement angles, to drive a measurement-based computation performed on a remote device while hiding critical aspects of the computation.
NASA Astrophysics Data System (ADS)
Ferrada, Gonzalo A.; Freitas, Saulo; Pereira, Gabriel; Paugam, Ronan
2017-04-01
This study had the aim to evaluate the new developments on the Plume Rise Model (PRM), embedded into the Brazilian developments on the Regional Atmospheric Modelling System (BRAMS). PRM computes the biomass burning plume injection heights and returns that information to the host model. Then, the atmospheric model releases all the fire emissions at this height. New developments are based on the initialization data used by the PRM, using fire size and fire radiative power (FRP) from remote sensing. The main difference between the two new versions is the conversion parameter (β) used to convert from FRP to the plume convective flux. In addition, a new scheme to generate daily fire emission fluxes is offered using the fire radiative energy (computed from remote sensing) in the Brazilian Biomass Burning Emission Model (3BEM-FRE). Model results using the three versions of the PRM are compared with observed airborne CO and O3 data from the SAMBBA campaign, which took place in southern Amazonia and Cerrado (savanna-like) regions in September 2012. Results show that improvements in both 3BEM-FRE and PRM models, had a better performance in the vertical and horizontal reproduction of CO and O3 than the original versions of them, especially in the middle and upper troposphere. Nevertheless, with some difficulty to reproduce the emissions by the end of the campaign, probably due to the cumulus parameterization used, which overestimated the precipitation in the region of study. Also, developments made in the 3BEM model show better agreement with the observed remote sensing data of daily fire emissions than the original version of it in the Amazon region, but with some difficulty in the Cerrado.
Running VisIt Software on the Peregrine System | High-Performance Computing
kilobyte range. VisIt features a robust remote visualization capability. VisIt can be started on a local machine and used to visualize data on a remote compute cluster.The remote machine must be able to send VisIt module must be loaded as part of this process. To enable remote visualization the 'module load
Accommodating Student Diversity in Remote Sensing Instruction.
ERIC Educational Resources Information Center
Hammen, John L., III.
1992-01-01
Discusses the difficulty of teaching computer-based remote sensing to students of varying levels of computer literacy. Suggests an instructional method that accommodates all levels of technical expertise through the use of microcomputers. Presents a curriculum that includes an introduction to remote sensing, digital image processing, and…
Concurrent Image Processing Executive (CIPE). Volume 1: Design overview
NASA Technical Reports Server (NTRS)
Lee, Meemong; Groom, Steven L.; Mazer, Alan S.; Williams, Winifred I.
1990-01-01
The design and implementation of a Concurrent Image Processing Executive (CIPE), which is intended to become the support system software for a prototype high performance science analysis workstation are described. The target machine for this software is a JPL/Caltech Mark 3fp Hypercube hosted by either a MASSCOMP 5600 or a Sun-3, Sun-4 workstation; however, the design will accommodate other concurrent machines of similar architecture, i.e., local memory, multiple-instruction-multiple-data (MIMD) machines. The CIPE system provides both a multimode user interface and an applications programmer interface, and has been designed around four loosely coupled modules: user interface, host-resident executive, hypercube-resident executive, and application functions. The loose coupling between modules allows modification of a particular module without significantly affecting the other modules in the system. In order to enhance hypercube memory utilization and to allow expansion of image processing capabilities, a specialized program management method, incremental loading, was devised. To minimize data transfer between host and hypercube, a data management method which distributes, redistributes, and tracks data set information was implemented. The data management also allows data sharing among application programs. The CIPE software architecture provides a flexible environment for scientific analysis of complex remote sensing image data, such as planetary data and imaging spectrometry, utilizing state-of-the-art concurrent computation capabilities.
ViPAR: a software platform for the Virtual Pooling and Analysis of Research Data.
Carter, Kim W; Francis, Richard W; Carter, K W; Francis, R W; Bresnahan, M; Gissler, M; Grønborg, T K; Gross, R; Gunnes, N; Hammond, G; Hornig, M; Hultman, C M; Huttunen, J; Langridge, A; Leonard, H; Newman, S; Parner, E T; Petersson, G; Reichenberg, A; Sandin, S; Schendel, D E; Schalkwyk, L; Sourander, A; Steadman, C; Stoltenberg, C; Suominen, A; Surén, P; Susser, E; Sylvester Vethanayagam, A; Yusof, Z
2016-04-01
Research studies exploring the determinants of disease require sufficient statistical power to detect meaningful effects. Sample size is often increased through centralized pooling of disparately located datasets, though ethical, privacy and data ownership issues can often hamper this process. Methods that facilitate the sharing of research data that are sympathetic with these issues and which allow flexible and detailed statistical analyses are therefore in critical need. We have created a software platform for the Virtual Pooling and Analysis of Research data (ViPAR), which employs free and open source methods to provide researchers with a web-based platform to analyse datasets housed in disparate locations. Database federation permits controlled access to remotely located datasets from a central location. The Secure Shell protocol allows data to be securely exchanged between devices over an insecure network. ViPAR combines these free technologies into a solution that facilitates 'virtual pooling' where data can be temporarily pooled into computer memory and made available for analysis without the need for permanent central storage. Within the ViPAR infrastructure, remote sites manage their own harmonized research dataset in a database hosted at their site, while a central server hosts the data federation component and a secure analysis portal. When an analysis is initiated, requested data are retrieved from each remote site and virtually pooled at the central site. The data are then analysed by statistical software and, on completion, results of the analysis are returned to the user and the virtually pooled data are removed from memory. ViPAR is a secure, flexible and powerful analysis platform built on open source technology that is currently in use by large international consortia, and is made publicly available at [http://bioinformatics.childhealthresearch.org.au/software/vipar/]. © The Author 2015. Published by Oxford University Press on behalf of the International Epidemiological Association.
Computers, Remote Teleprocessing and Mass Communication.
ERIC Educational Resources Information Center
Cropley, A. J.
Recent developments in computer technology are reducing the limitations of computers as mass communication devices. The growth of remote teleprocessing is one important step. Computers can now interact with users via terminals which may be hundreds of miles from the actual mainframe machine. Many terminals can be in operation at once, so that many…
Suzuki, Keishiro; Hirasawa, Yukinori; Yaegashi, Yuji; Miyamoto, Hideki; Shirato, Hiroki
2009-01-01
We developed a web-based, remote radiation treatment planning system which allowed staff at an affiliated hospital to obtain support from a fully staffed central institution. Network security was based on a firewall and a virtual private network (VPN). Client computers were installed at a cancer centre, at a university hospital and at a staff home. We remotely operated the treatment planning computer using the Remote Desktop function built in to the Windows operating system. Except for the initial setup of the VPN router, no special knowledge was needed to operate the remote radiation treatment planning system. There was a time lag that seemed to depend on the volume of data traffic on the Internet, but it did not affect smooth operation. The initial cost and running cost of the system were reasonable.
Track-to-track association for object matching in an inter-vehicle communication system
NASA Astrophysics Data System (ADS)
Yuan, Ting; Roth, Tobias; Chen, Qi; Breu, Jakob; Bogdanovic, Miro; Weiss, Christian A.
2015-09-01
Autonomous driving poses unique challenges for vehicle environment perception due to the complex driving environment the autonomous vehicle finds itself in and differentiates from remote vehicles. Due to inherent uncertainty of the traffic environments and incomplete knowledge due to sensor limitation, an autonomous driving system using only local onboard sensor information is generally not sufficiently enough for conducting a reliable intelligent driving with guaranteed safety. In order to overcome limitations of the local (host) vehicle sensing system and to increase the likelihood of correct detections and classifications, collaborative information from cooperative remote vehicles could substantially facilitate effectiveness of vehicle decision making process. Dedicated Short Range Communication (DSRC) system provides a powerful inter-vehicle wireless communication channel to enhance host vehicle environment perceiving capability with the aid of transmitted information from remote vehicles. However, there is a major challenge before one can fuse the DSRC-transmitted remote information and host vehicle Radar-observed information (in the present case): the remote DRSC data must be correctly associated with the corresponding onboard Radar data; namely, an object matching problem. Direct raw data association (i.e., measurement-to-measurement association - M2MA) is straightforward but error-prone, due to inherent uncertain nature of the observation data. The uncertainties could lead to serious difficulty in matching decision, especially, using non-stationary data. In this study, we present an object matching algorithm based on track-to-track association (T2TA) and evaluate the proposed approach with prototype vehicles in real traffic scenarios. To fully exploit potential of the DSRC system, only GPS position data from remote vehicle are used in fusion center (at host vehicle), i.e., we try to get what we need from the least amount of information; additional feature information can help the data association but are not currently considered. Comparing to M2MA, benefits of the T2TA object matching approach are: i) tracks taking into account important statistical information can provide more reliable inference results; ii) the track-formed smoothed trajectories can be used for an easier shape matching; iii) each local vehicle can design its own tracker and sends only tracks to fusion center to alleviate communication constraints. A real traffic study with different driving environments, based on a statistical hypothesis test, shows promising object matching results of significant practical implications.
Three-dimensional laser microvision.
Shimotahira, H; Iizuka, K; Chu, S C; Wah, C; Costen, F; Yoshikuni, Y
2001-04-10
A three-dimensional (3-D) optical imaging system offering high resolution in all three dimensions, requiring minimum manipulation and capable of real-time operation, is presented. The system derives its capabilities from use of the superstructure grating laser source in the implementation of a laser step frequency radar for depth information acquisition. A synthetic aperture radar technique was also used to further enhance its lateral resolution as well as extend the depth of focus. High-speed operation was made possible by a dual computer system consisting of a host and a remote microcomputer supported by a dual-channel Small Computer System Interface parallel data transfer system. The system is capable of operating near real time. The 3-D display of a tunneling diode, a microwave integrated circuit, and a see-through image taken by the system operating near real time are included. The depth resolution is 40 mum; lateral resolution with a synthetic aperture approach is a fraction of a micrometer and that without it is approximately 10 mum.
Remote sensing image ship target detection method based on visual attention model
NASA Astrophysics Data System (ADS)
Sun, Yuejiao; Lei, Wuhu; Ren, Xiaodong
2017-11-01
The traditional methods of detecting ship targets in remote sensing images mostly use sliding window to search the whole image comprehensively. However, the target usually occupies only a small fraction of the image. This method has high computational complexity for large format visible image data. The bottom-up selective attention mechanism can selectively allocate computing resources according to visual stimuli, thus improving the computational efficiency and reducing the difficulty of analysis. Considering of that, a method of ship target detection in remote sensing images based on visual attention model was proposed in this paper. The experimental results show that the proposed method can reduce the computational complexity while improving the detection accuracy, and improve the detection efficiency of ship targets in remote sensing images.
Impact of remote sensing upon the planning, management, and development of water resources
NASA Technical Reports Server (NTRS)
Loats, H. L.; Fowler, T. R.; Frech, S. L.
1974-01-01
A survey of the principal water resource users was conducted to determine the impact of new remote data streams on hydrologic computer models. The analysis of the responses and direct contact demonstrated that: (1) the majority of water resource effort of the type suitable to remote sensing inputs is conducted by major federal water resources agencies or through federally stimulated research, (2) the federal government develops most of the hydrologic models used in this effort; and (3) federal computer power is extensive. The computers, computer power, and hydrologic models in current use were determined.
Remotely accessible laboratory for MEMS testing
NASA Astrophysics Data System (ADS)
Sivakumar, Ganapathy; Mulsow, Matthew; Melinger, Aaron; Lacouture, Shelby; Dallas, Tim E.
2010-02-01
We report on the construction of a remotely accessible and interactive laboratory for testing microdevices (aka: MicroElectroMechancial Systems - MEMS). Enabling expanded utilization of microdevices for research, commercial, and educational purposes is very important for driving the creation of future MEMS devices and applications. Unfortunately, the relatively high costs associated with MEMS devices and testing infrastructure makes widespread access to the world of MEMS difficult. The creation of a virtual lab to control and actuate MEMS devices over the internet helps spread knowledge to a larger audience. A host laboratory has been established that contains a digital microscope, microdevices, controllers, and computers that can be logged into through the internet. The overall layout of the tele-operated MEMS laboratory system can be divided into two major parts: the server side and the client side. The server-side is present at Texas Tech University, and hosts a server machine that runs the Linux operating system and is used for interfacing the MEMS lab with the outside world via internet. The controls from the clients are transferred to the lab side through the server interface. The server interacts with the electronics required to drive the MEMS devices using a range of National Instruments hardware and LabView Virtual Instruments. An optical microscope (100 ×) with a CCD video camera is used to capture images of the operating MEMS. The server broadcasts the live video stream over the internet to the clients through the website. When the button is pressed on the website, the MEMS device responds and the video stream shows the movement in close to real time.
Name-Based Address Mapping for Virtual Private Networks
NASA Astrophysics Data System (ADS)
Surányi, Péter; Shinjo, Yasushi; Kato, Kazuhiko
IPv4 private addresses are commonly used in local area networks (LANs). With the increasing popularity of virtual private networks (VPNs), it has become common that a user connects to multiple LANs at the same time. However, private address ranges for LANs frequently overlap. In such cases, existing systems do not allow the user to access the resources on all LANs at the same time. In this paper, we propose name-based address mapping for VPNs, a novel method that allows connecting to hosts through multiple VPNs at the same time, even when the address ranges of the VPNs overlap. In name-based address mapping, rather than using the IP addresses used on the LANs (the real addresses), we assign a unique virtual address to each remote host based on its domain name. The local host uses the virtual addresses to communicate with remote hosts. We have implemented name-based address mapping for layer 3 OpenVPN connections on Linux and measured its performance. The communication overhead of our system is less than 1.5% for throughput and less than 0.2ms for each name resolution.
DAMT - DISTRIBUTED APPLICATION MONITOR TOOL (HP9000 VERSION)
NASA Technical Reports Server (NTRS)
Keith, B.
1994-01-01
Typical network monitors measure status of host computers and data traffic among hosts. A monitor to collect statistics about individual processes must be unobtrusive and possess the ability to locate and monitor processes, locate and monitor circuits between processes, and report traffic back to the user through a single application program interface (API). DAMT, Distributed Application Monitor Tool, is a distributed application program that will collect network statistics and make them available to the user. This distributed application has one component (i.e., process) on each host the user wishes to monitor as well as a set of components at a centralized location. DAMT provides the first known implementation of a network monitor at the application layer of abstraction. Potential users only need to know the process names of the distributed application they wish to monitor. The tool locates the processes and the circuit between them, and reports any traffic between them at a user-defined rate. The tool operates without the cooperation of the processes it monitors. Application processes require no changes to be monitored by this tool. Neither does DAMT require the UNIX kernel to be recompiled. The tool obtains process and circuit information by accessing the operating system's existing process database. This database contains all information available about currently executing processes. Expanding the information monitored by the tool can be done by utilizing more information from the process database. Traffic on a circuit between processes is monitored by a low-level LAN analyzer that has access to the raw network data. The tool also provides features such as dynamic event reporting and virtual path routing. A reusable object approach was used in the design of DAMT. The tool has four main components; the Virtual Path Switcher, the Central Monitor Complex, the Remote Monitor, and the LAN Analyzer. All of DAMT's components are independent, asynchronously executing processes. The independent processes communicate with each other via UNIX sockets through a Virtual Path router, or Switcher. The Switcher maintains a routing table showing the host of each component process of the tool, eliminating the need for each process to do so. The Central Monitor Complex provides the single application program interface (API) to the user and coordinates the activities of DAMT. The Central Monitor Complex is itself divided into independent objects that perform its functions. The component objects are the Central Monitor, the Process Locator, the Circuit Locator, and the Traffic Reporter. Each of these objects is an independent, asynchronously executing process. User requests to the tool are interpreted by the Central Monitor. The Process Locator identifies whether a named process is running on a monitored host and which host that is. The circuit between any two processes in the distributed application is identified using the Circuit Locator. The Traffic Reporter handles communication with the LAN Analyzer and accumulates traffic updates until it must send a traffic report to the user. The Remote Monitor process is replicated on each monitored host. It serves the Central Monitor Complex processes with application process information. The Remote Monitor process provides access to operating systems information about currently executing processes. It allows the Process Locator to find processes and the Circuit Locator to identify circuits between processes. It also provides lifetime information about currently monitored processes. The LAN Analyzer consists of two processes. Low-level monitoring is handled by the Sniffer. The Sniffer analyzes the raw data on a single, physical LAN. It responds to commands from the Analyzer process, which maintains the interface to the Traffic Reporter and keeps track of which circuits to monitor. DAMT is written in C-language for HP-9000 series computers running HP-UX and Sun 3 and 4 series computers running SunOS. DAMT requires 1Mb of disk space and 4Mb of RAM for execution. This package requires MIT's X Window System, Version 11 Revision 4, with OSF/Motif 1.1. The HP-9000 version (GSC-13589) includes sample HP-9000/375 and HP-9000/730 executables which were compiled under HP-UX, and the Sun version (GSC-13559) includes sample Sun3 and Sun4 executables compiled under SunOS. The standard distribution medium for the HP version of DAMT is a .25 inch HP pre-formatted streaming magnetic tape cartridge in UNIX tar format. It is also available on a 4mm magnetic tape in UNIX tar format. The standard distribution medium for the Sun version of DAMT is a .25 inch streaming magnetic tape cartridge in UNIX tar format. It is also available on a 3.5 inch diskette in UNIX tar format. DAMT was developed in 1992.
Remote Symbolic Computation of Loci
ERIC Educational Resources Information Center
Abanades, Miguel A.; Escribano, Jesus; Botana, Francisco
2010-01-01
This article presents a web-based tool designed to compute certified equations and graphs of geometric loci specified using standard Dynamic Geometry Systems (DGS). Complementing the graphing abilities of the considered DGS, the equations of the loci produced by the application are remotely computed using symbolic algebraic techniques from the…
Echinostomes in Felid Coprolites from Brazil.
Sianto, L; Duarte, A N; Borba, V H; Magalhães, J G; de Souza, S M; Chame, M
2016-06-01
The first record of Echinostoma (Trematoda: Echinostomatidae) in coprolites was from a mummified human body in Minas Gerais State, Brazil. The finding raised questions on this parasite's incidence in prehistoric populations and the natural hosts of each species in remote times. Echinostomes occur worldwide and, despite the wide range of hosts, there is no record of Echinostomatidae in felines in Brazil. This study reports the finding of Echinostomatidae eggs in felid coprolites in the Furna do Estrago Archaeological Site, located in Pernambuco State in the Brazilian semiarid. Despite the possibility of false parasitism, the finding expands the distribution of this Digenea in remote times and raises the hypothesis of other cases of echinostomiasis in pre-Colombian populations.
System design and implementation of digital-image processing using computational grids
NASA Astrophysics Data System (ADS)
Shen, Zhanfeng; Luo, Jiancheng; Zhou, Chenghu; Huang, Guangyu; Ma, Weifeng; Ming, Dongping
2005-06-01
As a special type of digital image, remotely sensed images are playing increasingly important roles in our daily lives. Because of the enormous amounts of data involved, and the difficulties of data processing and transfer, an important issue for current computer and geo-science experts is developing internet technology to implement rapid remotely sensed image processing. Computational grids are able to solve this problem effectively. These networks of computer workstations enable the sharing of data and resources, and are used by computer experts to solve imbalances of network resources and lopsided usage. In China, computational grids combined with spatial-information-processing technology have formed a new technology: namely, spatial-information grids. In the field of remotely sensed images, spatial-information grids work more effectively for network computing, data processing, resource sharing, task cooperation and so on. This paper focuses mainly on the application of computational grids to digital-image processing. Firstly, we describe the architecture of digital-image processing on the basis of computational grids, its implementation is then discussed in detail with respect to the technology of middleware. The whole network-based intelligent image-processing system is evaluated on the basis of the experimental analysis of remotely sensed image-processing tasks; the results confirm the feasibility of the application of computational grids to digital-image processing.
Educational Computer Utilization and Computer Communications.
ERIC Educational Resources Information Center
Singh, Jai P.; Morgan, Robert P.
As part of an analysis of educational needs and telecommunications requirements for future educational satellite systems, three studies were carried out. 1) The role of the computer in education was examined and both current status and future requirements were analyzed. Trade-offs between remote time sharing and remote batch process were explored…
Risk profiling of schistosomiasis using remote sensing: approaches, challenges and outlook.
Walz, Yvonne; Wegmann, Martin; Dech, Stefan; Raso, Giovanna; Utzinger, Jürg
2015-03-17
Schistosomiasis is a water-based disease that affects an estimated 250 million people, mainly in sub-Saharan Africa. The transmission of schistosomiasis is spatially and temporally restricted to freshwater bodies that contain schistosome cercariae released from specific snails that act as intermediate hosts. Our objective was to assess the contribution of remote sensing applications and to identify remaining challenges in its optimal application for schistosomiasis risk profiling in order to support public health authorities to better target control interventions. We reviewed the literature (i) to deepen our understanding of the ecology and the epidemiology of schistosomiasis, placing particular emphasis on remote sensing; and (ii) to fill an identified gap, namely interdisciplinary research that bridges different strands of scientific inquiry to enhance spatially explicit risk profiling. As a first step, we reviewed key factors that govern schistosomiasis risk. Secondly, we examined remote sensing data and variables that have been used for risk profiling of schistosomiasis. Thirdly, the linkage between the ecological consequence of environmental conditions and the respective measure of remote sensing data were synthesised. We found that the potential of remote sensing data for spatial risk profiling of schistosomiasis is - in principle - far greater than explored thus far. Importantly though, the application of remote sensing data requires a tailored approach that must be optimised by selecting specific remote sensing variables, considering the appropriate scale of observation and modelling within ecozones. Interestingly, prior studies that linked prevalence of Schistosoma infection to remotely sensed data did not reflect that there is a spatial gap between the parasite and intermediate host snail habitats where disease transmission occurs, and the location (community or school) where prevalence measures are usually derived from. Our findings imply that the potential of remote sensing data for risk profiling of schistosomiasis and other neglected tropical diseases has yet to be fully exploited.
RIP-REMOTE INTERACTIVE PARTICLE-TRACER
NASA Technical Reports Server (NTRS)
Rogers, S. E.
1994-01-01
Remote Interactive Particle-tracing (RIP) is a distributed-graphics program which computes particle traces for computational fluid dynamics (CFD) solution data sets. A particle trace is a line which shows the path a massless particle in a fluid will take; it is a visual image of where the fluid is going. The program is able to compute and display particle traces at a speed of about one trace per second because it runs on two machines concurrently. The data used by the program is contained in two files. The solution file contains data on density, momentum and energy quantities of a flow field at discrete points in three-dimensional space, while the grid file contains the physical coordinates of each of the discrete points. RIP requires two computers. A local graphics workstation interfaces with the user for program control and graphics manipulation, and a remote machine interfaces with the solution data set and performs time-intensive computations. The program utilizes two machines in a distributed mode for two reasons. First, the data to be used by the program is usually generated on the supercomputer. RIP avoids having to convert and transfer the data, eliminating any memory limitations of the local machine. Second, as computing the particle traces can be computationally expensive, RIP utilizes the power of the supercomputer for this task. Although the remote site code was developed on a CRAY, it is possible to port this to any supercomputer class machine with a UNIX-like operating system. Integration of a velocity field from a starting physical location produces the particle trace. The remote machine computes the particle traces using the particle-tracing subroutines from PLOT3D/AMES, a CFD post-processing graphics program available from COSMIC (ARC-12779). These routines use a second-order predictor-corrector method to integrate the velocity field. Then the remote program sends graphics tokens to the local machine via a remote-graphics library. The local machine interprets the graphics tokens and draws the particle traces. The program is menu driven. RIP is implemented on the silicon graphics IRIS 3000 (local workstation) with an IRIX operating system and on the CRAY2 (remote station) with a UNICOS 1.0 or 2.0 operating system. The IRIS 4D can be used in place of the IRIS 3000. The program is written in C (67%) and FORTRAN 77 (43%) and has an IRIS memory requirement of 4 MB. The remote and local stations must use the same user ID. PLOT3D/AMES unformatted data sets are required for the remote machine. The program was developed in 1988.
NASA Technical Reports Server (NTRS)
Wilcox, Brian H.
1994-01-01
System for remote control of robotic land vehicle requires only small radio-communication bandwidth. Twin video cameras on vehicle create stereoscopic images. Operator views cross-polarized images on two cathode-ray tubes through correspondingly polarized spectacles. By use of cursor on frozen image, remote operator designates path. Vehicle proceeds to follow path, by use of limited degree of autonomous control to cope with unexpected conditions. System concept, called "computer-aided remote driving" (CARD), potentially useful in exploration of other planets, military surveillance, firefighting, and clean-up of hazardous materials.
Patterns of computer usage among medical practitioners in rural and remote Queensland.
White, Col; Sheedy, Vicki; Lawrence, Nicola
2002-06-01
As part of a more detailed needs analysis, patterns of computer usage among medical practitioners in rural and remote Queensland were investigated. Utilising a questionnaire approach, a response rate of 23.82% (n = 131) was obtained. Results suggest that medical practitioners in rural and remote Queensland are relatively sophisticated in their use of computer and information technologies and have embraced computerisation to a substantially higher extent compared with their urban counterparts and previously published estimates. Findings also indicate that a substantial number of rural and remote practitioners are utilising computer and information technologies for clinical purposes such as pathology, patient information sheets, prescribing, education, patient records and patient recalls. Despite barriers such as bandwidth limitations, cost and the sometimes unreliable quality of Internet service providers, a majority of rural and remote respondents rated an Internet site with continuing medical education information and services as being important or very important. Suggestions that "rural doctors are slow to adapt to new technologies" are questioned, with findings indicating that rural and remote medical practitioners in Queensland have adapted to, and utilise, information technology to a far higher extent than has been previously documented.
The Machine / Job Features Mechanism
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alef, M.; Cass, T.; Keijser, J. J.
Within the HEPiX virtualization group and the Worldwide LHC Computing Grid’s Machine/Job Features Task Force, a mechanism has been developed which provides access to detailed information about the current host and the current job to the job itself. This allows user payloads to access meta information, independent of the current batch system or virtual machine model. The information can be accessed either locally via the filesystem on a worker node, or remotely via HTTP(S) from a webserver. This paper describes the final version of the specification from 2016 which was published as an HEP Software Foundation technical note, and themore » design of the implementations of this version for batch and virtual machine platforms. We discuss early experiences with these implementations and how they can be exploited by experiment frameworks.« less
The machine/job features mechanism
NASA Astrophysics Data System (ADS)
Alef, M.; Cass, T.; Keijser, J. J.; McNab, A.; Roiser, S.; Schwickerath, U.; Sfiligoi, I.
2017-10-01
Within the HEPiX virtualization group and the Worldwide LHC Computing Grid’s Machine/Job Features Task Force, a mechanism has been developed which provides access to detailed information about the current host and the current job to the job itself. This allows user payloads to access meta information, independent of the current batch system or virtual machine model. The information can be accessed either locally via the filesystem on a worker node, or remotely via HTTP(S) from a webserver. This paper describes the final version of the specification from 2016 which was published as an HEP Software Foundation technical note, and the design of the implementations of this version for batch and virtual machine platforms. We discuss early experiences with these implementations and how they can be exploited by experiment frameworks.
Additive Manufacturing of a Microbial Fuel Cell—A detailed study
Calignano, Flaviana; Tommasi, Tonia; Manfredi, Diego; Chiolerio, Alessandro
2015-01-01
In contemporary society we observe an everlasting permeation of electron devices, smartphones, portable computing tools. The tiniest living organisms on Earth could become the key to address this challenge: energy generation by bacterial processes from renewable stocks/waste through devices such as microbial fuel cells (MFCs). However, the application of this solution was limited by a moderately low efficiency. We explored the limits, if any, of additive manufacturing (AM) technology to fabricate a fully AM-based powering device, exploiting low density, open porosities able to host the microbes, systems easy to fuel continuously and to run safely. We obtained an optimal energy recovery close to 3 kWh m−3 per day that can power sensors and low-power appliances, allowing data processing and transmission from remote/harsh environments. PMID:26611142
Additive Manufacturing of a Microbial Fuel Cell—A detailed study
NASA Astrophysics Data System (ADS)
Calignano, Flaviana; Tommasi, Tonia; Manfredi, Diego; Chiolerio, Alessandro
2015-11-01
In contemporary society we observe an everlasting permeation of electron devices, smartphones, portable computing tools. The tiniest living organisms on Earth could become the key to address this challenge: energy generation by bacterial processes from renewable stocks/waste through devices such as microbial fuel cells (MFCs). However, the application of this solution was limited by a moderately low efficiency. We explored the limits, if any, of additive manufacturing (AM) technology to fabricate a fully AM-based powering device, exploiting low density, open porosities able to host the microbes, systems easy to fuel continuously and to run safely. We obtained an optimal energy recovery close to 3 kWh m-3 per day that can power sensors and low-power appliances, allowing data processing and transmission from remote/harsh environments.
NASA Astrophysics Data System (ADS)
Abdelazim, S.; Santoro, D.; Arend, M.; Moshary, F.; Ahmed, S.
2011-11-01
A field deployable all-fiber eye-safe Coherent Doppler LIDAR is being developed at the Optical Remote Sensing Lab at the City College of New York (CCNY) and is designed to monitor wind fields autonomously and continuously in urban settings. Data acquisition is accomplished by sampling lidar return signals at 400 MHz and performing onboard processing using field programmable gate arrays (FPGAs). The FPGA is programmed to accumulate signal information that is used to calculate the power spectrum of the atmospherically back scattered signal. The advantage of using FPGA is that signal processing will be performed at the hardware level, reducing the load on the host computer and allowing for 100% return signal processing. An experimental setup measured wind speeds at ranges of up to 3 km.
A New Computational Framework for Atmospheric and Surface Remote Sensing
NASA Technical Reports Server (NTRS)
Timucin, Dogan A.
2004-01-01
A Bayesian data-analysis framework is described for atmospheric and surface retrievals from remotely-sensed hyper-spectral data. Some computational techniques are high- lighted for improved accuracy in the forward physics model.
Lowering the Barrier for Standards-Compliant and Discoverable Hydrological Data Publication
NASA Astrophysics Data System (ADS)
Kadlec, J.
2013-12-01
The growing need for sharing and integration of hydrological and climate data across multiple organizations has resulted in the development of distributed, services-based, standards-compliant hydrological data management and data hosting systems. The problem with these systems is complicated set-up and deployment. Many existing systems assume that the data publisher has remote-desktop access to a locally managed server and experience with computer network setup. For corporate websites, shared web hosting services with limited root access provide an inexpensive, dynamic web presence solution using the Linux, Apache, MySQL and PHP (LAMP) software stack. In this paper, we hypothesize that a webhosting service provides an optimal, low-cost solution for hydrological data hosting. We propose a software architecture of a standards-compliant, lightweight and easy-to-deploy hydrological data management system that can be deployed on the majority of existing shared internet webhosting services. The architecture and design is validated by developing Hydroserver Lite: a PHP and MySQL-based hydrological data hosting package that is fully standards-compliant and compatible with the Consortium of Universities for Advancement of Hydrologic Sciences (CUAHSI) hydrologic information system. It is already being used for management of field data collection by students of the McCall Outdoor Science School in Idaho. For testing, the Hydroserver Lite software has been installed on multiple different free and low-cost webhosting sites including Godaddy, Bluehost and 000webhost. The number of steps required to set-up the server is compared with the number of steps required to set-up other standards-compliant hydrologic data hosting systems including THREDDS, IstSOS and MapServer SOS.
USDA-ARS?s Scientific Manuscript database
Remote sensing was used in a series of experiments over a three-year period to obtain spectral reflectance data for use in studying differences in vegetation indices between grasses, broadleaf plants, and grass/broadleaf plant mixtures. Empirical simulations of selected non-crop winter and spring h...
Remote Sensing of Landscapes with Spectral Images
NASA Astrophysics Data System (ADS)
Adams, John B.; Gillespie, Alan R.
2006-05-01
Remote Sensing of Landscapes with Spectral Images describes how to process and interpret spectral images using physical models to bridge the gap between the engineering and theoretical sides of remote-sensing and the world that we encounter when we venture outdoors. The emphasis is on the practical use of images rather than on theory and mathematical derivations. Examples are drawn from a variety of landscapes and interpretations are tested against the reality seen on the ground. The reader is led through analysis of real images (using figures and explanations); the examples are chosen to illustrate important aspects of the analytic framework. This textbook will form a valuable reference for graduate students and professionals in a variety of disciplines including ecology, forestry, geology, geography, urban planning, archeology and civil engineering. It is supplemented by a web-site hosting digital color versions of figures in the book as well as ancillary images (www.cambridge.org/9780521662214). Presents a coherent view of practical remote sensing, leading from imaging and field work to the generation of useful thematic maps Explains how to apply physical models to help interpret spectral images Supplemented by a website hosting digital colour versions of figures in the book, as well as additional colour figures
NASA Astrophysics Data System (ADS)
Hernandez, C.
2010-09-01
The weakness of small island electrical grids implies a handicap for the electrical generation with renewable energy sources. With the intention of maximizing the installation of photovoltaic generators in the Canary Islands, arises the need to develop a solar forecasting system that allows knowing in advance the amount of PV generated electricity that will be going into the grid, from the installed PV power plants installed in the island. The forecasting tools need to get feedback from real weather data in "real time" from remote weather stations. Nevertheless, the transference of this data to the calculation computer servers is very complicated with the old point to point telecommunication systems that, neither allow the transfer of data from several remote weather stations simultaneously nor high frequency of sampling of weather parameters due to slowness of the connection. This one project has developed a telecommunications infrastructure that allows sensorizadas remote stations, to send data of its sensors, once every minute and simultaneously, to the calculation server running the solar forecasting numerical models. For it, the Canary Islands Institute of Technology has added a sophisticated communications network to its 30 weather stations measuring irradiation at strategic sites, areas with high penetration of photovoltaic generation or that have potential to host in the future photovoltaic power plants connected to the grid. In each one of the stations, irradiance and temperature measurement instruments have been installed, over inclined silicon cell, global radiation on horizontal surface and room temperature. Mobile telephone devices have been installed and programmed in each one of the weather stations, which allow the transfer of their data taking advantage of the UMTS service offered by the local telephone operator. Every minute the computer server running the numerical weather forecasting models receives data inputs from 120 instruments distributed over the 30 radiometric stations. As a the result, currently it exist a stable, flexible, safe and economic infrastructure of radiometric stations and telecommunications that allows, on the one hand, to have data in real time from all 30 remote weather stations, and on the other hand allows to communicate with them in order to reprogram them and to carry out maintenance works.
Electronic collection system for spacelab mission timeline requirements
NASA Technical Reports Server (NTRS)
Lindberg, James P.; Piner, John R.; Huang, Allen K. H.
1995-01-01
This paper describes the Functional Objective Requirements Collection System (FORCS) software tool that has been developed for use by Principal Investigators (PI's) and Payload Element Developers (PED's) on their own personal computers to develop on-orbit timelining requirements for their payloads. The FORCS tool can be used either in a totally stand-alone mode, storing the information in a local file on the user's personal computer hard disk or in a remote mode where the user's computer is linked to a host computer containing the integrated database of the timeline requirements for all of the payloads on a mission. There are a number of features incorporated in the FORCS software to assist the user. The user may move freely back and forth between the various forms for inputting the data. Several methods are used to input the information, depending on the type of the information. These methods range from filling in text boxes, using check boxes and radio buttons, to inputting information into a spreadsheet format. There are automated features provided to assist in developing the proper format for the data, ranging from limit checking on some of the parameters to automatic conversion of different formats of time data inputs to the one standard format used for the timeline scheduling software.
Methods of training the graduate level and professional geologist in remote sensing technology
NASA Technical Reports Server (NTRS)
Kolm, K. E.
1981-01-01
Requirements for a basic course in remote sensing to accommodate the needs of the graduate level and professional geologist are described. The course should stress the general topics of basic remote sensing theory, the theory and data types relating to different remote sensing systems, an introduction to the basic concepts of computer image processing and analysis, the characteristics of different data types, the development of methods for geological interpretations, the integration of all scales and data types of remote sensing in a given study, the integration of other data bases (geophysical and geochemical) into a remote sensing study, and geological remote sensing applications. The laboratories should stress hands on experience to reinforce the concepts and procedures presented in the lecture. The geologist should then be encouraged to pursue a second course in computer image processing and analysis of remotely sensed data.
Remote sensing of vegetation structure using computer vision
NASA Astrophysics Data System (ADS)
Dandois, Jonathan P.
High-spatial resolution measurements of vegetation structure are needed for improving understanding of ecosystem carbon, water and nutrient dynamics, the response of ecosystems to a changing climate, and for biodiversity mapping and conservation, among many research areas. Our ability to make such measurements has been greatly enhanced by continuing developments in remote sensing technology---allowing researchers the ability to measure numerous forest traits at varying spatial and temporal scales and over large spatial extents with minimal to no field work, which is costly for large spatial areas or logistically difficult in some locations. Despite these advances, there remain several research challenges related to the methods by which three-dimensional (3D) and spectral datasets are joined (remote sensing fusion) and the availability and portability of systems for frequent data collections at small scale sampling locations. Recent advances in the areas of computer vision structure from motion (SFM) and consumer unmanned aerial systems (UAS) offer the potential to address these challenges by enabling repeatable measurements of vegetation structural and spectral traits at the scale of individual trees. However, the potential advances offered by computer vision remote sensing also present unique challenges and questions that need to be addressed before this approach can be used to improve understanding of forest ecosystems. For computer vision remote sensing to be a valuable tool for studying forests, bounding information about the characteristics of the data produced by the system will help researchers understand and interpret results in the context of the forest being studied and of other remote sensing techniques. This research advances understanding of how forest canopy and tree 3D structure and color are accurately measured by a relatively low-cost and portable computer vision personal remote sensing system: 'Ecosynth'. Recommendations are made for optimal conditions under which forest structure measurements should be obtained with UAS-SFM remote sensing. Ultimately remote sensing of vegetation by computer vision offers the potential to provide an 'ecologist's eye view', capturing not only canopy 3D and spectral properties, but also seeing the trees in the forest and the leaves on the trees.
NASA Astrophysics Data System (ADS)
Zhao, Shaoshuai; Ni, Chen; Cao, Jing; Li, Zhengqiang; Chen, Xingfeng; Ma, Yan; Yang, Leiku; Hou, Weizhen; Qie, Lili; Ge, Bangyu; Liu, Li; Xing, Jin
2018-03-01
The remote sensing image is usually polluted by atmosphere components especially like aerosol particles. For the quantitative remote sensing applications, the radiative transfer model based atmospheric correction is used to get the reflectance with decoupling the atmosphere and surface by consuming a long computational time. The parallel computing is a solution method for the temporal acceleration. The parallel strategy which uses multi-CPU to work simultaneously is designed to do atmospheric correction for a multispectral remote sensing image. The parallel framework's flow and the main parallel body of atmospheric correction are described. Then, the multispectral remote sensing image of the Chinese Gaofen-2 satellite is used to test the acceleration efficiency. When the CPU number is increasing from 1 to 8, the computational speed is also increasing. The biggest acceleration rate is 6.5. Under the 8 CPU working mode, the whole image atmospheric correction costs 4 minutes.
Using IKAROS as a data transfer and management utility within the KM3NeT computing model
NASA Astrophysics Data System (ADS)
Filippidis, Christos; Cotronis, Yiannis; Markou, Christos
2016-04-01
KM3NeT is a future European deep-sea research infrastructure hosting a new generation neutrino detectors that - located at the bottom of the Mediterranean Sea - will open a new window on the universe and answer fundamental questions both in particle physics and astrophysics. IKAROS is a framework that enables creating scalable storage formations on-demand and helps addressing several limitations that the current file systems face when dealing with very large scale infrastructures. It enables creating ad-hoc nearby storage formations and can use a huge number of I/O nodes in order to increase the available bandwidth (I/O and network). IKAROS unifies remote and local access in the overall data flow, by permitting direct access to each I/O node. In this way we can handle the overall data flow at the network layer, limiting the interaction with the operating system. This approach allows virtually connecting, at the users level, the several different computing facilities used (Grids, Clouds, HPCs, Data Centers, Local computing Clusters and personal storage devices), on-demand, based on the needs, by using well known standards and protocols, like HTTP.
Impact of remote sensing upon the planning, management, and development of water resources
NASA Technical Reports Server (NTRS)
Castruccio, P. A.; Loats, H. L.; Fowler, T. R.; Frech, S. L.
1975-01-01
Principal water resources users were surveyed to determine the impact of remote data streams on hydrologic computer models. Analysis of responses demonstrated that: most water resources effort suitable to remote sensing inputs is conducted through federal agencies or through federally stimulated research; and, most hydrologic models suitable to remote sensing data are federally developed. Computer usage by major water resources users was analyzed to determine the trends of usage and costs for the principal hydrologic users/models. The laws and empirical relationships governing the growth of the data processing loads were described and applied to project the future data loads. Data loads for ERTS CCT image processing were computed and projected through the 1985 era.
Evolving technologies for Space Station Freedom computer-based workstations
NASA Technical Reports Server (NTRS)
Jensen, Dean G.; Rudisill, Marianne
1990-01-01
Viewgraphs on evolving technologies for Space Station Freedom computer-based workstations are presented. The human-computer computer software environment modules are described. The following topics are addressed: command and control workstation concept; cupola workstation concept; Japanese experiment module RMS workstation concept; remote devices controlled from workstations; orbital maneuvering vehicle free flyer; remote manipulator system; Japanese experiment module exposed facility; Japanese experiment module small fine arm; flight telerobotic servicer; human-computer interaction; and workstation/robotics related activities.
GISpark: A Geospatial Distributed Computing Platform for Spatiotemporal Big Data
NASA Astrophysics Data System (ADS)
Wang, S.; Zhong, E.; Wang, E.; Zhong, Y.; Cai, W.; Li, S.; Gao, S.
2016-12-01
Geospatial data are growing exponentially because of the proliferation of cost effective and ubiquitous positioning technologies such as global remote-sensing satellites and location-based devices. Analyzing large amounts of geospatial data can provide great value for both industrial and scientific applications. Data- and compute- intensive characteristics inherent in geospatial big data increasingly pose great challenges to technologies of data storing, computing and analyzing. Such challenges require a scalable and efficient architecture that can store, query, analyze, and visualize large-scale spatiotemporal data. Therefore, we developed GISpark - a geospatial distributed computing platform for processing large-scale vector, raster and stream data. GISpark is constructed based on the latest virtualized computing infrastructures and distributed computing architecture. OpenStack and Docker are used to build multi-user hosting cloud computing infrastructure for GISpark. The virtual storage systems such as HDFS, Ceph, MongoDB are combined and adopted for spatiotemporal data storage management. Spark-based algorithm framework is developed for efficient parallel computing. Within this framework, SuperMap GIScript and various open-source GIS libraries can be integrated into GISpark. GISpark can also integrated with scientific computing environment (e.g., Anaconda), interactive computing web applications (e.g., Jupyter notebook), and machine learning tools (e.g., TensorFlow/Orange). The associated geospatial facilities of GISpark in conjunction with the scientific computing environment, exploratory spatial data analysis tools, temporal data management and analysis systems make up a powerful geospatial computing tool. GISpark not only provides spatiotemporal big data processing capacity in the geospatial field, but also provides spatiotemporal computational model and advanced geospatial visualization tools that deals with other domains related with spatial property. We tested the performance of the platform based on taxi trajectory analysis. Results suggested that GISpark achieves excellent run time performance in spatiotemporal big data applications.
NASA Technical Reports Server (NTRS)
Fijany, Amir (Inventor); Bejczy, Antal K. (Inventor)
1993-01-01
This is a real-time robotic controller and simulator which is a MIMD-SIMD parallel architecture for interfacing with an external host computer and providing a high degree of parallelism in computations for robotic control and simulation. It includes a host processor for receiving instructions from the external host computer and for transmitting answers to the external host computer. There are a plurality of SIMD microprocessors, each SIMD processor being a SIMD parallel processor capable of exploiting fine grain parallelism and further being able to operate asynchronously to form a MIMD architecture. Each SIMD processor comprises a SIMD architecture capable of performing two matrix-vector operations in parallel while fully exploiting parallelism in each operation. There is a system bus connecting the host processor to the plurality of SIMD microprocessors and a common clock providing a continuous sequence of clock pulses. There is also a ring structure interconnecting the plurality of SIMD microprocessors and connected to the clock for providing the clock pulses to the SIMD microprocessors and for providing a path for the flow of data and instructions between the SIMD microprocessors. The host processor includes logic for controlling the RRCS by interpreting instructions sent by the external host computer, decomposing the instructions into a series of computations to be performed by the SIMD microprocessors, using the system bus to distribute associated data among the SIMD microprocessors, and initiating activity of the SIMD microprocessors to perform the computations on the data by procedure call.
Security Measures to Protect Mobile Agents
NASA Astrophysics Data System (ADS)
Dadhich, Piyanka; Govil, M. C.; Dutta, Kamlesh
2010-11-01
The security issues of mobile agent systems have embarrassed its widespread implementation. Mobile agents that move around the network are not safe because the remote hosts that accommodate the agents initiates all kinds of attacks. These hosts try to analyze the agent's decision logic and their accumulated data. So, mobile agent security is the most challenging unsolved problems. The paper analyzes various security measures deeply. Security especially the attacks performed by hosts to the visiting mobile agent (the malicious hosts problem) is a major obstacle that prevents mobile agent technology from being widely adopted. Being the running environment for mobile agent, the host has full control over them and could easily perform many kinds of attacks against them.
NASA Technical Reports Server (NTRS)
Dorais, Gregory A.; Kurien, James; Rajan, Kanna
1999-01-01
We describe the computer demonstration of the Remote Agent Experiment (RAX). The Remote Agent is a high-level, model-based, autonomous control agent being validated on the NASA Deep Space 1 spacecraft.
Estimating Water Levels with Google Earth Engine
NASA Astrophysics Data System (ADS)
Lucero, E.; Russo, T. A.; Zentner, M.; May, J.; Nguy-Robertson, A. L.
2016-12-01
Reservoirs serve multiple functions and are vital for storage, electricity generation, and flood control. For many areas, traditional ground-based reservoir measurements may not be available or data dissemination may be problematic. Consistent monitoring of reservoir levels in data-poor areas can be achieved through remote sensing, providing information to researchers and the international community. Estimates of trends and relative reservoir volume can be used to identify water supply vulnerability, anticipate low power generation, and predict flood risk. Image processing with automated cloud computing provides opportunities to study multiple geographic areas in near real-time. We demonstrate the prediction capability of a cloud environment for identifying water trends at reservoirs in the US, and then apply the method to data-poor areas in North Korea, Iran, Azerbaijan, Zambia, and India. The Google Earth Engine cloud platform hosts remote sensing data and can be used to automate reservoir level estimation with multispectral imagery. We combine automated cloud-based analysis from Landsat image classification to identify reservoir surface area trends and radar altimetry to identify reservoir level trends. The study estimates water level trends using three years of data from four domestic reservoirs to validate the remote sensing method, and five foreign reservoirs to demonstrate the method application. We report correlations between ground-based reservoir level measurements in the US and our remote sensing methods, and correlations between the cloud analysis and altimetry data for reservoirs in data-poor areas. The availability of regular satellite imagery and an automated, near real-time application method provides the necessary datasets for further temporal analysis, reservoir modeling, and flood forecasting. All statements of fact, analysis, or opinion are those of the author and do not reflect the official policy or position of the Department of Defense or any of its components or the U.S. Government
NASA Technical Reports Server (NTRS)
Edwards, J. W.; Deets, D. A.
1975-01-01
A cost-effective approach to flight testing advanced control concepts with remotely piloted vehicles is described. The approach utilizes a ground based digital computer coupled to the remotely piloted vehicle's motion sensors and control surface actuators through telemetry links to provide high bandwidth feedback control. The system was applied to the control of an unmanned 3/8-scale model of the F-15 airplane. The model was remotely augmented; that is, the F-15 mechanical and control augmentation flight control systems were simulated by the ground-based computer, rather than being in the vehicle itself. The results of flight tests of the model at high angles of attack are discussed.
NASA Astrophysics Data System (ADS)
Kerkez, B.; Zhang, Z.; Oroza, C.; Glaser, S. D.; Bales, R. C.
2012-12-01
We describe our improved, robust, and scalable architecture by which to rapidly instrument large-scale watersheds, while providing the resulting data in real-time. Our system consists of more than twenty wireless sensor networks and thousands of sensors, which will be deployed in the American River basin (5000 sq. km) of California. The core component of our system is known as a mote, a tiny, ultra-low-power, embedded wireless computer that can be used for any number of sensing applications. Our new generation of motes is equipped with IPv6 functionality, effectively giving each sensor in the field its own unique IP address, thus permitting users to remotely interact with the devices without going through intermediary services. Thirty to fifty motes will be deployed across 1-2 square kilometer regions to form a mesh-based wireless sensor network. Redundancy of local wireless links will ensure that data will always be able to traverse the network, even if hash wintertime conditions adversely affect some network nodes. These networks will be used to develop spatial estimates of a number of hydrologic parameters, focusing especially on snowpack. Each wireless sensor network has one main network controller, which is responsible with interacting with an embedded Linux computer to relay information across higher-powered, long-range wireless links (cell modems, satellite, WiFi) to neighboring networks and remote, offsite servers. The network manager is also responsible for providing an Internet connection to each mote. Data collected by the sensors can either be read directly by remote hosts, or stored on centralized servers for future access. With 20 such networks deployed in the American River, our system will comprise an unprecedented cyber-physical architecture for measuring hydrologic parameters in large-scale basins. The spatiotemporal density and real-time nature of the data is also expected to significantly improve operational hydrology and water resource management in the basin.
A high throughput geocomputing system for remote sensing quantitative retrieval and a case study
NASA Astrophysics Data System (ADS)
Xue, Yong; Chen, Ziqiang; Xu, Hui; Ai, Jianwen; Jiang, Shuzheng; Li, Yingjie; Wang, Ying; Guang, Jie; Mei, Linlu; Jiao, Xijuan; He, Xingwei; Hou, Tingting
2011-12-01
The quality and accuracy of remote sensing instruments have been improved significantly, however, rapid processing of large-scale remote sensing data becomes the bottleneck for remote sensing quantitative retrieval applications. The remote sensing quantitative retrieval is a data-intensive computation application, which is one of the research issues of high throughput computation. The remote sensing quantitative retrieval Grid workflow is a high-level core component of remote sensing Grid, which is used to support the modeling, reconstruction and implementation of large-scale complex applications of remote sensing science. In this paper, we intend to study middleware components of the remote sensing Grid - the dynamic Grid workflow based on the remote sensing quantitative retrieval application on Grid platform. We designed a novel architecture for the remote sensing Grid workflow. According to this architecture, we constructed the Remote Sensing Information Service Grid Node (RSSN) with Condor. We developed a graphic user interface (GUI) tools to compose remote sensing processing Grid workflows, and took the aerosol optical depth (AOD) retrieval as an example. The case study showed that significant improvement in the system performance could be achieved with this implementation. The results also give a perspective on the potential of applying Grid workflow practices to remote sensing quantitative retrieval problems using commodity class PCs.
Using NetMeeting for remote configuration of the Otto Bock C-Leg: technical considerations.
Lemaire, E D; Fawcett, J A
2002-08-01
Telehealth has the potential to be a valuable tool for technical and clinical support of computer controlled prosthetic devices. This pilot study examined the use of Internet-based, desktop video conferencing for remote configuration of the Otto Bock C-Leg. Laboratory tests involved connecting two computers running Microsoft NetMeeting over a local area network (IP protocol). Over 56 Kbs(-1), DSL/Cable, and 10 Mbs(-1) LAN speeds, a prosthetist remotely configured a user's C-Leg by using Application Sharing, Live Video, and Live Audio. A similar test between sites in Ottawa and Toronto, Canada was limited by the notebook computer's 28 Kbs(-1) modem. At the 28 Kbs(-1) Internet-connection speed, NetMeeting's application sharing feature was not able to update the remote Sliders window fast enough to display peak toe loads and peak knee angles. These results support the use of NetMeeting as an accessible and cost-effective tool for remote C-Leg configuration, provided that sufficient Internet data transfer speed is available.
Lab4CE: A Remote Laboratory for Computer Education
ERIC Educational Resources Information Center
Broisin, Julien; Venant, Rémi; Vidal, Philippe
2017-01-01
Remote practical activities have been demonstrated to be efficient when learners come to acquire inquiry skills. In computer science education, virtualization technologies are gaining popularity as this technological advance enables instructors to implement realistic practical learning activities, and learners to engage in authentic and…
Enabling a high throughput real time data pipeline for a large radio telescope array with GPUs
NASA Astrophysics Data System (ADS)
Edgar, R. G.; Clark, M. A.; Dale, K.; Mitchell, D. A.; Ord, S. M.; Wayth, R. B.; Pfister, H.; Greenhill, L. J.
2010-10-01
The Murchison Widefield Array (MWA) is a next-generation radio telescope currently under construction in the remote Western Australia Outback. Raw data will be generated continuously at 5 GiB s-1, grouped into 8 s cadences. This high throughput motivates the development of on-site, real time processing and reduction in preference to archiving, transport and off-line processing. Each batch of 8 s data must be completely reduced before the next batch arrives. Maintaining real time operation will require a sustained performance of around 2.5 TFLOP s-1 (including convolutions, FFTs, interpolations and matrix multiplications). We describe a scalable heterogeneous computing pipeline implementation, exploiting both the high computing density and FLOP-per-Watt ratio of modern GPUs. The architecture is highly parallel within and across nodes, with all major processing elements performed by GPUs. Necessary scatter-gather operations along the pipeline are loosely synchronized between the nodes hosting the GPUs. The MWA will be a frontier scientific instrument and a pathfinder for planned peta- and exa-scale facilities.
An introduction to quantitative remote sensing. [data processing
NASA Technical Reports Server (NTRS)
Lindenlaub, J. C.; Russell, J.
1974-01-01
The quantitative approach to remote sensing is discussed along with the analysis of remote sensing data. Emphasis is placed on the application of pattern recognition in numerically oriented remote sensing systems. A common background and orientation for users of the LARS computer software system is provided.
Remote Sensing between Liver and Intestine: Importance of Microbial Metabolites
Fu, Zidong Donna; Cui, Julia Yue
2017-01-01
Recent technological advancements including metagenomics sequencing and metabolomics have allowed the discovery of critical functions of gut microbiota in obesity, malnutrition, neurological disorders, asthma, and xenobiotic metabolism. Classification of the human gut microbiome into distinct “enterotypes” has been proposed to serve as a new paradigm for understanding the interplay between microbial variation and human disease phenotypes, as many organs are affected by gut microbiota modifications during the pathogenesis of diseases. Gut microbiota remotely interacts with liver and other metabolic organs of the host through various microbial metabolites that are absorbed into the systemic circulation. Purpose of review The present review summarizes recent literature regarding the importance of gut microbiota in modulating the physiological and pathological responses of various host organs, and describes the functions of the known microbial metabolites that are involved in this remote sensing process, with a primary focus on the gut microbiota-liver axis. Recent findings Under physiological conditions, gut microbiota modulates the hepatic transcriptome, proteome, and metabolome, most notably down-regulating cytochrome P450 3a mediated xenobiotic metabolism. Gut microbiome also modulates the rhythmicity in liver gene expression, likely through microbial metabolites, such as butyrate and propionate that serve as epigenetic modifiers. Additionally, the production of host hormones such as primary bile acids and glucagon like peptide 1 is altered by gut microbiota to modify intermediary metabolism of the host. Summary Dysregulation of gut microbiota is implicated in various liver diseases such as alcoholic liver disease, non-alcoholic steatohepatitis, liver cirrhosis, cholangitis, and liver cancer. Gut microbiota modifiers such as probiotics and prebiotics are increasingly recognized as novel therapeutic modalities for liver and other types of human diseases. PMID:28983453
NASA Technical Reports Server (NTRS)
Zwick, H.; Ward, V.; Beaudette, L.
1973-01-01
A critical evaluation of existing optical remote sensors for HCl vapor detection in solid propellant rocket plumes is presented. The P branch of the fundamental vibration-rotation band was selected as the most promising spectral feature to sense. A computation of transmittance for HCl vapor, an estimation of interferent spectra, the application of these spectra to computer modelled remote sensors, and a trade-off study for instrument recommendation are also included.
NASA Astrophysics Data System (ADS)
Heilmann, B. Z.; Vallenilla Ferrara, A. M.
2009-04-01
The constant growth of contaminated sites, the unsustainable use of natural resources, and, last but not least, the hydrological risk related to extreme meteorological events and increased climate variability are major environmental issues of today. Finding solutions for these complex problems requires an integrated cross-disciplinary approach, providing a unified basis for environmental science and engineering. In computer science, grid computing is emerging worldwide as a formidable tool allowing distributed computation and data management with administratively-distant resources. Utilizing these modern High Performance Computing (HPC) technologies, the GRIDA3 project bundles several applications from different fields of geoscience aiming to support decision making for reasonable and responsible land use and resource management. In this abstract we present a geophysical application called EIAGRID that uses grid computing facilities to perform real-time subsurface imaging by on-the-fly processing of seismic field data and fast optimization of the processing workflow. Even though, seismic reflection profiling has a broad application range spanning from shallow targets in a few meters depth to targets in a depth of several kilometers, it is primarily used by the hydrocarbon industry and hardly for environmental purposes. The complexity of data acquisition and processing poses severe problems for environmental and geotechnical engineering: Professional seismic processing software is expensive to buy and demands large experience from the user. In-field processing equipment needed for real-time data Quality Control (QC) and immediate optimization of the acquisition parameters is often not available for this kind of studies. As a result, the data quality will be suboptimal. In the worst case, a crucial parameter such as receiver spacing, maximum offset, or recording time turns out later to be inappropriate and the complete acquisition campaign has to be repeated. The EIAGRID portal provides an innovative solution to this problem combining state-of-the-art data processing methods and modern remote grid computing technology. In field-processing equipment is substituted by remote access to high performance grid computing facilities. The latter can be ubiquitously controlled by a user-friendly web-browser interface accessed from the field by any mobile computer using wireless data transmission technology such as UMTS (Universal Mobile Telecommunications System) or HSUPA/HSDPA (High-Speed Uplink/Downlink Packet Access). The complexity of data-manipulation and processing and thus also the time demanding user interaction is minimized by a data-driven, and highly automated velocity analysis and imaging approach based on the Common-Reflection-Surface (CRS) stack. Furthermore, the huge computing power provided by the grid deployment allows parallel testing of alternative processing sequences and parameter settings, a feature which considerably reduces the turn-around times. A shared data storage using georeferencing tools and data grid technology is under current development. It will allow to publish already accomplished projects, making results, processing workflows and parameter settings available in a transparent and reproducible way. Creating a unified database shared by all users will facilitate complex studies and enable the use of data-crossing techniques to incorporate results of other environmental applications hosted on the GRIDA3 portal.
Estimating costs and performance of systems for machine processing of remotely sensed data
NASA Technical Reports Server (NTRS)
Ballard, R. J.; Eastwood, L. F., Jr.
1977-01-01
This paper outlines a method for estimating computer processing times and costs incurred in producing information products from digital remotely sensed data. The method accounts for both computation and overhead, and may be applied to any serial computer. The method is applied to estimate the cost and computer time involved in producing Level II Land Use and Vegetative Cover Maps for a five-state midwestern region. The results show that the amount of data to be processed overloads some example computer systems, but that the processing is feasible on others.
Commodity Cluster Computing for Remote Sensing Applications using Red Hat LINUX
NASA Technical Reports Server (NTRS)
Dorband, John
2003-01-01
Since 1994, we have been doing research at Goddard Space Flight Center on implementing a wide variety of applications on commodity based computing clusters. This talk is about these clusters and haw they are used on these applications including ones for remote sensing.
RADIAL COMPUTED TOMOGRAPHY OF AIR CONTAMINANTS USING OPTICAL REMOTE SENSING
The paper describes the application of an optical remote-sensing (ORS) system to map air contaminants and locate fugitive emissions. Many ORD systems may utilize radial non-overlapping beam geometry and a computed tomography (CT) algorithm to map the concentrations in a plane. In...
An automated procedure for developing hybrid computer simulations of turbofan engines
NASA Technical Reports Server (NTRS)
Szuch, J. R.; Krosel, S. M.
1980-01-01
A systematic, computer-aided, self-documenting methodology for developing hybrid computer simulations of turbofan engines is presented. The methodology makes use of a host program that can run on a large digital computer and a machine-dependent target (hybrid) program. The host program performs all of the calculations and date manipulations needed to transform user-supplied engine design information to a form suitable for the hybrid computer. The host program also trims the self contained engine model to match specified design point information. A test case is described and comparisons between hybrid simulation and specified engine performance data are presented.
Lay Theories Regarding Computer-Mediated Communication in Remote Collaboration
ERIC Educational Resources Information Center
Parke, Karl; Marsden, Nicola; Connolly, Cornelia
2017-01-01
Computer-mediated communication and remote collaboration has become an unexceptional norm as an educational modality for distance and open education, therefore the need to research and analyze students' online learning experience is necessary. This paper seeks to examine the assumptions and expectations held by students in regard to…
The paper presents a new approach to quantifying emissions from fugitive gaseous air pollution sources. Computed tomography (CT) and path-integrated optical remote sensing (PI-ORS) concentration data are combined in a new field beam geometry. Path-integrated concentrations are ...
Automated Content Synthesis for Interactive Remote Instruction.
ERIC Educational Resources Information Center
Maly, K.; Overstreet, C. M.; Gonzalez, A.; Denbar, M. L.; Cutaran, R.; Karunaratne, N.
This paper describes IRI (Interactive Remote Instruction), a computer-based system built at Old Dominion University (Virginia) in order to support distance education. The system is based on the concept of a virtual classroom where students at different locations have the same synchronous class experience, using networked computers to communicate…
Techtalk: Telecommunications for Improving Developmental Education.
ERIC Educational Resources Information Center
Caverly, David C.; Broderick, Bill
1993-01-01
Explains how to access the Internet, discussing hardware and software considerations, connectivity, and types of access available to users. Describes the uses of electronic mail; TELNET, a method for remotely logging onto another computer; and anonymous File Transfer Protocol (FTP), a method for downloading files from a remote computer. (MAB)
Remote Viewing and Computer Communications--An Experiment.
ERIC Educational Resources Information Center
Vallee, Jacques
1988-01-01
A series of remote viewing experiments were run with 12 participants who communicated through a computer conferencing network. The correct target sample was identified in 8 out of 33 cases. This represented more than double the pure chance expectation. Appendices present protocol, instructions, and results of the experiments. (Author/YP)
Commercial future: making remote sensing a media event
NASA Astrophysics Data System (ADS)
Lurie, Ian
1999-12-01
The rapid growth of commercial remote sensing has made high quality digital sensing data widely available -- now, remote sensing must become and remain a strong, commercially viable industry. However, this new industry cannot survive without an educated consumer base. To access markets, remote sensing providers must make their product more accessible, both literally and figuratively: Potential customers must be able to find the data they require, when they require it, and they must understand the utility of the information available to them. The Internet and the World Wide Web offer the perfect medium to educate potential customers and to sell remote sensing data to those customers. A well-designed web presence can provide both an information center and a market place for companies offering their data for sale. A very high potential web-based market for remote sensing lies in media. News agencies, web sites, and a host of other visual media services can use remote sensing data to provide current, relevant information regarding news around the world. This paper will provide a model for promotion and sale of remote sensing data via the Internet.
NASA Technical Reports Server (NTRS)
Byer, R. L. (Editor); Trebino, R. (Editor); Gustafson, E. K. (Editor)
1985-01-01
Papers are presented on solid-state lasers for remote sensing, diode-pumped Nd:YAG lasers, and tunable solid-state-laser systems. Topics discussed include titanium-sapphire tunable laser systems, the performance of slab geometry, and the development of slab lasers. Consideration is given to garnet host solid-state lasers, the growth of lasers and nonlinear materials, and nonlinear frequency conversion and tunable sources.
REMOTE SENSING AND GIS FOR WETLANDS
In identifying and characterizing wetland and adjacent features, the use of remote sensor and Geographic Information Systems (GIS) technologies has been valuable. Remote sensors such as photographs and computer-sensor generated images can illustrate conditions of hydrology, exten...
Remote Visualization and Remote Collaboration On Computational Fluid Dynamics
NASA Technical Reports Server (NTRS)
Watson, Val; Lasinski, T. A. (Technical Monitor)
1995-01-01
A new technology has been developed for remote visualization that provides remote, 3D, high resolution, dynamic, interactive viewing of scientific data (such as fluid dynamics simulations or measurements). Based on this technology, some World Wide Web sites on the Internet are providing fluid dynamics data for educational or testing purposes. This technology is also being used for remote collaboration in joint university, industry, and NASA projects in computational fluid dynamics and wind tunnel testing. Previously, remote visualization of dynamic data was done using video format (transmitting pixel information) such as video conferencing or MPEG movies on the Internet. The concept for this new technology is to send the raw data (e.g., grids, vectors, and scalars) along with viewing scripts over the Internet and have the pixels generated by a visualization tool running on the viewer's local workstation. The visualization tool that is currently used is FAST (Flow Analysis Software Toolkit).
System architecture for asynchronous multi-processor robotic control system
NASA Technical Reports Server (NTRS)
Steele, Robert D.; Long, Mark; Backes, Paul
1993-01-01
The architecture for the Modular Telerobot Task Execution System (MOTES) as implemented in the Supervisory Telerobotics (STELER) Laboratory is described. MOTES is the software component of the remote site of a local-remote telerobotic system which is being developed for NASA for space applications, in particular Space Station Freedom applications. The system is being developed to provide control and supervised autonomous control to support both space based operation and ground-remote control with time delay. The local-remote architecture places task planning responsibilities at the local site and task execution responsibilities at the remote site. This separation allows the remote site to be designed to optimize task execution capability within a limited computational environment such as is expected in flight systems. The local site task planning system could be placed on the ground where few computational limitations are expected. MOTES is written in the Ada programming language for a multiprocessor environment.
Scattering Properties of Heterogeneous Mineral Particles with Absorbing Inclusions
NASA Technical Reports Server (NTRS)
Dlugach, Janna M.; Mishchenko, Michael I.
2015-01-01
We analyze the results of numerically exact computer modeling of scattering and absorption properties of randomly oriented poly-disperse heterogeneous particles obtained by placing microscopic absorbing grains randomly on the surfaces of much larger spherical mineral hosts or by imbedding them randomly inside the hosts. These computations are paralleled by those for heterogeneous particles obtained by fully encapsulating fractal-like absorbing clusters in the mineral hosts. All computations are performed using the superposition T-matrix method. In the case of randomly distributed inclusions, the results are compared with the outcome of Lorenz-Mie computations for an external mixture of the mineral hosts and absorbing grains. We conclude that internal aggregation can affect strongly both the integral radiometric and differential scattering characteristics of the heterogeneous particle mixtures.
Using Avizo Software on the Peregrine System | High-Performance Computing |
be run remotely from the Peregrine visualization node. First, launch a TurboVNC remote desktop. Then from a terminal in that remote desktop: % module load avizo % vglrun avizo Running Locally Avizo can
The role of the host in a cooperating mainframe and workstation environment, volumes 1 and 2
NASA Technical Reports Server (NTRS)
Kusmanoff, Antone; Martin, Nancy L.
1989-01-01
In recent years, advancements made in computer systems have prompted a move from centralized computing based on timesharing a large mainframe computer to distributed computing based on a connected set of engineering workstations. A major factor in this advancement is the increased performance and lower cost of engineering workstations. The shift to distributed computing from centralized computing has led to challenges associated with the residency of application programs within the system. In a combined system of multiple engineering workstations attached to a mainframe host, the question arises as to how does a system designer assign applications between the larger mainframe host and the smaller, yet powerful, workstation. The concepts related to real time data processing are analyzed and systems are displayed which use a host mainframe and a number of engineering workstations interconnected by a local area network. In most cases, distributed systems can be classified as having a single function or multiple functions and as executing programs in real time or nonreal time. In a system of multiple computers, the degree of autonomy of the computers is important; a system with one master control computer generally differs in reliability, performance, and complexity from a system in which all computers share the control. This research is concerned with generating general criteria principles for software residency decisions (host or workstation) for a diverse yet coupled group of users (the clustered workstations) which may need the use of a shared resource (the mainframe) to perform their functions.
Remote Science Operation Center research
NASA Technical Reports Server (NTRS)
Banks, P. M.
1986-01-01
Progress in the following areas is discussed: the design, planning and operation of a remote science payload operations control center; design and planning of a data link via satellite; and the design and prototyping of an advanced workstation environment for multi-media (3-D computer aided design/computer aided engineering, voice, video, text) communications and operations.
Computational approaches to predict bacteriophage–host relationships
Edwards, Robert A.; McNair, Katelyn; Faust, Karoline; Raes, Jeroen; Dutilh, Bas E.
2015-01-01
Metagenomics has changed the face of virus discovery by enabling the accurate identification of viral genome sequences without requiring isolation of the viruses. As a result, metagenomic virus discovery leaves the first and most fundamental question about any novel virus unanswered: What host does the virus infect? The diversity of the global virosphere and the volumes of data obtained in metagenomic sequencing projects demand computational tools for virus–host prediction. We focus on bacteriophages (phages, viruses that infect bacteria), the most abundant and diverse group of viruses found in environmental metagenomes. By analyzing 820 phages with annotated hosts, we review and assess the predictive power of in silico phage–host signals. Sequence homology approaches are the most effective at identifying known phage–host pairs. Compositional and abundance-based methods contain significant signal for phage–host classification, providing opportunities for analyzing the unknowns in viral metagenomes. Together, these computational approaches further our knowledge of the interactions between phages and their hosts. Importantly, we find that all reviewed signals significantly link phages to their hosts, illustrating how current knowledge and insights about the interaction mechanisms and ecology of coevolving phages and bacteria can be exploited to predict phage–host relationships, with potential relevance for medical and industrial applications. PMID:26657537
Wireless Control of Miniaturized Mobile Vehicle for Indoor Surveillance
NASA Astrophysics Data System (ADS)
Taha Saquib, Syed M.; Hameed, Sarmad; Usman Ali, Syed M.; Jafri, Raza; Amin, Imran
2013-12-01
This work is based upon electronic automation and Smart Control techniques, which constitute the basis of Control Area Network (CAN) and Personal Area Network (PAN). Bluetooth technology has been interfaced with a programmable controller to provide multi-dimensional vehicle control. A network is proposed which contains a remote, mobile host controller and an android operating system based mobile set (Client). The client communicates with a host controller through a Bluetooth device. The system incorporates duplex communication after successful confirmation between the host and the client; the android based mobile unit controls the vehicle through the Bluetooth module.
Walz, Yvonne; Wegmann, Martin; Dech, Stefan; Vounatsou, Penelope; Poda, Jean-Noël; N'Goran, Eliézer K.; Utzinger, Jürg; Raso, Giovanna
2015-01-01
Background Schistosomiasis is the most widespread water-based disease in sub-Saharan Africa. Transmission is governed by the spatial distribution of specific freshwater snails that act as intermediate hosts and human water contact patterns. Remote sensing data have been utilized for spatially explicit risk profiling of schistosomiasis. We investigated the potential of remote sensing to characterize habitat conditions of parasite and intermediate host snails and discuss the relevance for public health. Methodology We employed high-resolution remote sensing data, environmental field measurements, and ecological data to model environmental suitability for schistosomiasis-related parasite and snail species. The model was developed for Burkina Faso using a habitat suitability index (HSI). The plausibility of remote sensing habitat variables was validated using field measurements. The established model was transferred to different ecological settings in Côte d’Ivoire and validated against readily available survey data from school-aged children. Principal Findings Environmental suitability for schistosomiasis transmission was spatially delineated and quantified by seven habitat variables derived from remote sensing data. The strengths and weaknesses highlighted by the plausibility analysis showed that temporal dynamic water and vegetation measures were particularly useful to model parasite and snail habitat suitability, whereas the measurement of water surface temperature and topographic variables did not perform appropriately. The transferability of the model showed significant relations between the HSI and infection prevalence in study sites of Côte d’Ivoire. Conclusions/Significance A predictive map of environmental suitability for schistosomiasis transmission can support measures to gain and sustain control. This is particularly relevant as emphasis is shifting from morbidity control to interrupting transmission. Further validation of our mechanistic model needs to be complemented by field data of parasite- and snail-related fitness. Our model provides a useful tool to monitor the development of new hotspots of potential schistosomiasis transmission based on regularly updated remote sensing data. PMID:26587839
Walz, Yvonne; Wegmann, Martin; Dech, Stefan; Vounatsou, Penelope; Poda, Jean-Noël; N'Goran, Eliézer K; Utzinger, Jürg; Raso, Giovanna
2015-11-01
Schistosomiasis is the most widespread water-based disease in sub-Saharan Africa. Transmission is governed by the spatial distribution of specific freshwater snails that act as intermediate hosts and human water contact patterns. Remote sensing data have been utilized for spatially explicit risk profiling of schistosomiasis. We investigated the potential of remote sensing to characterize habitat conditions of parasite and intermediate host snails and discuss the relevance for public health. We employed high-resolution remote sensing data, environmental field measurements, and ecological data to model environmental suitability for schistosomiasis-related parasite and snail species. The model was developed for Burkina Faso using a habitat suitability index (HSI). The plausibility of remote sensing habitat variables was validated using field measurements. The established model was transferred to different ecological settings in Côte d'Ivoire and validated against readily available survey data from school-aged children. Environmental suitability for schistosomiasis transmission was spatially delineated and quantified by seven habitat variables derived from remote sensing data. The strengths and weaknesses highlighted by the plausibility analysis showed that temporal dynamic water and vegetation measures were particularly useful to model parasite and snail habitat suitability, whereas the measurement of water surface temperature and topographic variables did not perform appropriately. The transferability of the model showed significant relations between the HSI and infection prevalence in study sites of Côte d'Ivoire. A predictive map of environmental suitability for schistosomiasis transmission can support measures to gain and sustain control. This is particularly relevant as emphasis is shifting from morbidity control to interrupting transmission. Further validation of our mechanistic model needs to be complemented by field data of parasite- and snail-related fitness. Our model provides a useful tool to monitor the development of new hotspots of potential schistosomiasis transmission based on regularly updated remote sensing data.
Remote observing with the Nickel Telescope at Lick Observatory
NASA Astrophysics Data System (ADS)
Grigsby, Bryant; Chloros, Konstantinos; Gates, John; Deich, William T. S.; Gates, Elinor; Kibrick, Robert
2008-07-01
We describe a project to enable remote observing on the Nickel 1-meter Telescope at Lick Observatory. The purpose was to increase the subscription rate and create more economical means for graduate- and undergraduate students to observe with this telescope. The Nickel Telescope resides in a 125 year old dome on Mount Hamilton. Remote observers may work from any of the University of California (UC) remote observing facilities that have been created to support remote work at both Keck Observatory and Lick Observatory. The project included hardware and software upgrades to enable computer control of all equipment that must be operated by the astronomer; a remote observing architecture that is closely modeled on UCO/Lick's work to implement remote observing between UC campuses and Keck Observatory; new policies to ensure safety of Observatory staff and equipment, while ensuring that the telescope subsystems would be suitably configured for remote use; and new software to enforce the safety-related policies. The results increased the subscription rate from a few nights per month to nearly full subscription, and has spurred the installation of remote observing sites at more UC campuses. Thanks to the increased automation and computer control, local observing has also benefitted and is more efficient. Remote observing is now being implemented for the Shane 3- meter telescope.
Grid Computing for Earth Science
NASA Astrophysics Data System (ADS)
Renard, Philippe; Badoux, Vincent; Petitdidier, Monique; Cossu, Roberto
2009-04-01
The fundamental challenges facing humankind at the beginning of the 21st century require an effective response to the massive changes that are putting increasing pressure on the environment and society. The worldwide Earth science community, with its mosaic of disciplines and players (academia, industry, national surveys, international organizations, and so forth), provides a scientific basis for addressing issues such as the development of new energy resources; a secure water supply; safe storage of nuclear waste; the analysis, modeling, and mitigation of climate changes; and the assessment of natural and industrial risks. In addition, the Earth science community provides short- and medium-term prediction of weather and natural hazards in real time, and model simulations of a host of phenomena relating to the Earth and its space environment. These capabilities require that the Earth science community utilize, both in real and remote time, massive amounts of data, which are usually distributed among many different organizations and data centers.
The Design of an Autonomous Underwater Vehicle for Water Quality Monitoring
NASA Astrophysics Data System (ADS)
Li, Yulong; Liu, Rong; Liu, Shujin
2018-01-01
This paper describes the development of a civilian-used autonomous underwater vehicle (AUV) for water quality monitoring at reservoirs and watercourses that can obtain realtime visual and locational information. The mechanical design was completed with CAD software Solidworks. Four thrusters—two horizontal and two vertical—on board enable the vehicle to surge, heave, yaw, and pitch. A specialized water sample collection compartment is designed to perform water collection at target locations. The vehicle has a central controller—STM32—and a sub-coordinate controller—Arduino MEGA 2560—that coordinates multiple sensors including an inertial sensor, ultrasonic sensors, etc. Global Navigation Satellite System (GNSS) and the inertial sensor enable the vehicle’s localization. Remote operators monitor and control the vehicle via a host computer system. Operators choose either semi-autonomous mode in which they set target locations or manual mode. The experimental results show that the vehicle is able to perform well in either mode.
Hiding the Disk and Network Latency of Out-of-Core Visualization
NASA Technical Reports Server (NTRS)
Ellsworth, David
2001-01-01
This paper describes an algorithm that improves the performance of application-controlled demand paging for out-of-core visualization by hiding the latency of reading data from both local disks or disks on remote servers. The performance improvements come from better overlapping the computation with the page reading process, and by performing multiple page reads in parallel. The paper includes measurements that show that the new multithreaded paging algorithm decreases the time needed to compute visualizations by one third when using one processor and reading data from local disk. The time needed when using one processor and reading data from remote disk decreased by two thirds. Visualization runs using data from remote disk actually ran faster than ones using data from local disk because the remote runs were able to make use of the remote server's high performance disk array.
NASA Astrophysics Data System (ADS)
Tanner, Steve; Stein, Cara; Graves, Sara J.
Networks of remote sensors are becoming more common as technology improves and costs decline. In the past, a remote sensor was usually a device that collected data to be retrieved at a later time by some other mechanism. This collected data were usually processed well after the fact at a computer greatly removed from the in situ sensing location. This has begun to change as sensor technology, on-board processing, and network communication capabilities have increased and their prices have dropped. There has been an explosion in the number of sensors and sensing devices, not just around the world, but literally throughout the solar system. These sensors are not only becoming vastly more sophisticated, accurate, and detailed in the data they gather but they are also becoming cheaper, lighter, and smaller. At the same time, engineers have developed improved methods to embed computing systems, memory, storage, and communication capabilities into the platforms that host these sensors. Now, it is not unusual to see large networks of sensors working in cooperation with one another. Nor does it seem strange to see the autonomous operation of sensorbased systems, from space-based satellites to smart vacuum cleaners that keep our homes clean and robotic toys that help to entertain and educate our children. But access to sensor data and computing power is only part of the story. For all the power of these systems, there are still substantial limits to what they can accomplish. These include the well-known limits to current Artificial Intelligence capabilities and our limited ability to program the abstract concepts, goals, and improvisation needed for fully autonomous systems. But it also includes much more basic engineering problems such as lack of adequate power, communications bandwidth, and memory, as well as problems with the geolocation and real-time georeferencing required to integrate data from multiple sensors to be used together.
Arranging computer architectures to create higher-performance controllers
NASA Technical Reports Server (NTRS)
Jacklin, Stephen A.
1988-01-01
Techniques for integrating microprocessors, array processors, and other intelligent devices in control systems are reviewed, with an emphasis on the (re)arrangement of components to form distributed or parallel processing systems. Consideration is given to the selection of the host microprocessor, increasing the power and/or memory capacity of the host, multitasking software for the host, array processors to reduce computation time, the allocation of real-time and non-real-time events to different computer subsystems, intelligent devices to share the computational burden for real-time events, and intelligent interfaces to increase communication speeds. The case of a helicopter vibration-suppression and stabilization controller is analyzed as an example, and significant improvements in computation and throughput rates are demonstrated.
Treecode with a Special-Purpose Processor
NASA Astrophysics Data System (ADS)
Makino, Junichiro
1991-08-01
We describe an implementation of the modified Barnes-Hut tree algorithm for a gravitational N-body calculation on a GRAPE (GRAvity PipE) backend processor. GRAPE is a special-purpose computer for N-body calculations. It receives the positions and masses of particles from a host computer and then calculates the gravitational force at each coordinate specified by the host. To use this GRAPE processor with the hierarchical tree algorithm, the host computer must maintain a list of all nodes that exert force on a particle. If we create this list for each particle of the system at each timestep, the number of floating-point operations on the host and that on GRAPE would become comparable, and the increased speed obtained by using GRAPE would be small. In our modified algorithm, we create a list of nodes for many particles. Thus, the amount of the work required of the host is significantly reduced. This algorithm was originally developed by Barnes in order to vectorize the force calculation on a Cyber 205. With this algorithm, the computing time of the force calculation becomes comparable to that of the tree construction, if the GRAPE backend processor is sufficiently fast. The obtained speed-up factor is 30 to 50 for a RISC-based host computer and GRAPE-1A with a peak speed of 240 Mflops.
NASA Astrophysics Data System (ADS)
Meurant, Robert C.
Sorting of Korean English-as-a-Foreign-Language (EFL) university students by Second Language (L2) aptitude allocates students to classes of compatible ability level, and was here used to screen candidates for interview. Paper-and-pen versions of the Oxford Quick Placement Test were adapted to computer-based testing via online hosting using FSCreations ExamView. Problems with their online hosting site led to conversion to the popular computer-based learning management system Moodle, hosted on www.ninehub.com. 317 sophomores were tested online to encourage L2 digital literacy. Strategies for effective hybrid implementation of Learning Management Systems in L2 tertiary education include computer-based Internet-hosted L2 aptitude tests. These potentially provide a convenient measure of student progress in developing L2 fluency, and offer a more objective and relevant means of teacher- and course-assessment than student evaluations, which tend to confuse entertainment value and teacher popularity with academic credibility and pedagogical effectiveness.
Encryption for Remote Control via Internet or Intranet
NASA Technical Reports Server (NTRS)
Lineberger, Lewis
2005-01-01
A data-communication protocol has been devised to enable secure, reliable remote control of processes and equipment via a collision-based network, while using minimal bandwidth and computation. The network could be the Internet or an intranet. Control is made secure by use of both a password and a dynamic key, which is sent transparently to a remote user by the controlled computer (that is, the computer, located at the site of the equipment or process to be controlled, that exerts direct control over the process). The protocol functions in the presence of network latency, overcomes errors caused by missed dynamic keys, and defeats attempts by unauthorized remote users to gain control. The protocol is not suitable for real-time control, but is well suited for applications in which control latencies up to about 0.5 second are acceptable. The encryption scheme involves the use of both a dynamic and a private key, without any additional overhead that would degrade performance. The dynamic key is embedded in the equipment- or process-monitor data packets sent out by the controlled computer: in other words, the dynamic key is a subset of the data in each such data packet. The controlled computer maintains a history of the last 3 to 5 data packets for use in decrypting incoming control commands. In addition, the controlled computer records a private key (password) that is given to the remote computer. The encrypted incoming command is permuted by both the dynamic and private key. A person who records the command data in a given packet for hostile purposes cannot use that packet after the public key expires (typically within 3 seconds). Even a person in possession of an unauthorized copy of the command/remote-display software cannot use that software in the absence of the password. The use of a dynamic key embedded in the outgoing data makes the central-processing unit overhead very small. The use of a National Instruments DataSocket(TradeMark) (or equivalent) protocol or the User Datagram Protocol makes it possible to obtain reasonably short response times: Typical response times in event-driven control, using packets sized .300 bytes, are <0.2 second for commands issued from locations anywhere on Earth. The protocol requires that control commands represent absolute values of controlled parameters (e.g., a specified temperature), as distinguished from changes in values of controlled parameters (e.g., a specified increment of temperature). Each command is issued three or more times to ensure delivery in crowded networks. The use of absolute-value commands prevents additional (redundant) commands from causing trouble. Because a remote controlling computer receives "talkback" in the form of data packets from the controlled computer, typically within a time interval < or =1 s, the controlling computer can re-issue a command if network failure has occurred. The controlled computer, the process or equipment that it controls, and any human operator(s) at the site of the controlled equipment or process should be equipped with safety measures to prevent damage to equipment or injury to humans. These features could be a combination of software, external hardware, and intervention by the human operator(s). The protocol is not fail-safe, but by adopting these safety measures as part of the protocol, one makes the protocol a robust means of controlling remote processes and equipment by use of typical office computers via intranets and/or the Internet.
Protecting software agents from malicious hosts using quantum computing
NASA Astrophysics Data System (ADS)
Reisner, John; Donkor, Eric
2000-07-01
We evaluate how quantum computing can be applied to security problems for software agents. Agent-based computing, which merges technological advances in artificial intelligence and mobile computing, is a rapidly growing domain, especially in applications such as electronic commerce, network management, information retrieval, and mission planning. System security is one of the more eminent research areas in agent-based computing, and the specific problem of protecting a mobile agent from a potentially hostile host is one of the most difficult of these challenges. In this work, we describe our agent model, and discuss the capabilities and limitations of classical solutions to the malicious host problem. Quantum computing may be extremely helpful in addressing the limitations of classical solutions to this problem. This paper highlights some of the areas where quantum computing could be applied to agent security.
Accelerating Demand Paging for Local and Remote Out-of-Core Visualization
NASA Technical Reports Server (NTRS)
Ellsworth, David
2001-01-01
This paper describes a new algorithm that improves the performance of application-controlled demand paging for the out-of-core visualization of data sets that are on either local disks or disks on remote servers. The performance improvements come from better overlapping the computation with the page reading process, and by performing multiple page reads in parallel. The new algorithm can be applied to many different visualization algorithms since application-controlled demand paging is not specific to any visualization algorithm. The paper includes measurements that show that the new multi-threaded paging algorithm decreases the time needed to compute visualizations by one third when using one processor and reading data from local disk. The time needed when using one processor and reading data from remote disk decreased by up to 60%. Visualization runs using data from remote disk ran about as fast as ones using data from local disk because the remote runs were able to make use of the remote server's high performance disk array.
a Hadoop-Based Distributed Framework for Efficient Managing and Processing Big Remote Sensing Images
NASA Astrophysics Data System (ADS)
Wang, C.; Hu, F.; Hu, X.; Zhao, S.; Wen, W.; Yang, C.
2015-07-01
Various sensors from airborne and satellite platforms are producing large volumes of remote sensing images for mapping, environmental monitoring, disaster management, military intelligence, and others. However, it is challenging to efficiently storage, query and process such big data due to the data- and computing- intensive issues. In this paper, a Hadoop-based framework is proposed to manage and process the big remote sensing data in a distributed and parallel manner. Especially, remote sensing data can be directly fetched from other data platforms into the Hadoop Distributed File System (HDFS). The Orfeo toolbox, a ready-to-use tool for large image processing, is integrated into MapReduce to provide affluent image processing operations. With the integration of HDFS, Orfeo toolbox and MapReduce, these remote sensing images can be directly processed in parallel in a scalable computing environment. The experiment results show that the proposed framework can efficiently manage and process such big remote sensing data.
NASA Astrophysics Data System (ADS)
Liu, Jia; Liu, Longli; Xue, Yong; Dong, Jing; Hu, Yingcui; Hill, Richard; Guang, Jie; Li, Chi
2017-01-01
Workflow for remote sensing quantitative retrieval is the ;bridge; between Grid services and Grid-enabled application of remote sensing quantitative retrieval. Workflow averts low-level implementation details of the Grid and hence enables users to focus on higher levels of application. The workflow for remote sensing quantitative retrieval plays an important role in remote sensing Grid and Cloud computing services, which can support the modelling, construction and implementation of large-scale complicated applications of remote sensing science. The validation of workflow is important in order to support the large-scale sophisticated scientific computation processes with enhanced performance and to minimize potential waste of time and resources. To research the semantic correctness of user-defined workflows, in this paper, we propose a workflow validation method based on tacit knowledge research in the remote sensing domain. We first discuss the remote sensing model and metadata. Through detailed analysis, we then discuss the method of extracting the domain tacit knowledge and expressing the knowledge with ontology. Additionally, we construct the domain ontology with Protégé. Through our experimental study, we verify the validity of this method in two ways, namely data source consistency error validation and parameters matching error validation.
NASA Technical Reports Server (NTRS)
Szuch, J. R.; Krosel, S. M.; Bruton, W. M.
1982-01-01
A systematic, computer-aided, self-documenting methodology for developing hybrid computer simulations of turbofan engines is presented. The methodology that is pesented makes use of a host program that can run on a large digital computer and a machine-dependent target (hybrid) program. The host program performs all the calculations and data manipulations that are needed to transform user-supplied engine design information to a form suitable for the hybrid computer. The host program also trims the self-contained engine model to match specified design-point information. Part I contains a general discussion of the methodology, describes a test case, and presents comparisons between hybrid simulation and specified engine performance data. Part II, a companion document, contains documentation, in the form of computer printouts, for the test case.
Core Technical Capability Laboratory Management System
NASA Technical Reports Server (NTRS)
Shaykhian, Linda; Dugger, Curtis; Griffin, Laurie
2008-01-01
The Core Technical Capability Lab - oratory Management System (CTCLMS) consists of dynamically generated Web pages used to access a database containing detailed CTC lab data with the software hosted on a server that allows users to have remote access.
Somayaji, R; Lynch, T; Powell, J N; Gregson, D
2016-11-04
Lactobacillus spp. are uncommon pathogens in immunocompetent hosts, and even rarer causes of prosthetic device infections. A case of chronic hip prosthetic joint infection (PJI) caused by L. animalis is described. This occurred 5 years after a transient bacteremia with the same organism. Whole genome sequencing of both isolates proved this PJI infection resulted from this remote bacteremia. We document that prosthetic joint infections may be a consequence of bacteremia as much as 3 years before the onset of symptoms.
Value-Range Analysis of C Programs
NASA Astrophysics Data System (ADS)
Simon, Axel
In 1988, Robert T. Morris exploited a so-called buffer-overflow bug in finger (a dæmon whose job it is to return information on local users) to mount a denial-of-service attack on hundreds of VAX and Sun-3 computers [159]. He created what is nowadays called a worm; that is, a crafted stream of bytes that, when sent to a computer over the network, utilises a buffer-overflow bug in the software of that computer to execute code encoded in the byte stream. In the case of a worm, this code will send the very same byte stream to other computers on the network, thereby creating an avalanche of network traffic that ultimately renders the network and all computers involved in replicating the worm inaccessible. Besides duplicating themselves, worms can alter data on the host that they are running on. The most famous example in recent years was the MSBlaster32 worm, which altered the configuration database on many Microsoft Windows machines, thereby forcing the computers to reboot incessantly. Although this worm was rather benign, it caused huge damage to businesses who were unable to use their IT infrastructure for hours or even days after the appearance of the worm. A more malicious worm is certainly conceivable [187] due to the fact that worms are executed as part of a dæmon (also known as "service" on Windows machines) and thereby run at a privileged level, allowing access to any data stored on the remote computer. While the deletion of data presents a looming threat to valuable information, even more serious uses are espionage and theft, in particular because worms do not have to affect the running system and hence may be impossible to detect.
Mobile Transactional Modelling: From Concepts to Incremental Knowledge
NASA Astrophysics Data System (ADS)
Launders, Ivan; Polovina, Simon; Hill, Richard
In 1988, Robert T. Morris exploited a so-called buffer-overflow bug in finger (a dæmon whose job it is to return information on local users) to mount a denial-of-service attack on hundreds of VAX and Sun-3 computers [159]. He created what is nowadays called a worm; that is, a crafted stream of bytes that, when sent to a computer over the network, utilises a buffer-overflow bug in the software of that computer to execute code encoded in the byte stream. In the case of a worm, this code will send the very same byte stream to other computers on the network, thereby creating an avalanche of network traffic that ultimately renders the network and all computers involved in replicating the worm inaccessible. Besides duplicating themselves, worms can alter data on the host that they are running on. The most famous example in recent years was the MSBlaster32 worm, which altered the configuration database on many Microsoft Windows machines, thereby forcing the computers to reboot incessantly. Although this worm was rather benign, it caused huge damage to businesses who were unable to use their IT infrastructure for hours or even days after the appearance of the worm. A more malicious worm is certainly conceivable [187] due to the fact that worms are executed as part of a dæmon (also known as "service" on Windows machines) and thereby run at a privileged level, allowing access to any data stored on the remote computer. While the deletion of data presents a looming threat to valuable information, even more serious uses are espionage and theft, in particular because worms do not have to affect the running system and hence may be impossible to detect.
The Penn State ORSER system for processing and analyzing ERTS and other MSS data
NASA Technical Reports Server (NTRS)
Mcmurtry, G. J.; Petersen, G. W. (Principal Investigator); Borden, F. Y.; Weeden, H. A.
1974-01-01
The author has identified the following significant results. The office for Remote Sensing of Earth Resources (ORSER) of the Space Science and Engineering Laboratory at the Pennsylvania State University has developed an extensive operational system for processing and analyzing ERTS-1 and similar multispectral data. The ORSER system was developed for use by a wide variety of researchers working in remote sensing. Both photointerpretive techniques and automatic computer processing methods have been developed and used, separately and in a combined approach. A remote Job Entry system permits use of an IBM 370/168 computer from any compatible remote terminal, including equipment tied in by long distance telephone connections. An elementary cost analysis has been prepared for the processing of ERTS data.
Design and Verification of Remote Sensing Image Data Center Storage Architecture Based on Hadoop
NASA Astrophysics Data System (ADS)
Tang, D.; Zhou, X.; Jing, Y.; Cong, W.; Li, C.
2018-04-01
The data center is a new concept of data processing and application proposed in recent years. It is a new method of processing technologies based on data, parallel computing, and compatibility with different hardware clusters. While optimizing the data storage management structure, it fully utilizes cluster resource computing nodes and improves the efficiency of data parallel application. This paper used mature Hadoop technology to build a large-scale distributed image management architecture for remote sensing imagery. Using MapReduce parallel processing technology, it called many computing nodes to process image storage blocks and pyramids in the background to improve the efficiency of image reading and application and sovled the need for concurrent multi-user high-speed access to remotely sensed data. It verified the rationality, reliability and superiority of the system design by testing the storage efficiency of different image data and multi-users and analyzing the distributed storage architecture to improve the application efficiency of remote sensing images through building an actual Hadoop service system.
Tracking Avian Reservoirs of Arboviruses using Remote Sensing and Radiotelemetry
NASA Technical Reports Server (NTRS)
Beck, L.; Wright, S.; Schmidt, C.; Lobitz, B.; Bell, D.; Brown, D.; Brass, James A. (Technical Monitor)
2002-01-01
Encephalitis is caused by a virus that is transmitted by mosquitoes between mammalian hosts. The virus is closely related to the West Nile virus (WNV), which started in New York in 1999, and has since spread to 25 states. Like encephalitis, WNV is vectored by mosquitoes, and the primary hosts are birds; humans are accidental, or'dead-end' hosts. Very little is understood about the behavior of these bird populations, and how they intersect - both in time and in space - with mosquito populations. Exploring these relationships is the first step in developing models for encephalitis and WNV transmission risk. This project combines remotely sensed data with radiotelemetry to create a spatiotemporal map of encephalitis viral activity in bird and mosquito populations in the Sacramento Valley of California. Specifically, remote sensing (RS) and geographic information system (GIS) technologies were used to characterize habitats utilized by both avian viral reservoirs and the mosquitoes that vector encephalitis. Radiotelemetry and serosurveys (blood) were then used to spatially and temporally track the patterns of infection. The project uses Landsat ETM+ multitemporal satellite data to characterize habitats utilized by both birds and the mosquito vectors. Mist nets were used to sample members of individual flocks of blackbirds and cowbirds over a period of several months; these birds were then bled to assess their viral status, banded, and fitted with transmitters. Radiotelemetry was used to spatially and temporally track the distribution of banded birds and their associated flocks. The movement of these indicator flocks were compared with the location of remotely sensed (adult and larval) mosquito habitats to determine the intersection of bird's and vectors; this is key in understanding where and when transmission occurs from bird to bird, as well as from bird to mammal, via mosquito. The relationships found during the project are being used to generate a model of encephalitis transmission risk in California.
Towards Dynamic Remote Data Auditing in Computational Clouds
Khurram Khan, Muhammad; Anuar, Nor Badrul
2014-01-01
Cloud computing is a significant shift of computational paradigm where computing as a utility and storing data remotely have a great potential. Enterprise and businesses are now more interested in outsourcing their data to the cloud to lessen the burden of local data storage and maintenance. However, the outsourced data and the computation outcomes are not continuously trustworthy due to the lack of control and physical possession of the data owners. To better streamline this issue, researchers have now focused on designing remote data auditing (RDA) techniques. The majority of these techniques, however, are only applicable for static archive data and are not subject to audit the dynamically updated outsourced data. We propose an effectual RDA technique based on algebraic signature properties for cloud storage system and also present a new data structure capable of efficiently supporting dynamic data operations like append, insert, modify, and delete. Moreover, this data structure empowers our method to be applicable for large-scale data with minimum computation cost. The comparative analysis with the state-of-the-art RDA schemes shows that the proposed scheme is secure and highly efficient in terms of the computation and communication overhead on the auditor and server. PMID:25121114
Towards dynamic remote data auditing in computational clouds.
Sookhak, Mehdi; Akhunzada, Adnan; Gani, Abdullah; Khurram Khan, Muhammad; Anuar, Nor Badrul
2014-01-01
Cloud computing is a significant shift of computational paradigm where computing as a utility and storing data remotely have a great potential. Enterprise and businesses are now more interested in outsourcing their data to the cloud to lessen the burden of local data storage and maintenance. However, the outsourced data and the computation outcomes are not continuously trustworthy due to the lack of control and physical possession of the data owners. To better streamline this issue, researchers have now focused on designing remote data auditing (RDA) techniques. The majority of these techniques, however, are only applicable for static archive data and are not subject to audit the dynamically updated outsourced data. We propose an effectual RDA technique based on algebraic signature properties for cloud storage system and also present a new data structure capable of efficiently supporting dynamic data operations like append, insert, modify, and delete. Moreover, this data structure empowers our method to be applicable for large-scale data with minimum computation cost. The comparative analysis with the state-of-the-art RDA schemes shows that the proposed scheme is secure and highly efficient in terms of the computation and communication overhead on the auditor and server.
NASA Astrophysics Data System (ADS)
Kulinkina, A. V.; Walz, Y.; Liss, A.; Kosinski, K. C.; Biritwum, N. K.; Naumova, E. N.
2016-06-01
Schistosoma haematobium transmission is influenced by environmental conditions that determine the suitability of the parasite and intermediate host snail habitats, as well as by socioeconomic conditions, access to water and sanitation infrastructure, and human behaviors. Remote sensing is a demonstrated valuable tool to characterize environmental conditions that support schistosomiasis transmission. Socioeconomic and behavioral conditions that propagate repeated domestic and recreational surface water contact are more difficult to quantify at large spatial scales. We present a mixed-methods approach that builds on the remotely sensed ecological variables by exploring water and sanitation related community characteristics as independent risk factors of schistosomiasis transmission.
Wolinski, Christophe Czeslaw [Los Alamos, NM; Gokhale, Maya B [Los Alamos, NM; McCabe, Kevin Peter [Los Alamos, NM
2011-01-18
Fabric-based computing systems and methods are disclosed. A fabric-based computing system can include a polymorphous computing fabric that can be customized on a per application basis and a host processor in communication with said polymorphous computing fabric. The polymorphous computing fabric includes a cellular architecture that can be highly parameterized to enable a customized synthesis of fabric instances for a variety of enhanced application performances thereof. A global memory concept can also be included that provides the host processor random access to all variables and instructions associated with the polymorphous computing fabric.
Hand controller commonality evaluation process
NASA Technical Reports Server (NTRS)
Stuart, Mark A.; Bierschwale, John M.; Wilmington, Robert P.; Adam, Susan C.; Diaz, Manuel F.; Jensen, Dean G.
1990-01-01
A hand controller evaluation process has been developed to determine the appropriate hand controller configurations for supporting remotely controlled devices. These devices include remote manipulator systems (RMS), dexterous robots, and remotely-piloted free flyers. Standard interfaces were developed to evaluate six different hand controllers in three test facilities including dynamic computer simulations, kinematic computer simulations, and physical simulations. The hand controllers under consideration were six degree-of-freedom (DOF) position and rate minimaster and joystick controllers, and three-DOF rate controllers. Task performance data, subjective comments, and anthropometric data obtained during tests were used for controller configuration recommendations to the SSF Program.
Study to design and develop remote manipulator system
NASA Technical Reports Server (NTRS)
Hill, J. W.; Sword, A. J.
1973-01-01
Human performance measurement techniques for remote manipulation tasks and remote sensing techniques for manipulators are described for common manipulation tasks, performance is monitored by means of an on-line computer capable of measuring the joint angles of both master and slave arms as a function of time. The computer programs allow measurements of the operator's strategy and physical quantities such as task time and power consumed. The results are printed out after a test run to compare different experimental conditions. For tracking tasks, we describe a method of displaying errors in three dimensions and measuring the end-effector position in three dimensions.
NASA Technical Reports Server (NTRS)
Maxwell, E. L.
1980-01-01
The need for degree programs in remote sensing is considered. Any education program which claims to train remote sensing specialists must include expertise in the physical principles upon which remote sensing is based. These principles dictate the limits of engineering and design, computer analysis, photogrammetry, and photointerpretation. Faculty members must be hired to provide emphasis in those five areas.
Gritzo, R.E.
1985-09-12
A remote reset circuit acts as a stand-along monitor and controller by clocking in each character sent by a terminal to a computer and comparing it to a given reference character. When a match occurs, the remote reset circuit activates the system's hardware reset line. The remote reset circuit is hardware based centered around monostable multivibrators and is unaffected by system crashes, partial serial transmissions, or power supply transients. 4 figs.
Gritzo, Russell E.
1987-01-01
A remote reset circuit acts as a stand-alone monitor and controller by clocking in each character sent by a terminal to a computer and comparing it to a given reference character. When a match occurs, the remote reset circuit activates the system's hardware reset line. The remote reset circuit is hardware based centered around monostable multivibrators and is unaffected by system crashes, partial serial transmissions, or power supply transients.
Overview of the land analysis system (LAS)
Quirk, Bruce K.; Olseson, Lyndon R.
1987-01-01
The Land Analysis System (LAS) is a fully integrated digital analysis system designed to support remote sensing, image processing, and geographic information systems research. LAS is being developed through a cooperative effort between the National Aeronautics and Space Administration Goddard Space Flight Center and the U. S. Geological Survey Earth Resources Observation Systems (EROS) Data Center. LAS has over 275 analysis modules capable to performing input and output, radiometric correction, geometric registration, signal processing, logical operations, data transformation, classification, spatial analysis, nominal filtering, conversion between raster and vector data types, and display manipulation of image and ancillary data. LAS is currently implant using the Transportable Applications Executive (TAE). While TAE was designed primarily to be transportable, it still provides the necessary components for a standard user interface, terminal handling, input and output services, display management, and intersystem communications. With TAE the analyst uses the same interface to the processing modules regardless of the host computer or operating system. LAS was originally implemented at EROS on a Digital Equipment Corporation computer system under the Virtual Memorial System operating system with DeAnza displays and is presently being converted to run on a Gould Power Node and Sun workstation under the Berkeley System Distribution UNIX operating system.
NASA Technical Reports Server (NTRS)
Grantham, C.
1979-01-01
The Interactive Software Invocation (ISIS), an interactive data management system, was developed to act as a buffer between the user and host computer system. The user is provided by ISIS with a powerful system for developing software or systems in the interactive environment. The user is protected from the idiosyncracies of the host computer system by providing such a complete range of capabilities that the user should have no need for direct access to the host computer. These capabilities are divided into four areas: desk top calculator, data editor, file manager, and tool invoker.
Elementary Age Children and Remote Sensing: Research from Project Omega.
ERIC Educational Resources Information Center
Kirman, Joseph M.
1991-01-01
Discusses remote sensing technology use in teaching elementary school students about science and social studies. Reviews findings dealing with the use of remote sensing and considering children's abilities, teacher training, computer applications, gifted children, and sex-related differences. Concludes that children as young as grade three can…
Inquiry-Based Learning in Remote Sensing: A Space Balloon Educational Experiment
ERIC Educational Resources Information Center
Mountrakis, Giorgos; Triantakonstantis, Dimitrios
2012-01-01
Teaching remote sensing in higher education has been traditionally restricted in lecture and computer-aided laboratory activities. This paper presents and evaluates an engaging inquiry-based educational experiment. The experiment was incorporated in an introductory remote sensing undergraduate course to bridge the gap between theory and…
Fully Engaging Students in the Remote Sensing Process through Field Experience
ERIC Educational Resources Information Center
Rundquist, Bradley C.; Vandeberg, Gregory S.
2013-01-01
Field data collection is often crucial to the success of investigations based upon remotely sensed data. Students of environmental remote sensing typically learn about the discipline through classroom lectures, a textbook, and computer laboratory sessions focused on the interpretation and processing of aircraft and satellite data. The importance…
Design and Implementation Issues for Modern Remote Laboratories
ERIC Educational Resources Information Center
Guimaraes, E. G.; Cardozo, E.; Moraes, D. H.; Coelho, P. R.
2011-01-01
The design and implementation of remote laboratories present different levels of complexity according to the nature of the equipments operated by the remote laboratory, the requirements imposed on the accessing computers, the network linking the user to the laboratory, and the type of experiments the laboratory supports. This paper addresses the…
NASA Astrophysics Data System (ADS)
Moreno, H. A.; Ogden, F. L.; Steinke, R. C.; Alvarez, L. V.
2015-12-01
Triangulated Irregular Networks (TINs) are increasingly popular for terrain representation in high performance surface and hydrologic modeling by their skill to capture significant changes in surface forms such as topographical summits, slope breaks, ridges, valley floors, pits and cols. This work presents a methodology for estimating slope, aspect and the components of the incoming solar radiation by using a vectorial approach within a topocentric coordinate system by establishing geometric relations between groups of TIN elements and the sun position. A normal vector to the surface of each TIN element describes slope and aspect while spherical trigonometry allows computing a unit vector defining the position of the sun at each hour and DOY. Thus, a dot product determines the radiation flux at each TIN element. Remote shading is computed by scanning the projection of groups of TIN elements in the direction of the closest perpendicular plane to the sun vector. Sky view fractions are computed by a simplified scanning algorithm in prescribed directions and are useful to determine diffuse radiation. Finally, remote radiation scattering is computed from the sky view factor complementary functions for prescribed albedo values of the surrounding terrain only for significant angles above the horizon. This methodology represents an improvement on the current algorithms to compute terrain and radiation parameters on TINs in an efficient manner. All terrain features (e.g. slope, aspect, sky view factors and remote sheltering) can be pre-computed and stored for easy access for a subsequent ground surface or hydrologic simulation.
On-demand provisioning of HEP compute resources on cloud sites and shared HPC centers
NASA Astrophysics Data System (ADS)
Erli, G.; Fischer, F.; Fleig, G.; Giffels, M.; Hauth, T.; Quast, G.; Schnepf, M.; Heese, J.; Leppert, K.; Arnaez de Pedro, J.; Sträter, R.
2017-10-01
This contribution reports on solutions, experiences and recent developments with the dynamic, on-demand provisioning of remote computing resources for analysis and simulation workflows. Local resources of a physics institute are extended by private and commercial cloud sites, ranging from the inclusion of desktop clusters over institute clusters to HPC centers. Rather than relying on dedicated HEP computing centers, it is nowadays more reasonable and flexible to utilize remote computing capacity via virtualization techniques or container concepts. We report on recent experience from incorporating a remote HPC center (NEMO Cluster, Freiburg University) and resources dynamically requested from the commercial provider 1&1 Internet SE into our intitute’s computing infrastructure. The Freiburg HPC resources are requested via the standard batch system, allowing HPC and HEP applications to be executed simultaneously, such that regular batch jobs run side by side to virtual machines managed via OpenStack [1]. For the inclusion of the 1&1 commercial resources, a Python API and SDK as well as the possibility to upload images were available. Large scale tests prove the capability to serve the scientific use case in the European 1&1 datacenters. The described environment at the Institute of Experimental Nuclear Physics (IEKP) at KIT serves the needs of researchers participating in the CMS and Belle II experiments. In total, resources exceeding half a million CPU hours have been provided by remote sites.
Airborne remote sensing for geology and the environment; present and future
Watson, Ken; Knepper, Daniel H.
1994-01-01
In 1988, a group of leading experts from government, academia, and industry attended a workshop on airborne remote sensing sponsored by the U.S. Geological Survey (USGS) and hosted by the Branch of Geophysics. The purpose of the workshop was to examine the scientific rationale for airborne remote sensing in support of government earth science in the next decade. This report has arranged the six resulting working-group reports under two main headings: (1) Geologic Remote Sensing, for the reports on geologic mapping, mineral resources, and fossil fuels and geothermal resources; and (2) Environmental Remote Sensing, for the reports on environmental geology, geologic hazards, and water resources. The intent of the workshop was to provide an evaluation of demonstrated capabilities, their direct extensions, and possible future applications, and this was the organizational format used for the geologic remote sensing reports. The working groups in environmental remote sensing chose to present their reports in a somewhat modified version of this format. A final section examines future advances and limitations in the field. There is a large, complex, and often bewildering array of remote sensing data available. Early remote sensing studies were based on data collected from airborne platforms. Much of that technology was later extended to satellites. The original 80-m-resolution Landsat Multispectral Scanner System (MSS) has now been largely superseded by the 30-m-resolution Thematic Mapper (TM) system that has additional spectral channels. The French satellite SPOT provides higher spatial resolution for channels equivalent to MSS. Low-resolution (1 km) data are available from the National Oceanographic and Atmospheric Administration's AVHRR system, which acquires reflectance and day and night thermal data daily. Several experimental satellites have acquired limited data, and there are extensive plans for future satellites including those of Japan (JERS), Europe (ESA), Canada (Radarsat), and the United States (EOS). There are currently two national airborne remote sensing programs (photography, radar) with data archived at the USGS' EROS Data Center. Airborne broadband multispectral data (comparable to Landsat MSS and TM but involving several more channels) for limited geographic areas also are available for digital processing and analysis. Narrow-band imaging spectrometer data are available for some NASA experiment sites and can be acquired for other locations commercially. Remote sensing data and derivative images, because of the uniform spatial coverage, availability at different resolutions, and digital format, are becoming important data sets for geographic information system (GIS) analyses. Examples range from overlaying digitized geologic maps on remote sensing images and draping these over topography, to maps of mineral distribution and inferred abundance. A large variety of remote sensing data sets are available, with costs ranging from a few dollars per square mile for satellite digital data to a few hundred dollars per square mile for airborne imaging spectrometry. Computer processing and analysis costs routinely surpass these expenses because of the equipment and expertise necessary for information extraction and interpretation. Effective use requires both an understanding of the current methodology and an appreciation of the most cost-effective solution.
Griffin, Bronwyn R; Kimble, Roy M; Watt, Kerrianne; Shields, Linda
2018-04-01
The main objective of this study is to describe incidence rates of low-speed vehicle run-over (LSVRO) events among children aged 0-14 years residing in Queensland from 1999 to 2009. A second objective was to describe the associated patterns of injury, with respect to gender, age group, severity, characteristics (host, vehicle and environment), and trends over time in relation to geographical remoteness. Final results are hoped to inform prevention policies. In this statewide, retrospective, population-based study, data were collected on LSVRO events that occurred among children aged 0-14 years in Queensland from 1999 to 2009 from all relevant data sources across the continuum of care, and manually linked to obtain the most comprehensive estimate possible of the magnitude and nature of LSVRO events to date. Crude incidence rates were calculated separately for males and females, for fatal events, non-fatal events (hospital admissions and non-admissions, respectively), and for all LSVRO events, for each area of geographical remoteness (major cities, inner regional, outer regional, remote/very remote). Relative risks and 95% confidence interval were calculated, and trends over time were examined. Data on host, injury and event characteristics were also obtained to investigate whether these characteristics varied between areas of remoteness. Incidence rates were lowest among children (0-14 years) living in major cities (13.8/100 000/annum, with the highest recorded incidence in outer regional areas (incidence rate =42.5/100 000/annum). Incidence rates were higher for children residing outside major cities for both males and females, for every age group, for each of the 11 years of the study, and consequences of LSVRO events were worse. Young children aged 0-4 years were identified as those most at risk for these events, regardless of geographical location. Differences were observed as a function of remoteness category in relation to injury characteristics (eg injury type), and host characteristics (eg sociodemographic status), but there were no observed differences in environmental characteristics (eg time of day, day of week). Heavy vehicles such as four-wheel drives, utilities, trucks and tractors were more frequently involved in LSVRO events that occurred outside major cities. The results confirmed that children of all ages and genders residing outside of major cities in Queensland are more at risk of being involved in an LSVRO incident, and experience more severe consequences compared to children in major cities. Future research should address the specific risk factors and focus on engaging rural communities to assist in the prevention of LSVRO incidents.
Processing module operating methods, processing modules, and communications systems
McCown, Steven Harvey; Derr, Kurt W.; Moore, Troy
2014-09-09
A processing module operating method includes using a processing module physically connected to a wireless communications device, requesting that the wireless communications device retrieve encrypted code from a web site and receiving the encrypted code from the wireless communications device. The wireless communications device is unable to decrypt the encrypted code. The method further includes using the processing module, decrypting the encrypted code, executing the decrypted code, and preventing the wireless communications device from accessing the decrypted code. Another processing module operating method includes using a processing module physically connected to a host device, executing an application within the processing module, allowing the application to exchange user interaction data communicated using a user interface of the host device with the host device, and allowing the application to use the host device as a communications device for exchanging information with a remote device distinct from the host device.
1994-06-28
developing Unmanned Aerial Vehicles, not for military use, but for civilian use3, such as remote news coverage and remote tourism by broadcasting live...Interoperability, and Integration of (’ommand, (Control, (’ ommunications , Computers, and Intelligence Systems. CJCS Instruction no. 6212.01, Washington, D.C.: U.S
Distributed solar radiation fast dynamic measurement for PV cells
NASA Astrophysics Data System (ADS)
Wan, Xuefen; Yang, Yi; Cui, Jian; Du, Xingjing; Zheng, Tao; Sardar, Muhammad Sohail
2017-10-01
To study the operating characteristics about PV cells, attention must be given to the dynamic behavior of the solar radiation. The dynamic behaviors of annual, monthly, daily and hourly averages of solar radiation have been studied in detail. But faster dynamic behaviors of solar radiation need more researches. The solar radiation random fluctuations in minute-long or second-long range, which lead to alternating radiation and cool down/warm up PV cell frequently, decrease conversion efficiency. Fast dynamic processes of solar radiation are mainly relevant to stochastic moving of clouds. Even in clear sky condition, the solar irradiations show a certain degree of fast variation. To evaluate operating characteristics of PV cells under fast dynamic irradiation, a solar radiation measuring array (SRMA) based on large active area photodiode, LoRa spread spectrum communication and nanoWatt MCU is proposed. This cross photodiodes structure tracks fast stochastic moving of clouds. To compensate response time of pyranometer and reduce system cost, the terminal nodes with low-cost fast-responded large active area photodiode are placed besides positions of tested PV cells. A central node, consists with pyranometer, large active area photodiode, wind detector and host computer, is placed in the center of the central topologies coordinate to scale temporal envelope of solar irradiation and get calibration information between pyranometer and large active area photodiodes. In our SRMA system, the terminal nodes are designed based on Microchip's nanoWatt XLP PIC16F1947. FDS-100 is adopted for large active area photodiode in terminal nodes and host computer. The output current and voltage of each PV cell are monitored by I/V measurement. AS62-T27/SX1278 LoRa communication modules are used for communicating between terminal nodes and host computer. Because the LoRa LPWAN (Low Power Wide Area Network) specification provides seamless interoperability among Smart Things without the need of complex local installations, configuring of our SRMA system is very easy. Lora also provides SRMA a means to overcome the short communication distance and weather signal propagation decline such as in ZigBee and WiFi. The host computer in SRMA system uses the low power single-board PC EMB-3870 which was produced by NORCO. Wind direction sensor SM5386B and wind-force sensor SM5387B are installed to host computer through RS-485 bus for wind reference data collection. And Davis 6450 solar radiation sensor, which is a precision instrument that detects radiation at wavelengths of 300 to 1100 nanometers, allow host computer to follow real-time solar radiation. A LoRa polling scheme is adopt for the communication between host computer and terminal nodes in SRMA. An experimental SRMA has been established. This system was tested in Ganyu, Jiangshu province from May to August, 2016. In the test, the distances between the nodes and the host computer were between 100m and 1900m. At work, SRMA system showed higher reliability. Terminal nodes could follow the instructions from host computer and collect solar radiation data of distributed PV cells effectively. And the host computer managed the SRAM and achieves reference parameters well. Communications between the host computer and terminal nodes were almost unaffected by the weather. In conclusion, the testing results show that SRMA could be a capable method for fast dynamic measuring about solar radiation and related PV cell operating characteristics.
CSI computer system/remote interface unit acceptance test results
NASA Technical Reports Server (NTRS)
Sparks, Dean W., Jr.
1992-01-01
The validation tests conducted on the Control/Structures Interaction (CSI) Computer System (CCS)/Remote Interface Unit (RIU) is discussed. The CCS/RIU consists of a commercially available, Langley Research Center (LaRC) programmed, space flight qualified computer and a flight data acquisition and filtering computer, developed at LaRC. The tests were performed in the Space Structures Research Laboratory (SSRL) and included open loop excitation, closed loop control, safing, RIU digital filtering, and RIU stand alone testing with the CSI Evolutionary Model (CEM) Phase-0 testbed. The test results indicated that the CCS/RIU system is comparable to ground based systems in performing real-time control-structure experiments.
Technology and the future of medical equipment maintenance.
Wear, J O
1999-05-01
Maintenance of medical equipment has been changing rapidly in the past few years. It is changing more rapidly in developed countries, but changes are also occurring in developing countries. Some of the changes may permit improved maintenance on the higher technology equipment in developing countries, since they do not require onsite expertise. Technology has had an increasing impact on the development of medical equipment with the increased use of microprocessors and computers. With miniaturization from space technology and electronic chip design, powerful microprocessors and computers have been built into medical equipment. The improvement in manufacturing technology has increased the quality of parts and therefore the medical equipment. This has resulted in increased mean time between failures and reduced maintenance needs. This has made equipment more reliable in remote areas and developing countries. The built-in computers and advances in software design have brought about self-diagnostics in medical equipment. The technicians now have a strong tool to be used in maintenance. One problem in this area is getting access to the self-diagnostics. Some manufacturers will not readily provide this access to the owner of the equipment. Advances in telecommunications in conjunction with self-diagnostics make available remote diagnosis and repair. Since components can no longer be repaired, a remote repair technician can instruct an operator or an on-site repairman on board replacement. In case of software problems, the remote repair technician may perform the repairs over the telephone. It is possible for the equipment to be monitored remotely by modern without interfering with the operation of the equipment. These changes in technology require the training of biomedical engineering technicians (BMETs) to change. They must have training in computers and telecommunications. Some of this training can be done with telecommunications and computers.
Synchronous, Remote, Internet Conferencing with Unique Populations in Various Settings.
ERIC Educational Resources Information Center
Mallory, James R.; MacKenzie, Douglas
This paper focuses on the authors' experiences with interactive, synchronous Internet video conferencing using Microsoft's NetMeeting software with deaf and hard-of-hearing students in two different settings. One setting involved teaching and tutoring computer programming to remote deaf and hard-of-hearing students in a remote situation using…
Distributed computing testbed for a remote experimental environment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Butner, D.N.; Casper, T.A.; Howard, B.C.
1995-09-18
Collaboration is increasing as physics research becomes concentrated on a few large, expensive facilities, particularly in magnetic fusion energy research, with national and international participation. These facilities are designed for steady state operation and interactive, real-time experimentation. We are developing tools to provide for the establishment of geographically distant centers for interactive operations; such centers would allow scientists to participate in experiments from their home institutions. A testbed is being developed for a Remote Experimental Environment (REE), a ``Collaboratory.`` The testbed will be used to evaluate the ability of a remotely located group of scientists to conduct research on themore » DIII-D Tokamak at General Atomics. The REE will serve as a testing environment for advanced control and collaboration concepts applicable to future experiments. Process-to-process communications over high speed wide area networks provide real-time synchronization and exchange of data among multiple computer networks, while the ability to conduct research is enhanced by adding audio/video communication capabilities. The Open Software Foundation`s Distributed Computing Environment is being used to test concepts in distributed control, security, naming, remote procedure calls and distributed file access using the Distributed File Services. We are exploring the technology and sociology of remotely participating in the operation of a large scale experimental facility.« less
ERIC Educational Resources Information Center
Zadahmad, Manouchehr; Yousefzadehfard, Parisa
2016-01-01
Mobile Cloud Computing (MCC) aims to improve all mobile applications such as m-learning systems. This study presents an innovative method to use web technology and software engineering's best practices to provide m-learning functionalities hosted in a MCC-learning system as service. Components hosted by MCC are used to empower developers to create…
Defense Acquisitions Acronyms and Terms
2012-12-01
Computer-Aided Design CADD Computer-Aided Design and Drafting CAE Component Acquisition Executive; Computer-Aided Engineering CAIV Cost As an...Radiation to Ordnance HFE Human Factors Engineering HHA Health Hazard Assessment HNA Host-Nation Approval HNS Host-Nation Support HOL High -Order...Engineering Change Proposal VHSIC Very High Speed Integrated Circuit VLSI Very Large Scale Integration VOC Volatile Organic Compound W WAN Wide
NASA Technical Reports Server (NTRS)
Lewis, A. J.; Isaacson, D. L.; Schrumpf, B. J. (Principal Investigator)
1980-01-01
Projects completed for the NASA Office of University Affairs include the application of remote sensing data in support of rehabilitation of wild fire damaged areas and the use of LANDSAT 3 return beam vidicon in forestry mapping applications. Continuing projects for that office include monitoring western Oregon timber clearcut; detecting and monitoring wheat disease; land use monitoring for tax assessment in Umatilla, Lake, and Morrow Counties; and the use of Oregon Air National Guard thermal infrared scanning data. Projects funded through other agencies include the remote sensing inventory of elk in the Blue Mountains; the estimation of burned agricultural acreage in the Willamette Valley; a resource inventory of Deschutes County; and hosting a LANDSAT digital workshop.
Karpievitch, Yuliya V; Almeida, Jonas S
2006-01-01
Background Matlab, a powerful and productive language that allows for rapid prototyping, modeling and simulation, is widely used in computational biology. Modeling and simulation of large biological systems often require more computational resources then are available on a single computer. Existing distributed computing environments like the Distributed Computing Toolbox, MatlabMPI, Matlab*G and others allow for the remote (and possibly parallel) execution of Matlab commands with varying support for features like an easy-to-use application programming interface, load-balanced utilization of resources, extensibility over the wide area network, and minimal system administration skill requirements. However, all of these environments require some level of access to participating machines to manually distribute the user-defined libraries that the remote call may invoke. Results mGrid augments the usual process distribution seen in other similar distributed systems by adding facilities for user code distribution. mGrid's client-side interface is an easy-to-use native Matlab toolbox that transparently executes user-defined code on remote machines (i.e. the user is unaware that the code is executing somewhere else). Run-time variables are automatically packed and distributed with the user-defined code and automated load-balancing of remote resources enables smooth concurrent execution. mGrid is an open source environment. Apart from the programming language itself, all other components are also open source, freely available tools: light-weight PHP scripts and the Apache web server. Conclusion Transparent, load-balanced distribution of user-defined Matlab toolboxes and rapid prototyping of many simple parallel applications can now be done with a single easy-to-use Matlab command. Because mGrid utilizes only Matlab, light-weight PHP scripts and the Apache web server, installation and configuration are very simple. Moreover, the web-based infrastructure of mGrid allows for it to be easily extensible over the Internet. PMID:16539707
Karpievitch, Yuliya V; Almeida, Jonas S
2006-03-15
Matlab, a powerful and productive language that allows for rapid prototyping, modeling and simulation, is widely used in computational biology. Modeling and simulation of large biological systems often require more computational resources then are available on a single computer. Existing distributed computing environments like the Distributed Computing Toolbox, MatlabMPI, Matlab*G and others allow for the remote (and possibly parallel) execution of Matlab commands with varying support for features like an easy-to-use application programming interface, load-balanced utilization of resources, extensibility over the wide area network, and minimal system administration skill requirements. However, all of these environments require some level of access to participating machines to manually distribute the user-defined libraries that the remote call may invoke. mGrid augments the usual process distribution seen in other similar distributed systems by adding facilities for user code distribution. mGrid's client-side interface is an easy-to-use native Matlab toolbox that transparently executes user-defined code on remote machines (i.e. the user is unaware that the code is executing somewhere else). Run-time variables are automatically packed and distributed with the user-defined code and automated load-balancing of remote resources enables smooth concurrent execution. mGrid is an open source environment. Apart from the programming language itself, all other components are also open source, freely available tools: light-weight PHP scripts and the Apache web server. Transparent, load-balanced distribution of user-defined Matlab toolboxes and rapid prototyping of many simple parallel applications can now be done with a single easy-to-use Matlab command. Because mGrid utilizes only Matlab, light-weight PHP scripts and the Apache web server, installation and configuration are very simple. Moreover, the web-based infrastructure of mGrid allows for it to be easily extensible over the Internet.
NASA Astrophysics Data System (ADS)
Ramalingam, Srikumar
2001-11-01
A highly secure mobile agent system is very important for a mobile computing environment. The security issues in mobile agent system comprise protecting mobile hosts from malicious agents, protecting agents from other malicious agents, protecting hosts from other malicious hosts and protecting agents from malicious hosts. Using traditional security mechanisms the first three security problems can be solved. Apart from using trusted hardware, very few approaches exist to protect mobile code from malicious hosts. Some of the approaches to solve this problem are the use of trusted computing, computing with encrypted function, steganography, cryptographic traces, Seal Calculas, etc. This paper focuses on the simulation of some of these existing techniques in the designed mobile language. Some new approaches to solve malicious network problem and agent tampering problem are developed using public key encryption system and steganographic concepts. The approaches are based on encrypting and hiding the partial solutions of the mobile agents. The partial results are stored and the address of the storage is destroyed as the agent moves from one host to another host. This allows only the originator to make use of the partial results. Through these approaches some of the existing problems are solved.
Test-bed for the remote health monitoring system for bridge structures using FBG sensors
NASA Astrophysics Data System (ADS)
Lee, Chin-Hyung; Park, Ki-Tae; Joo, Bong-Chul; Hwang, Yoon-Koog
2009-05-01
This paper reports on test-bed for the long-term health monitoring system for bridge structures employing fiber Bragg grating (FBG) sensors, which is remotely accessible via the web, to provide real-time quantitative information on a bridge's response to live loading and environmental changes, and fast prediction of the structure's integrity. The sensors are attached on several locations of the structure and connected to a data acquisition system permanently installed onsite. The system can be accessed through remote communication using an optical cable network, through which the evaluation of the bridge behavior under live loading can be allowed at place far away from the field. Live structural data are transmitted continuously to the server computer at the central office. The server computer is connected securely to the internet, where data can be retrieved, processed and stored for the remote web-based health monitoring. Test-bed revealed that the remote health monitoring technology will enable practical, cost-effective, and reliable condition assessment and maintenance of bridge structures.
Analyzing Resources of United States Marine Corps for Humanitarian Operations
2014-08-26
years in the petrochemical, semiconductor, paper and pulp products, and steel industries, focusing on enabling corporate strategy by using the supply... Disease monitoring in remote areas Accurate information from host nation Clear procedures from DOS Clear areas of responsibility Collaboration
Remote voice training: A case study on space shuttle applications, appendix C
NASA Technical Reports Server (NTRS)
Mollakarimi, Cindy; Hamid, Tamin
1990-01-01
The Tile Automation System includes applications of automation and robotics technology to all aspects of the Shuttle tile processing and inspection system. An integrated set of rapid prototyping testbeds was developed which include speech recognition and synthesis, laser imaging systems, distributed Ada programming environments, distributed relational data base architectures, distributed computer network architectures, multi-media workbenches, and human factors considerations. Remote voice training in the Tile Automation System is discussed. The user is prompted over a headset by synthesized speech for the training sequences. The voice recognition units and the voice output units are remote from the user and are connected by Ethernet to the main computer system. A supervisory channel is used to monitor the training sequences. Discussions include the training approaches as well as the human factors problems and solutions for this system utilizing remote training techniques.
Web Environment for Programming and Control of a Mobile Robot in a Remote Laboratory
ERIC Educational Resources Information Center
dos Santos Lopes, Maísa Soares; Gomes, Iago Pacheco; Trindade, Roque M. P.; da Silva, Alzira F.; de C. Lima, Antonio C.
2017-01-01
Remote robotics laboratories have been successfully used for engineering education. However, few of them use mobile robots to to teach computer science. This article describes a mobile robot Control and Programming Environment (CPE) and its pedagogical applications. The system comprises a remote laboratory for robotics, an online programming tool,…
Information Power Grid Posters
NASA Technical Reports Server (NTRS)
Vaziri, Arsi
2003-01-01
This document is a summary of the accomplishments of the Information Power Grid (IPG). Grids are an emerging technology that provide seamless and uniform access to the geographically dispersed, computational, data storage, networking, instruments, and software resources needed for solving large-scale scientific and engineering problems. The goal of the NASA IPG is to use NASA's remotely located computing and data system resources to build distributed systems that can address problems that are too large or complex for a single site. The accomplishments outlined in this poster presentation are: access to distributed data, IPG heterogeneous computing, integration of large-scale computing node into distributed environment, remote access to high data rate instruments,and exploratory grid environment.
Telepresence system development for application to the control of remote robotic systems
NASA Technical Reports Server (NTRS)
Crane, Carl D., III; Duffy, Joseph; Vora, Rajul; Chiang, Shih-Chien
1989-01-01
The recent developments of techniques which assist an operator in the control of remote robotic systems are described. In particular, applications are aimed at two specific scenarios: The control of remote robot manipulators; and motion planning for remote transporter vehicles. Common to both applications is the use of realistic computer graphics images which provide the operator with pertinent information. The specific system developments for several recently completed and ongoing telepresence research projects are described.
Evidence acquisition tools for cyber sex crimes investigations
NASA Astrophysics Data System (ADS)
Novotny, Jon M.; Meehan, A.; Schulte, D.; Manes, Gavin W.; Shenoi, Sujeet
2002-08-01
Sexually explicit Internet chat rooms are increasingly used by pedophiles to reach potential victims. Logging and linking suspects to chat room conversations and e-mails exchanged with undercover detectives are crucial to prosecuting travelers, i.e., pedophiles who travel across state lines to engage in sexual acts with minors. This paper describes two tools, a chat room monitor and a remote fingerprinter, for acquiring and preserving evidence. The chat room monitor logs online communications as well as screen images and keystrokes of the undercover detective. stored to allow the chronological reconstruction and replay of the investigation. The remote fingerprinter uses sophisticated scanning techniques to capture and preserve a unique fingerprint of the suspect's computer over the Internet. Once the suspect's computer is seized, it is scanned again; matching this new fingerprint with the remotely acquired fingerprint establishes that the suspect's computer was used to communicate with the detective.
Watanabe, S; Tanaka, M; Wada, Y; Suzuki, H; Takagi, S; Mori, S; Fukai, K; Kanazawa, Y; Takagi, M; Hirakawa, K; Ogasawara, K; Tsumura, K; Ogawa, K; Matsumoto, K; Nagaoka, S; Suzuki, T; Shimura, D; Yamashita, M; Nishio, S
1994-07-01
The telescience testbed experiments were carried out to test and investigate the tele-manipulation techniques in the intracellular potential recording of amphibian eggs. Implementation of telescience testbed was set up in the two separated laboratories of the Tsukuba Space center of NASDA, which were connected by tele-communication links. Manipulators respective for a microelectrode and a sample stage of microscope were moved by computers, of which command signals were transmitted from a computer in a remote control room. The computer in the control room was operated by an investigator (PI) who controlled the movement of each manipulator remotely. A stereoscopic vision of the microscope image were prepared by using a head mounted display (HMD) and were indispensable to the intracellular single cell recording. The fertilization potential of amphibian eggs was successfully obtained through the remote operating system.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-01-11
... agency social media sites, and other agency communications or publicity media as a consistent branding... NARA web and social media sites (intranet and internet), whether hosted internally, remotely, or on a... agency correspondence and other communications and publicity media. The other three logos are for new...
Phosphor suspended in silicone, molded/formed and used in a remote phosphor configuration
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kolodin, Boris; Deshpande, Anirudha R
A light emitting package comprising a support hosting at least one light emitting diode. A light transmissive dome comprised of a silicone including a phosphor material positioned to receive light emitted by the diode. A glass cap overlies said dome.
EPA uses high-end scientific computing, geospatial services and remote sensing/imagery analysis to support EPA's mission. The Center for Environmental Computing (CEC) assists the Agency's program offices and regions to meet staff needs in these areas.
Case report: teledermatology and epiluminescence microscopy for the diagnosis of scabies.
Weinstock, M A; Kempton, S A
2000-07-01
We wish to share images from a patient seen in our teledermatology program. Due to the absence of on-site dermatology services at the Togus, Maine, Department of Veterans Affairs, and associated community clinics for veterans in Aroostook, Bangor, Calais, and Rumford, we created a program to provide dermatologic expertise from Providence, Rhode Island. Patients referred for this service were evaluated by a nurse practitioner, who obtained a history, performed a physical examination, and captured digital images of the affected area of skin, including epiluminescence microscopic images where indicated. These data were then retrieved at the Providence (host) site and reviewed by a dermatologist, who formulated an impression and plan that was then implemented by the remote site in Maine. This approach, which involves image capture at the remote site and later review of images at the host site, is the "store-and-forward" method, which appears to be a relatively cost-effective means of providing this service from a distance.
Transparent process migration: Design alternatives and the Sprite implementation
NASA Technical Reports Server (NTRS)
Douglis, Fred; Ousterhout, John
1991-01-01
The Sprite operating system allows executing processes to be moved between hosts at any time. We use this process migration mechanism to offload work onto idle machines, and also to evict migrated processes when idle workstations are reclaimed by their owners. Sprite's migration mechanism provides a high degree of transparency both for migrated processes and for users. Idle machines are identified, and eviction is invoked, automatically by daemon processes. On Sprite it takes up to a few hundred milliseconds on SPARCstation 1 workstations to perform a remote exec, while evictions typically occur in a few seconds. The pmake program uses remote invocation to invoke tasks concurrently. Compilations commonly obtain speedup factors in the range of three to six; they are limited primarily by contention for centralized resources such as file servers. CPU-bound tasks such as simulations can make more effective use of idle hosts, obtaining as much as eight-fold speedup over a period of hours. Process migration has been in regular service for over two years.
Study to design and develop remote manipulator system. [computer simulation of human performance
NASA Technical Reports Server (NTRS)
Hill, J. W.; Mcgovern, D. E.; Sword, A. J.
1974-01-01
Modeling of human performance in remote manipulation tasks is reported by automated procedures using computers to analyze and count motions during a manipulation task. Performance is monitored by an on-line computer capable of measuring the joint angles of both master and slave and in some cases the trajectory and velocity of the hand itself. In this way the operator's strategies with different transmission delays, displays, tasks, and manipulators can be analyzed in detail for comparison. Some progress is described in obtaining a set of standard tasks and difficulty measures for evaluating manipulator performance.
NASA Technical Reports Server (NTRS)
Barber, Bryan; Kahn, Laura; Wong, David
1990-01-01
Offshore operations such as oil drilling and radar monitoring require semisubmersible platforms to remain stationary at specific locations in the Gulf of Mexico. Ocean currents, wind, and waves in the Gulf of Mexico tend to move platforms away from their desired locations. A computer model was created to predict the station keeping requirements of a platform. The computer simulation uses remote sensing data from satellites and buoys as input. A background of the project, alternate approaches to the project, and the details of the simulation are presented.
A Robust and Effective Smart-Card-Based Remote User Authentication Mechanism Using Hash Function
Odelu, Vanga; Goswami, Adrijit
2014-01-01
In a remote user authentication scheme, a remote server verifies whether a login user is genuine and trustworthy, and also for mutual authentication purpose a login user validates whether the remote server is genuine and trustworthy. Several remote user authentication schemes using the password, the biometrics, and the smart card have been proposed in the literature. However, most schemes proposed in the literature are either computationally expensive or insecure against several known attacks. In this paper, we aim to propose a new robust and effective password-based remote user authentication scheme using smart card. Our scheme is efficient, because our scheme uses only efficient one-way hash function and bitwise XOR operations. Through the rigorous informal and formal security analysis, we show that our scheme is secure against possible known attacks. We perform the simulation for the formal security analysis using the widely accepted AVISPA (Automated Validation Internet Security Protocols and Applications) tool to ensure that our scheme is secure against passive and active attacks. Furthermore, our scheme supports efficiently the password change phase always locally without contacting the remote server and correctly. In addition, our scheme performs significantly better than other existing schemes in terms of communication, computational overheads, security, and features provided by our scheme. PMID:24892078
A robust and effective smart-card-based remote user authentication mechanism using hash function.
Das, Ashok Kumar; Odelu, Vanga; Goswami, Adrijit
2014-01-01
In a remote user authentication scheme, a remote server verifies whether a login user is genuine and trustworthy, and also for mutual authentication purpose a login user validates whether the remote server is genuine and trustworthy. Several remote user authentication schemes using the password, the biometrics, and the smart card have been proposed in the literature. However, most schemes proposed in the literature are either computationally expensive or insecure against several known attacks. In this paper, we aim to propose a new robust and effective password-based remote user authentication scheme using smart card. Our scheme is efficient, because our scheme uses only efficient one-way hash function and bitwise XOR operations. Through the rigorous informal and formal security analysis, we show that our scheme is secure against possible known attacks. We perform the simulation for the formal security analysis using the widely accepted AVISPA (Automated Validation Internet Security Protocols and Applications) tool to ensure that our scheme is secure against passive and active attacks. Furthermore, our scheme supports efficiently the password change phase always locally without contacting the remote server and correctly. In addition, our scheme performs significantly better than other existing schemes in terms of communication, computational overheads, security, and features provided by our scheme.
NASA Technical Reports Server (NTRS)
Lefebvre, D. R.; Sanderson, A. C.
1994-01-01
Robot coordination and control systems for remote teleoperation applications are by necessity implemented on distributed computers. Modeling and performance analysis of these distributed robotic systems is difficult, but important for economic system design. Performance analysis methods originally developed for conventional distributed computer systems are often unsatisfactory for evaluating real-time systems. The paper introduces a formal model of distributed robotic control systems; and a performance analysis method, based on scheduling theory, which can handle concurrent hard-real-time response specifications. Use of the method is illustrated by a case of remote teleoperation which assesses the effect of communication delays and the allocation of robot control functions on control system hardware requirements.
Mpeg2 codec HD improvements with medical and robotic imaging benefits
NASA Astrophysics Data System (ADS)
Picard, Wayne F. J.
2010-02-01
In this report, we propose an efficient scheme to use High Definition Television (HDTV) in a console or notebook format as a computer terminal in addition to their role as TV display unit. In the proposed scheme, we assume that the main computer is situated at a remote location. The computer raster in the remote server is compressed using an HD E- >Mpeg2 encoder and transmitted to the terminal at home. The built-in E->Mpeg2 decoder in the terminal decompresses the compressed bit stream, and displays the raster. The terminal will be fitted with a mouse and keyboard, through which the interaction with the remote computer server can be performed via a communications back channel. The terminal in a notebook format can thus be used as a high resolution computer and multimedia device. We will consider developments such as the required HD enhanced Mpeg2 resolution (E->Mpeg2) and its medical ramifications due to improvements on compressed image quality with 2D to 3D conversion (Mpeg3) and using the compressed Discrete Cosine Transform coefficients in the reality compression of vision and control of medical robotic surgeons.
ERIC Educational Resources Information Center
Childers, Gina; Jones, M. Gail
2015-01-01
Remote access technologies enable students to investigate science by utilizing scientific tools and communicating in real-time with scientists and researchers with only a computer and an Internet connection. Very little is known about student perceptions of how real remote investigations are and how immersed the students are in the experience.…
Practical applications of remote sensing technology
NASA Technical Reports Server (NTRS)
Whitmore, Roy A., Jr.
1990-01-01
Land managers increasingly are becoming dependent upon remote sensing and automated analysis techniques for information gathering and synthesis. Remote sensing and geographic information system (GIS) techniques provide quick and economical information gathering for large areas. The outputs of remote sensing classification and analysis are most effective when combined with a total natural resources data base within the capabilities of a computerized GIS. Some examples are presented of the successes, as well as the problems, in integrating remote sensing and geographic information systems. The need to exploit remotely sensed data and the potential that geographic information systems offer for managing and analyzing such data continues to grow. New microcomputers with vastly enlarged memory, multi-fold increases in operating speed and storage capacity that was previously available only on mainframe computers are a reality. Improved raster GIS software systems have been developed for these high performance microcomputers. Vector GIS systems previously reserved for mini and mainframe systems are available to operate on these enhanced microcomputers. One of the more exciting areas that is beginning to emerge is the integration of both raster and vector formats on a single computer screen. This technology will allow satellite imagery or digital aerial photography to be presented as a background to a vector display.
Writing Apprehension, Computer Anxiety and Telecomputing: A Pilot Study.
ERIC Educational Resources Information Center
Harris, Judith; Grandgenett, Neal
1992-01-01
A study measured graduate students' writing apprehension and computer anxiety levels before and after using electronic mail, computer conferencing, and remote database searching facilities during an educational technology course. Results indicted postcourse computer anxiety levels significantly related to usage statistics. Precourse writing…
A New Definition for Ground Control
NASA Technical Reports Server (NTRS)
2002-01-01
LandForm(R) VisualFlight(R) blends the power of a geographic information system with the speed of a flight simulator to transform a user's desktop computer into a "virtual cockpit." The software product, which is fully compatible with all Microsoft(R) Windows(R) operating systems, provides distributed, real-time three-dimensional flight visualization over a host of networks. From a desktop, a user can immediately obtain a cockpit view, a chase-plane view, or an airborne tracker view. A customizable display also allows the user to overlay various flight parameters, including latitude, longitude, altitude, pitch, roll, and heading information. Rapid Imaging Software sought assistance from NASA, and the VisualFlight technology came to fruition under a Phase II SBIR contract with Johnson Space Center in 1998. Three years later, on December 13, 2001, Ken Ham successfully flew NASA's X-38 spacecraft from a remote, ground-based cockpit using LandForm VisualFlight as part of his primary situation awareness display in a flight test at Edwards Air Force Base, California.
A Case Study in Ethical Decision Making Regarding Remote Mitigation of Botnets
NASA Astrophysics Data System (ADS)
Dittrich, David; Leder, Felix; Werner, Tillmann
It is becoming more common for researchers to find themselves in a position of being able to take over control of a malicious botnet. If this happens, should they use this knowledge to clean up all the infected hosts? How would this affect not only the owners and operators of the zombie computers, but also other researchers, law enforcement agents serving justice, or even the criminals themselves? What dire circumstances would change the calculus about what is or is not appropriate action to take? We review two case studies of long-lived malicious botnets that present serious challenges to researchers and responders and use them to illuminate many ethical issues regarding aggressive mitigation. We make no judgments about the questions raised, instead laying out the pros and cons of possible choices and allowing workshop attendees to consider how and where they would draw lines. By this, we hope to expose where there is clear community consensus as well as where controversy or uncertainty exists.
NASA Astrophysics Data System (ADS)
Su, Yun-Ting; Hu, Shuowen; Bethel, James S.
2017-05-01
Light detection and ranging (LIDAR) has become a widely used tool in remote sensing for mapping, surveying, modeling, and a host of other applications. The motivation behind this work is the modeling of piping systems in industrial sites, where cylinders are the most common primitive or shape. We focus on cylinder parameter estimation in three-dimensional point clouds, proposing a mathematical formulation based on angular distance to determine the cylinder orientation. We demonstrate the accuracy and robustness of the technique on synthetically generated cylinder point clouds (where the true axis orientation is known) as well as on real LIDAR data of piping systems. The proposed algorithm is compared with a discrete space Hough transform-based approach as well as a continuous space inlier approach, which iteratively discards outlier points to refine the cylinder parameter estimates. Results show that the proposed method is more computationally efficient than the Hough transform approach and is more accurate than both the Hough transform approach and the inlier method.
NASA Technical Reports Server (NTRS)
Ruiz, Ian B.; Burke, Gary R.; Lung, Gerald; Whitaker, William D.; Nowicki, Robert M.
2004-01-01
The Jet Propulsion Laboratory (JPL) has developed a command interface chip-set that primarily consists of two mixed-signal ASICs'; the Command Interface ASIC (CIA) and Analog Interface ASIC (AIA). The Open-systems architecture employed during the design of this chip-set enables its use as both an intelligent gateway between the system's flight computer and the control, actuation, and activation of the spacecraft's loads, valves, and pyrotechnics respectfully as well as the regulator of the spacecraft power bus. Furthermore, the architecture is highly adaptable and employed fault-tolerant design methods enabling a host of other mission uses including reliable remote data collection. The objective of this design is to both provide a needed flight component that meets the stringent environmental requirements of current deep space missions and to add a new element to a growing library that can be used as a standard building block for future missions to the outer planets.
NASA Astrophysics Data System (ADS)
Perini, Federico; Bortolotti, Claudio; Roma, Mauro; Ambrosini, Roberto; Negusini, Monia; Maccaferri, Giuseppe; Stagni, Matteo; Nanni, Mauro; Clivati, Cecilia; Frittelli, Matteo; Mura, Alberto; Levi, Filippo; Zucco, Massimo; Calonico, Davide; Bertarini, Alessandra; Artz, Thomas
2016-12-01
We present the first field test of the implementation of a coherent optical fiber link for remote antenna synchronization realized in Italy between the Italian Metrological Institute (INRIM) and the Medicina radio observatory of the National Institute for Astrophysics (INAF). The Medicina VLBI antenna participated in the EUR137 experiment carried out in September 2015 using, as reference systems, both the local H-maser and a remote H-maser hosted at the INRIM labs in Turin, separated by about 550 km. In order to assess the quality of the remote clock, the observed radio sources were split into two sets, using either the local or the remote H-maser. A system to switch automatically between the two references was integrated into the antenna field system. The observations were correlated in Bonn and preliminary results are encouraging since fringes were detected with both time references along the full 24 hours of the session. The experimental set-up, the results, and the perspectives for future radio astronomical and geodetic experiments are presented.
NASA Technical Reports Server (NTRS)
Seinfeld, J. H. (Principal Investigator)
1982-01-01
The problem of the assimilation of remote sensing data into mathematical models of atmospheric pollutant species was investigated. The data assimilation problem is posed in terms of the matching of spatially integrated species burden measurements to the predicted three-dimensional concentration fields from atmospheric diffusion models. General conditions were derived for the reconstructability of atmospheric concentration distributions from data typical of remote sensing applications, and a computational algorithm (filter) for the processing of remote sensing data was developed.
NASA Technical Reports Server (NTRS)
Seinfeld, J. H. (Principal Investigator)
1982-01-01
The problem of the assimilation of remote sensing data into mathematical models of atmospheric pollutant species was investigated. The problem is posed in terms of the matching of spatially integrated species burden measurements to the predicted three dimensional concentration fields from atmospheric diffusion models. General conditions are derived for the "reconstructability' of atmospheric concentration distributions from data typical of remote sensing applications, and a computational algorithm (filter) for the processing of remote sensing data is developed.
Searches over graphs representing geospatial-temporal remote sensing data
Brost, Randolph; Perkins, David Nikolaus
2018-03-06
Various technologies pertaining to identifying objects of interest in remote sensing images by searching over geospatial-temporal graph representations are described herein. Graphs are constructed by representing objects in remote sensing images as nodes, and connecting nodes with undirected edges representing either distance or adjacency relationships between objects and directed edges representing changes in time. Geospatial-temporal graph searches are made computationally efficient by taking advantage of characteristics of geospatial-temporal data in remote sensing images through the application of various graph search techniques.
Fueling and Stabilizing a Biomolecular Motor-Powered Biosensor for Remote Detection Scenarios
2007-10-01
streptavidin, cyclodextrin host - guest 0 , malachite green - aptamer", DNA - DNA12, and antibody - antigen . Using functionalized microtubules gliding on...3), 646-648 (2005). 11 Hirabayashi, M. et al. Malachite green-conjugated microtubules as mobile bioprobes selective for malachite green aptamers with
Telerobotics control of ExoGeoLab lander instruments
NASA Astrophysics Data System (ADS)
Lillo, A.; Foing, B. H.
2017-09-01
This document is about the improvement of the autonomy and capabilities of the prototype lander ExoGeoLab, designed to host remote controlled instruments for analogue Moon/Mars manned missions. Accent is put on new exploration capabilities for the lander to reduce the need for EVA.
FRR: fair remote retrieval of outsourced private medical records in electronic health networks.
Wang, Huaqun; Wu, Qianhong; Qin, Bo; Domingo-Ferrer, Josep
2014-08-01
Cloud computing is emerging as the next-generation IT architecture. However, cloud computing also raises security and privacy concerns since the users have no physical control over the outsourced data. This paper focuses on fairly retrieving encrypted private medical records outsourced to remote untrusted cloud servers in the case of medical accidents and disputes. Our goal is to enable an independent committee to fairly recover the original private medical records so that medical investigation can be carried out in a convincing way. We achieve this goal with a fair remote retrieval (FRR) model in which either t investigation committee members cooperatively retrieve the original medical data or none of them can get any information on the medical records. We realize the first FRR scheme by exploiting fair multi-member key exchange and homomorphic privately verifiable tags. Based on the standard computational Diffie-Hellman (CDH) assumption, our scheme is provably secure in the random oracle model (ROM). A detailed performance analysis and experimental results show that our scheme is efficient in terms of communication and computation. Copyright © 2014 Elsevier Inc. All rights reserved.
Regulation of inflammation by microbiota interactions with the host.
Blander, J Magarian; Longman, Randy S; Iliev, Iliyan D; Sonnenberg, Gregory F; Artis, David
2017-07-19
The study of the intestinal microbiota has begun to shift from cataloging individual members of the commensal community to understanding their contributions to the physiology of the host organism in health and disease. Here, we review the effects of the microbiome on innate and adaptive immunological players from epithelial cells and antigen-presenting cells to innate lymphoid cells and regulatory T cells. We discuss recent studies that have identified diverse microbiota-derived bioactive molecules and their effects on inflammation within the intestine and distally at sites as anatomically remote as the brain. Finally, we highlight new insights into how the microbiome influences the host response to infection, vaccination and cancer, as well as susceptibility to autoimmune and neurodegenerative disorders.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ruff, T.M.
1992-01-01
A prototype mucking machine designed to operate in narrow vein stopes was developed by Foster-Miller, Inc., Waltham, MA, under contract with the U.S. Bureau of Mines. The machine, called a compact loader/trammer, or minimucker, was designed to replace slusher muckers in narrow-vein underground mines. The minimucker is a six-wheel-drive, skid-steered, load-haul-dump machine that loads muck at the front with a novel slide-bucket system and ejects it out the rear so that the machine does not have to be turned around. To correct deficiencies of the tether remote control system, a computer-based, radio remote control was retrofitted to the minimucker. Initialmore » tests indicated a need to assist the operator in guiding the machine in narrow stopes and an automatic guidance system that used ultrasonic ranging sensors and a wall-following algorithm was installed. Additional tests in a simulated test stope showed that these changes improved the operation of the minimucker. The design and functions of the minimucker and its computer-based, remote control system are reviewed, and an ultrasonic, sensor-based guidance system is described.« less
CBRAIN: a web-based, distributed computing platform for collaborative neuroimaging research
Sherif, Tarek; Rioux, Pierre; Rousseau, Marc-Etienne; Kassis, Nicolas; Beck, Natacha; Adalat, Reza; Das, Samir; Glatard, Tristan; Evans, Alan C.
2014-01-01
The Canadian Brain Imaging Research Platform (CBRAIN) is a web-based collaborative research platform developed in response to the challenges raised by data-heavy, compute-intensive neuroimaging research. CBRAIN offers transparent access to remote data sources, distributed computing sites, and an array of processing and visualization tools within a controlled, secure environment. Its web interface is accessible through any modern browser and uses graphical interface idioms to reduce the technical expertise required to perform large-scale computational analyses. CBRAIN's flexible meta-scheduling has allowed the incorporation of a wide range of heterogeneous computing sites, currently including nine national research High Performance Computing (HPC) centers in Canada, one in Korea, one in Germany, and several local research servers. CBRAIN leverages remote computing cycles and facilitates resource-interoperability in a transparent manner for the end-user. Compared with typical grid solutions available, our architecture was designed to be easily extendable and deployed on existing remote computing sites with no tool modification, administrative intervention, or special software/hardware configuration. As October 2013, CBRAIN serves over 200 users spread across 53 cities in 17 countries. The platform is built as a generic framework that can accept data and analysis tools from any discipline. However, its current focus is primarily on neuroimaging research and studies of neurological diseases such as Autism, Parkinson's and Alzheimer's diseases, Multiple Sclerosis as well as on normal brain structure and development. This technical report presents the CBRAIN Platform, its current deployment and usage and future direction. PMID:24904400
CBRAIN: a web-based, distributed computing platform for collaborative neuroimaging research.
Sherif, Tarek; Rioux, Pierre; Rousseau, Marc-Etienne; Kassis, Nicolas; Beck, Natacha; Adalat, Reza; Das, Samir; Glatard, Tristan; Evans, Alan C
2014-01-01
The Canadian Brain Imaging Research Platform (CBRAIN) is a web-based collaborative research platform developed in response to the challenges raised by data-heavy, compute-intensive neuroimaging research. CBRAIN offers transparent access to remote data sources, distributed computing sites, and an array of processing and visualization tools within a controlled, secure environment. Its web interface is accessible through any modern browser and uses graphical interface idioms to reduce the technical expertise required to perform large-scale computational analyses. CBRAIN's flexible meta-scheduling has allowed the incorporation of a wide range of heterogeneous computing sites, currently including nine national research High Performance Computing (HPC) centers in Canada, one in Korea, one in Germany, and several local research servers. CBRAIN leverages remote computing cycles and facilitates resource-interoperability in a transparent manner for the end-user. Compared with typical grid solutions available, our architecture was designed to be easily extendable and deployed on existing remote computing sites with no tool modification, administrative intervention, or special software/hardware configuration. As October 2013, CBRAIN serves over 200 users spread across 53 cities in 17 countries. The platform is built as a generic framework that can accept data and analysis tools from any discipline. However, its current focus is primarily on neuroimaging research and studies of neurological diseases such as Autism, Parkinson's and Alzheimer's diseases, Multiple Sclerosis as well as on normal brain structure and development. This technical report presents the CBRAIN Platform, its current deployment and usage and future direction.
Information Processing of Remote-Sensing Data.
ERIC Educational Resources Information Center
Berry, P. A. M.; Meadows, A. J.
1987-01-01
Reviews the current status of satellite remote sensing data, including problems with efficient storage and rapid retrieval of the data, and appropriate computer graphics to process images. Areas of research concerned with overcoming these problems are described. (16 references) (CLB)
ARACHNE: A neural-neuroglial network builder with remotely controlled parallel computing
Rusakov, Dmitri A.; Savtchenko, Leonid P.
2017-01-01
Creating and running realistic models of neural networks has hitherto been a task for computing professionals rather than experimental neuroscientists. This is mainly because such networks usually engage substantial computational resources, the handling of which requires specific programing skills. Here we put forward a newly developed simulation environment ARACHNE: it enables an investigator to build and explore cellular networks of arbitrary biophysical and architectural complexity using the logic of NEURON and a simple interface on a local computer or a mobile device. The interface can control, through the internet, an optimized computational kernel installed on a remote computer cluster. ARACHNE can combine neuronal (wired) and astroglial (extracellular volume-transmission driven) network types and adopt realistic cell models from the NEURON library. The program and documentation (current version) are available at GitHub repository https://github.com/LeonidSavtchenko/Arachne under the MIT License (MIT). PMID:28362877
NASA Technical Reports Server (NTRS)
1982-01-01
End user concerns about the content and accessibility of libraries of remote sensing data in general are addressed. Recommendations pertaining to the United States' satellite remote sensing programs urge: (1) the continuation of the NASA/EROS Data Center program to convert pre-1979 scenes to computer readable tapes and create a historical archive of this valuable data; (2) improving the EROS archive by adding geologically interesting scenes, data from other agencies (including previously classified data), and by adopting a policy to retire data from the archive; (3) establishing a computer data base inquiry system that includes remote sensing data from all publically available sources; (4) capability for prepurchase review and evaluation; (5) a flexible price structure; and (6) adoption of standard digital data products format. Information about LANDSAT 4, the status of worldwide LANDSAT receiving stations, future non-U.S. remote sensing satellites, a list of sources for LANDSAT data, and the results of a survey of GEOSAT members' remote sensing data processing systems are also considered.
Advanced electronics for the CTF MEG system.
McCubbin, J; Vrba, J; Spear, P; McKenzie, D; Willis, R; Loewen, R; Robinson, S E; Fife, A A
2004-11-30
Development of the CTF MEG system has been advanced with the introduction of a computer processing cluster between the data acquisition electronics and the host computer. The advent of fast processors, memory, and network interfaces has made this innovation feasible for large data streams at high sampling rates. We have implemented tasks including anti-alias filter, sample rate decimation, higher gradient balancing, crosstalk correction, and optional filters with a cluster consisting of 4 dual Intel Xeon processors operating on up to 275 channel MEG systems at 12 kHz sample rate. The architecture is expandable with additional processors to implement advanced processing tasks which may include e.g., continuous head localization/motion correction, optional display filters, coherence calculations, or real time synthetic channels (via beamformer). We also describe an electronics configuration upgrade to provide operator console access to the peripheral interface features such as analog signal and trigger I/O. This allows remote location of the acoustically noisy electronics cabinet and fitting of the cabinet with doors for improved EMI shielding. Finally, we present the latest performance results available for the CTF 275 channel MEG system including an unshielded SEF (median nerve electrical stimulation) measurement enhanced by application of an adaptive beamformer technique (SAM) which allows recognition of the nominal 20-ms response in the unaveraged signal.
Multiuser Collaboration with Networked Mobile Devices
NASA Technical Reports Server (NTRS)
Tso, Kam S.; Tai, Ann T.; Deng, Yong M.; Becks, Paul G.
2006-01-01
In this paper we describe a multiuser collaboration infrastructure that enables multiple mission scientists to remotely and collaboratively interact with visualization and planning software, using wireless networked personal digital assistants(PDAs) and other mobile devices. During ground operations of planetary rover and lander missions, scientists need to meet daily to review downlinked data and plan science activities. For example, scientists use the Science Activity Planner (SAP) in the Mars Exploration Rover (MER) mission to visualize downlinked data and plan rover activities during the science meetings [1]. Computer displays are projected onto large screens in the meeting room to enable the scientists to view and discuss downlinked images and data displayed by SAP and other software applications. However, only one person can interact with the software applications because input to the computer is limited to a single mouse and keyboard. As a result, the scientists have to verbally express their intentions, such as selecting a target at a particular location on the Mars terrain image, to that person in order to interact with the applications. This constrains communication and limits the returns of science planning. Furthermore, ground operations for Mars missions are fundamentally constrained by the short turnaround time for science and engineering teams to process and analyze data, plan the next uplink, generate command sequences, and transmit the uplink to the vehicle [2]. Therefore, improving ground operations is crucial to the success of Mars missions. The multiuser collaboration infrastructure enables users to control software applications remotely and collaboratively using mobile devices. The infrastructure includes (1) human-computer interaction techniques to provide natural, fast, and accurate inputs, (2) a communications protocol to ensure reliable and efficient coordination of the input devices and host computers, (3) an application-independent middleware that maintains the states, sessions, and interactions of individual users of the software applications, (4) an application programming interface to enable tight integration of applications and the middleware. The infrastructure is able to support any software applications running under the Windows or Unix platforms. The resulting technologies not only are applicable to NASA mission operations, but also useful in other situations such as design reviews, brainstorming sessions, and business meetings, as they can benefit from having the participants concurrently interact with the software applications (e.g., presentation applications and CAD design tools) to illustrate their ideas and provide inputs.
Singh, Tarkeshwar; Perry, Christopher M; Herter, Troy M
2016-01-26
Robotic and virtual-reality systems offer tremendous potential for improving assessment and rehabilitation of neurological disorders affecting the upper extremity. A key feature of these systems is that visual stimuli are often presented within the same workspace as the hands (i.e., peripersonal space). Integrating video-based remote eye tracking with robotic and virtual-reality systems can provide an additional tool for investigating how cognitive processes influence visuomotor learning and rehabilitation of the upper extremity. However, remote eye tracking systems typically compute ocular kinematics by assuming eye movements are made in a plane with constant depth (e.g. frontal plane). When visual stimuli are presented at variable depths (e.g. transverse plane), eye movements have a vergence component that may influence reliable detection of gaze events (fixations, smooth pursuits and saccades). To our knowledge, there are no available methods to classify gaze events in the transverse plane for monocular remote eye tracking systems. Here we present a geometrical method to compute ocular kinematics from a monocular remote eye tracking system when visual stimuli are presented in the transverse plane. We then use the obtained kinematics to compute velocity-based thresholds that allow us to accurately identify onsets and offsets of fixations, saccades and smooth pursuits. Finally, we validate our algorithm by comparing the gaze events computed by the algorithm with those obtained from the eye-tracking software and manual digitization. Within the transverse plane, our algorithm reliably differentiates saccades from fixations (static visual stimuli) and smooth pursuits from saccades and fixations when visual stimuli are dynamic. The proposed methods provide advancements for examining eye movements in robotic and virtual-reality systems. Our methods can also be used with other video-based or tablet-based systems in which eye movements are performed in a peripersonal plane with variable depth.
Near-infrared remotely triggered drug-release strategies for cancer treatment
NASA Astrophysics Data System (ADS)
Goodman, Amanda M.; Neumann, Oara; Nørregaard, Kamilla; Henderson, Luke; Choi, Mi-Ran; Clare, Susan E.; Halas, Naomi J.
2017-11-01
Remotely controlled, localized drug delivery is highly desirable for potentially minimizing the systemic toxicity induced by the administration of typically hydrophobic chemotherapy drugs by conventional means. Nanoparticle-based drug delivery systems provide a highly promising approach for localized drug delivery, and are an emerging field of interest in cancer treatment. Here, we demonstrate near-IR light-triggered release of two drug molecules from both DNA-based and protein-based hosts that have been conjugated to near-infrared-absorbing Au nanoshells (SiO2 core, Au shell), each forming a light-responsive drug delivery complex. We show that, depending upon the drug molecule, the type of host molecule, and the laser illumination method (continuous wave or pulsed laser), in vitro light-triggered release can be achieved with both types of nanoparticle-based complexes. Two breast cancer drugs, docetaxel and HER2-targeted lapatinib, were delivered to MDA-MB-231 and SKBR3 (overexpressing HER2) breast cancer cells and compared with release in noncancerous RAW 264.7 macrophage cells. Continuous wave laser-induced release of docetaxel from a nanoshell-based DNA host complex showed increased cell death, which also coincided with nonspecific cell death from photothermal heating. Using a femtosecond pulsed laser, lapatinib release from a nanoshell-based human serum albumin protein host complex resulted in increased cancerous cell death while noncancerous control cells were unaffected. Both methods provide spatially and temporally localized drug-release strategies that can facilitate high local concentrations of chemotherapy drugs deliverable at a specific treatment site over a specific time window, with the potential for greatly minimized side effects.
Active optical control system design of the SONG-China Telescope
NASA Astrophysics Data System (ADS)
Ye, Yu; Kou, Songfeng; Niu, Dongsheng; Li, Cheng; Wang, Guomin
2012-09-01
The standard SONG node structure of control system is presented. The active optical control system of the project is a distributed system, and a host computer and a slave intelligent controller are included. The host control computer collects the information from wave front sensor and sends commands to the slave computer to realize a closed loop model. For intelligent controller, a programmable logic controller (PLC) system is used. This system combines with industrial personal computer (IPC) and PLC to make up a control system with powerful and reliable.
Maia, João P; Harris, D James; Carranza, Salvador; Goméz-Díaz, Elena
2016-11-01
Understanding the processes that shape parasite diversification, their distribution and abundance provides valuable information on the dynamics and evolution of disease. In this study, we assessed the diversity, distribution, host-specificity and infection patterns of apicomplexan parasites in amphibians and reptiles from Oman, Arabia. Using a quantitative PCR approach we detected three apicomplexan parasites (haemogregarines, lankesterellids and sarcocystids). A total of 13 haemogregarine haplotypes were identified, which fell into four main clades in a phylogenetic framework. Phylogenetic analysis of six new lankesterellid haplotypes revealed that these parasites were distinct from, but phylogenetically related to, known Lankesterella species and might represent new taxa. The percentage of infected hosts (prevalence) and the number of haemogregarines in the blood (parasitaemia) varied significantly between gecko species. We also found significant differences in parasitaemia between haemogregarine parasite lineages (defined by phylogenetic clustering of haplotypes), suggesting differences in host-parasite compatibility between these lineages. For Pristurus rupestris, we found significant differences in haemogregarine prevalence between geographical areas. Our results suggest that host ecology and host relatedness may influence haemogregarine distributions and, more generally, highlight the importance of screening wild hosts from remote regions to provide new insights into parasite diversity.
Cloud Based Educational Systems and Its Challenges and Opportunities and Issues
ERIC Educational Resources Information Center
Paul, Prantosh Kr.; Lata Dangwal, Kiran
2014-01-01
Cloud Computing (CC) is actually is a set of hardware, software, networks, storage, services an interface combines to deliver aspects of computing as a service. Cloud Computing (CC) actually uses the central remote servers to maintain data and applications. Practically Cloud Computing (CC) is extension of Grid computing with independency and…
Data Storage and Transfer | High-Performance Computing | NREL
High-Performance Computing (HPC) systems. Photo of computer server wiring and lights, blurred to show data. WinSCP for Windows File Transfers Use to transfer files from a local computer to a remote computer. Robinhood for File Management Use this tool to manage your data files on Peregrine. Best
Grönlund, Rasmus; Lundqvist, Mats; Svanberg, Sune
2006-08-01
A mobile lidar system was used in remote imaging laser-induced breakdown spectroscopy (LIBS) and laser-induced fluorescence (LIF) experiments. Also, computer-controlled remote ablation of a chosen area was demonstrated, relevant to cleaning of cultural heritage items. Nanosecond frequency-tripled Nd:YAG laser pulses at 355 nm were employed in experiments with a stand-off distance of 60 meters using pulse energies of up to 170 mJ. By coaxial transmission and common folding of the transmission and reception optical paths using a large computer-controlled mirror, full elemental imaging capability was achieved on composite targets. Different spectral identification algorithms were compared in producing thematic data based on plasma or fluorescence light.
A Low Cost Remote Sensing System Using PC and Stereo Equipment
NASA Technical Reports Server (NTRS)
Campbell, Joel F.; Flood, Michael A.; Prasad, Narasimha S.; Hodson, Wade D.
2011-01-01
A system using a personal computer, speaker, and a microphone is used to detect objects, and make crude measurements using a carrier modulated by a pseudorandom noise (PN) code. This system can be constructed using a personal computer and audio equipment commonly found in the laboratory or at home, or more sophisticated equipment that can be purchased at reasonable cost. We demonstrate its value as an instructional tool for teaching concepts of remote sensing and digital signal processing.
Model studies of laser absorption computed tomography for remote air pollution measurement
NASA Technical Reports Server (NTRS)
Wolfe, D. C., Jr.; Byer, R. L.
1982-01-01
Model studies of the potential of laser absorption-computed tomography are presented which demonstrate the possibility of sensitive remote atmospheric pollutant measurements, over kilometer-sized areas, with two-dimensional resolution, at modest laser source powers. An analysis of this tomographic reconstruction process as a function of measurement SNR, laser power, range, and system geometry, shows that the system is able to yield two-dimensional maps of pollutant concentrations at ranges and resolutions superior to those attainable with existing, direct-detection laser radars.
Demonstration of versatile whispering-gallery micro-lasers for remote refractive index sensing.
Wan, Lei; Chandrahalim, Hengky; Zhou, Jian; Li, Zhaohui; Chen, Cong; Cho, Sangha; Zhang, Hui; Mei, Ting; Tian, Huiping; Oki, Yuji; Nishimura, Naoya; Fan, Xudong; Guo, L Jay
2018-03-05
We developed chip-scale remote refractive index sensors based on Rhodamine 6G (R6G)-doped polymer micro-ring lasers. The chemical, temperature, and mechanical sturdiness of the fused-silica host guaranteed a flexible deployment of dye-doped polymers for refractive index sensing. The introduction of the dye as gain medium demonstrated the feasibility of remote sensing based on the free-space optics measurement setup. Compared to the R6G-doped TZ-001, the lasing behavior of R6G-doped SU-8 polymer micro-ring laser under an aqueous environment had a narrower spectrum linewidth, producing the minimum detectable refractive index change of 4 × 10 -4 RIU. The maximum bulk refractive index sensitivity (BRIS) of 75 nm/RIU was obtained for SU-8 laser-based refractive index sensors. The economical, rapid, and simple realization of polymeric micro-scale whispering-gallery-mode (WGM) laser-based refractive index sensors will further expand pathways of static and dynamic remote environmental, chemical, biological, and bio-chemical sensing.
Videodisc-Computer Interfaces.
ERIC Educational Resources Information Center
Zollman, Dean
1984-01-01
Lists microcomputer-videodisc interfaces currently available from 26 sources, including home use systems connected through remote control jack and industrial/educational systems utilizing computer ports and new laser reflective and stylus technology. Information provided includes computer and videodisc type, language, authoring system, educational…
Guidelines for Outsourcing Remote Access.
ERIC Educational Resources Information Center
Hassler, Ardoth; Neuman, Michael
1996-01-01
Discusses the advantages and disadvantages of outsourcing remote access to campus computer networks and the Internet, focusing on improved service, cost-sharing, partnerships with vendors, supported protocols, bandwidth, scope of access, implementation, support, network security, and pricing. Includes a checklist for a request for proposals on…
Remote file inquiry (RFI) system
NASA Technical Reports Server (NTRS)
1975-01-01
System interrogates and maintains user-definable data files from remote terminals, using English-like, free-form query language easily learned by persons not proficient in computer programming. System operates in asynchronous mode, allowing any number of inquiries within limitation of available core to be active concurrently.
NASA Technical Reports Server (NTRS)
1984-01-01
Topics discussed at the symposium include hardware, geographic information system (GIS) implementation, processing remotely sensed data, spatial data structures, and NASA programs in remote sensing information systems. Attention is also given GIS applications, advanced techniques, artificial intelligence, graphics, spatial navigation, and classification. Papers are included on the design of computer software for geographic image processing, concepts for a global resource information system, algorithm development for spatial operators, and an application of expert systems technology to remotely sensed image analysis.
Energy and remote sensing. [satellite exploration, monitoring, siting
NASA Technical Reports Server (NTRS)
Summers, R. A.; Smith, W. L.; Short, N. M.
1977-01-01
Exploration for uranium, thorium, oil, gas and geothermal activity through remote sensing techniques is considered; satellite monitoring of coal-derived CO2 in the atmosphere, and the remote assessment of strip mining and land restoration are also mentioned. Reference is made to color ratio composites based on Landsat data, which may aid in the detection of uranium deposits, and to computer-enhanced black and white airborne scanning imagery, which may locate geothermal anomalies. Other applications of remote sensing to energy resources management, including mapping of transportation networks and power plant siting, are discussed.
Practical lessons in remote connectivity.
Kouroubali, A.; Starren, J.; Barrows, R. C.; Clayton, P. D.
1997-01-01
Community Health Information Networks (CHINs) require the ability to provide computer network connections to many remote sites. During the implementation of the Washington Heights and Inwood Community Health Management Information System (WHICHIS) at the Columbia-Presbyterian Medical Center (CPMC), a number of remote connectivity issues have been encountered. Both technical and non-technical issues were significant during the installation. We developed a work-flow model for this process which may be helpful to any health care institution attempting to provide seamless remote connectivity. This model is presented and implementation lessons are discussed. PMID:9357643
Remote sensing impact on corridor selection and placement
NASA Technical Reports Server (NTRS)
Thomson, F. J.; Sellman, A. N.
1975-01-01
Computer-aided corridor selection techniques, utilizing digitized data bases of socio-economic, census, and cadastral data, and developed for highway corridor routing are considered. Land resource data generated from various remote sensing data sources were successfully merged with the ancillary data files of a corridor selection model and prototype highway corridors were designed using the combined data set. Remote sensing derived information considered useful for highway corridor location, special considerations in geometric correction of remote sensing data to facilitate merging it with ancillary data files, and special interface requirements are briefly discussed.
Hans-Erik Andersen; Strunk Jacob; Hailemariam Temesgen; Donald Atwood; Ken Winterberger
2012-01-01
The emergence of a new generation of remote sensing and geopositioning technologies, as well as increased capabilities in image processing, computing, and inferential techniques, have enabled the development and implementation of increasingly efficient and cost-effective multilevel sampling designs for forest inventory. In this paper, we (i) describe the conceptual...
Extending IPsec for Efficient Remote Attestation
NASA Astrophysics Data System (ADS)
Sadeghi, Ahmad-Reza; Schulz, Steffen
When establishing a VPN to connect different sites of a network, the integrity of the involved VPN endpoints is often a major security concern. Based on the Trusted Platform Module (TPM), available in many computing platforms today, remote attestation mechanisms can be used to evaluate the internal state of remote endpoints automatically. However, existing protocols and extensions are either unsuited for use with IPsec or impose considerable additional implementation complexity and protocol overhead.
3D Medical Collaboration Technology to Enhance Emergency Healthcare
Welch, Greg; Sonnenwald, Diane H; Fuchs, Henry; Cairns, Bruce; Mayer-Patel, Ketan; Söderholm, Hanna M.; Yang, Ruigang; State, Andrei; Towles, Herman; Ilie, Adrian; Ampalam, Manoj; Krishnan, Srinivas; Noel, Vincent; Noland, Michael; Manning, James E.
2009-01-01
Two-dimensional (2D) videoconferencing has been explored widely in the past 15–20 years to support collaboration in healthcare. Two issues that arise in most evaluations of 2D videoconferencing in telemedicine are the difficulty obtaining optimal camera views and poor depth perception. To address these problems, we are exploring the use of a small array of cameras to reconstruct dynamic three-dimensional (3D) views of a remote environment and of events taking place within. The 3D views could be sent across wired or wireless networks to remote healthcare professionals equipped with fixed displays or with mobile devices such as personal digital assistants (PDAs). The remote professionals’ viewpoints could be specified manually or automatically (continuously) via user head or PDA tracking, giving the remote viewers head-slaved or hand-slaved virtual cameras for monoscopic or stereoscopic viewing of the dynamic reconstructions. We call this idea remote 3D medical collaboration. In this article we motivate and explain the vision for 3D medical collaboration technology; we describe the relevant computer vision, computer graphics, display, and networking research; we present a proof-of-concept prototype system; and we present evaluation results supporting the general hypothesis that 3D remote medical collaboration technology could offer benefits over conventional 2D videoconferencing in emergency healthcare. PMID:19521951
3D medical collaboration technology to enhance emergency healthcare.
Welch, Gregory F; Sonnenwald, Diane H; Fuchs, Henry; Cairns, Bruce; Mayer-Patel, Ketan; Söderholm, Hanna M; Yang, Ruigang; State, Andrei; Towles, Herman; Ilie, Adrian; Ampalam, Manoj K; Krishnan, Srinivas; Noel, Vincent; Noland, Michael; Manning, James E
2009-04-19
Two-dimensional (2D) videoconferencing has been explored widely in the past 15-20 years to support collaboration in healthcare. Two issues that arise in most evaluations of 2D videoconferencing in telemedicine are the difficulty obtaining optimal camera views and poor depth perception. To address these problems, we are exploring the use of a small array of cameras to reconstruct dynamic three-dimensional (3D) views of a remote environment and of events taking place within. The 3D views could be sent across wired or wireless networks to remote healthcare professionals equipped with fixed displays or with mobile devices such as personal digital assistants (PDAs). The remote professionals' viewpoints could be specified manually or automatically (continuously) via user head or PDA tracking, giving the remote viewers head-slaved or hand-slaved virtual cameras for monoscopic or stereoscopic viewing of the dynamic reconstructions. We call this idea remote 3D medical collaboration. In this article we motivate and explain the vision for 3D medical collaboration technology; we describe the relevant computer vision, computer graphics, display, and networking research; we present a proof-of-concept prototype system; and we present evaluation results supporting the general hypothesis that 3D remote medical collaboration technology could offer benefits over conventional 2D videoconferencing in emergency healthcare.
DNET: A communications facility for distributed heterogeneous computing
NASA Technical Reports Server (NTRS)
Tole, John; Nagappan, S.; Clayton, J.; Ruotolo, P.; Williamson, C.; Solow, H.
1989-01-01
This document describes DNET, a heterogeneous data communications networking facility. DNET allows programs operating on hosts on dissimilar networks to communicate with one another without concern for computer hardware, network protocol, or operating system differences. The overall DNET network is defined as the collection of host machines/networks on which the DNET software is operating. Each underlying network is considered a DNET 'domain'. Data communications service is provided between any two processes on any two hosts on any of the networks (domains) that may be reached via DNET. DNET provides protocol transparent, reliable, streaming data transmission between hosts (restricted, initially to DECnet and TCP/IP networks). DNET also provides variable length datagram service with optional return receipts.
NASA Technical Reports Server (NTRS)
Montag, Bruce C.; Bishop, Alfred M.; Redfield, Joe B.
1989-01-01
The findings of a preliminary investigation by Southwest Research Institute (SwRI) in simulation host computer concepts is presented. It is designed to aid NASA in evaluating simulation technologies for use in spaceflight training. The focus of the investigation is on the next generation of space simulation systems that will be utilized in training personnel for Space Station Freedom operations. SwRI concludes that NASA should pursue a distributed simulation host computer system architecture for the Space Station Training Facility (SSTF) rather than a centralized mainframe based arrangement. A distributed system offers many advantages and is seen by SwRI as the only architecture that will allow NASA to achieve established functional goals and operational objectives over the life of the Space Station Freedom program. Several distributed, parallel computing systems are available today that offer real-time capabilities for time critical, man-in-the-loop simulation. These systems are flexible in terms of connectivity and configurability, and are easily scaled to meet increasing demands for more computing power.
TCS and peripheral robotization and upgrade on the ESO 1-meter telescope at La Silla Observatory
NASA Astrophysics Data System (ADS)
Ropert, S.; Suc, V.; Jordán, A.; Tala, M.; Liedtke, P.; Royo, S.
2016-07-01
In this work we describe the robotization and upgrade of the ESO 1m telescope located at La Silla Observatory. The ESO 1m telescope was the first telescope installed in La Silla, in 1966. It now hosts as a main instrument the FIber Dual EchellE Optical Spectrograph (FIDEOS), a high resolution spectrograph designed for precise Radial Velocity (RV) measurements on bright stars. In order to meet this project's requirements, the Telescope Control System (TCS) and some of its mechanical peripherals needed to be upgraded. The TCS was also upgraded into a modern and robust software running on a group of single board computers interacting together as a network with the CoolObs TCS developed by ObsTech. One of the particularities of the CoolObs TCS is that it allows to fuse the input signals of 2 encoders per axis in order to achieve high precision and resolution of the tracking with moderate cost encoders. One encoder is installed on axis at the telescope and the other on axis at the motor. The TCS was also integrated with the FIDEOS instrument system so that all the system can be controlled through the same remote user interface. Our modern TCS unit allows the user to run observations remotely through a secured internet web interface, minimizing the need of an on-site observer and opening a new age in robotic astronomy for the ESO 1m telescope.
Remote sensing entropy to assess the sustainability of rainfall in tropical catchment
NASA Astrophysics Data System (ADS)
Mahmud, M. R.; Reba, M. N. M.; Wei, J. S.; Razak, N. H. Abdul
2018-02-01
This study demonstrated the utility of entropy computation using the satellite precipitation remote sensing data to assess the sustainability of rainfall in tropical catchments. There were two major issues need to be anticipated in monitoring the tropical catchments; first is the frequent monitoring of the rainfall and second is the appropriate indicator that sensitive to rainfall pattern changes or disorder. For the first issue, the use of satellite remote sensing precipitation data is suggested. Meanwhile for the second issue, the utilization of entropy concept in interpreting the disorder of temporal rainfall can be used to assess the sustain ability had been successfully adopted in some studies. Therefore, we hypothesized that the use of satellite precipitation as main data to compute entropy can be a novel tool in anticipating the above-mentioned conflict earlier. The remote sensing entropy results and in-situ river level showed good agreement indicating its reliability. 72% of the catchment has moderate to good rainfall supply during normal or non-drought condition. However, our result showed that the catchments were highly sensitive to drought especially in the west coast and southern part of the Peninsular Malaysia. High resiliency was identified in the east coast. We summarized that the proposed entropy-quantity scheme was a useful tool for cost-effective, quick, and operational sustainability assessment This study demonstrated the utility of entropy computation using the satellite precipitation remote sensing data to assess the sustainability of rainfall in tropical catchments.
Controlling multiple security robots in a warehouse environment
NASA Technical Reports Server (NTRS)
Everett, H. R.; Gilbreath, G. A.; Heath-Pastore, T. A.; Laird, R. T.
1994-01-01
The Naval Command Control and Ocean Surveillance Center (NCCOSC) has developed an architecture to provide coordinated control of multiple autonomous vehicles from a single host console. The multiple robot host architecture (MRHA) is a distributed multiprocessing system that can be expanded to accommodate as many as 32 robots. The initial application will employ eight Cybermotion K2A Navmaster robots configured as remote security platforms in support of the Mobile Detection Assessment and Response System (MDARS) Program. This paper discusses developmental testing of the MRHA in an operational warehouse environment, with two actual and four simulated robotic platforms.
Host-Nation Operations: Soldier Training on Governance (HOST-G) Training Support Package
2011-07-01
restricted this webpage from running scripts or ActiveX controls that could access your computer. Click here for options…” • If this occurs, select that...scripts and ActiveX controls can be useful, but active content might also harm your computer. Are you sure you want to let this file run active
Imagine, Invent, Program, Share: A Library-Hosted Computer Club Promotes 21st Century Skills
ERIC Educational Resources Information Center
Myers, Brian
2009-01-01
During at least one afternoon each month, Wilmette (Illinois) Public Library (WPL) hosts a local group of computer programmers, designers, and artists, who meet to discuss digital projects and resources, technical challenges, and successful design or programming strategies. WPL's Game Design Club, now in its third year, owes its existence to a…
Finite-data-size study on practical universal blind quantum computation
NASA Astrophysics Data System (ADS)
Zhao, Qiang; Li, Qiong
2018-07-01
The universal blind quantum computation with weak coherent pulses protocol is a practical scheme to allow a client to delegate a computation to a remote server while the computation hidden. However, in the practical protocol, a finite data size will influence the preparation efficiency in the remote blind qubit state preparation (RBSP). In this paper, a modified RBSP protocol with two decoy states is studied in the finite data size. The issue of its statistical fluctuations is analyzed thoroughly. The theoretical analysis and simulation results show that two-decoy-state case with statistical fluctuation is closer to the asymptotic case than the one-decoy-state case with statistical fluctuation. Particularly, the two-decoy-state protocol can achieve a longer communication distance than the one-decoy-state case in this statistical fluctuation situation.
NASA Astrophysics Data System (ADS)
Druken, K. A.; Trenham, C. E.; Wang, J.; Bastrakova, I.; Evans, B. J. K.; Wyborn, L. A.; Ip, A. I.; Poudjom Djomani, Y.
2016-12-01
The National Computational Infrastructure (NCI) hosts one of Australia's largest repositories (10+ PBytes) of research data, colocated with a petascale High Performance Computer and a highly integrated research cloud. Key to maximizing benefit of NCI's collections and computational capabilities is ensuring seamless interoperable access to these datasets. This presents considerable data management challenges across the diverse range of geoscience data; spanning disciplines where netCDF-CF is commonly utilized (e.g., climate, weather, remote-sensing), through to the geophysics and seismology fields that employ more traditional domain- and study-specific data formats. These data are stored in a variety of gridded, irregularly spaced (i.e., trajectories, point clouds, profiles), and raster image structures. They often have diverse coordinate projections and resolutions, thus complicating the task of comparison and inter-discipline analysis. Nevertheless, much can be learned from the netCDF-CF model that has long served the climate community, providing a common data structure for the atmospheric, ocean and cryospheric sciences. We are extending the application of the existing Climate and Forecast (CF) metadata conventions to NCI's broader geoscience data collections. We present simple implementations that can significantly improve interoperability of the research collections, particularly in the case of line survey data. NCI has developed a compliance checker to assist with the data quality across all hosted netCDF-CF collections. The tool is an extension to one of the main existing CF Convention checkers, that we have modified to incorporate the Attribute Convention for Data Discovery (ACDD) and ISO19115 standards, and to perform parallelised checks over collections of files, ensuring compliance and consistency across the NCI data collections as a whole. It is complemented by a checker that also verifies functionality against a range of scientific analysis, programming, and data visualisation tools. By design, these tests are not necessarily domain-specific, and demonstrate that verified data is accessible to end-users, thus allowing for seamless interoperability with other datasets across a wide range of fields.
Gagne, Roderick B; Heins, David C; McIntyre, Peter B; Gilliam, James F; Blum, Michael J
2016-10-01
The presence of introduced hosts can increase or decrease infections of co-introduced parasites in native species of conservation concern. In this study, we compared parasite abundance, intensity, and prevalence between native Awaous stamineus and introduced poeciliid fishes by a co-introduced nematode parasite (Camallanus cotti) in 42 watersheds across the Hawaiian Islands. We found that parasite abundance, intensity and prevalence were greater in native than introduced hosts. Parasite abundance, intensity and prevalence within A. stamineus varied between years, which largely reflected a transient spike in infection in three remote watersheds on Molokai. At each site we measured host factors (length, density of native host, density of introduced host) and environmental factors (per cent agricultural and urban land use, water chemistry, watershed area and precipitation) hypothesized to influence C. cotti abundance, intensity and prevalence. Factors associated with parasitism differed between native and introduced hosts. Notably, parasitism of native hosts was higher in streams with lower water quality, whereas parasitism of introduced hosts was lower in streams with lower water quality. We also found that parasite burdens were lower in both native and introduced hosts when coincident. Evidence of a mutual dilution effect indicates that introduced hosts can ameliorate parasitism of native fishes by co-introduced parasites, which raises questions about the value of remediation actions, such as the removal of introduced hosts, in stemming the rise of infectious disease in species of conservation concern.
A comparison of operational remote sensing-based models for estimating crop evapotranspiration
USDA-ARS?s Scientific Manuscript database
The integration of remotely sensed data into models of actual evapotranspiration has allowed for the estimation of water consumption across agricultural regions. Two modeling approaches have been successfully applied. The first approach computes a surface energy balance using the radiometric surface...
A computer simulation experiment of supervisory control of remote manipulation. M.S. Thesis
NASA Technical Reports Server (NTRS)
Mccandlish, S. G.
1966-01-01
A computer simulation of a remote manipulation task and a rate-controlled manipulator is described. Some low-level automatic decision making ability which could be used at the operator's discretion to augment his direct continuous control was built into the manipulator. Experiments were made on the effect of transmission delay, dynamic lag, and intermittent vision on human manipulative ability. Delay does not make remote manipulation impossible. Intermittent visual feedback, and the absence of rate information in the display presented to the operator do not seem to impair the operator's performance. A small-capacity visual feedback channel may be sufficient for remote manipulation tasks, or one channel might be time-shared between several operators. In other experiments the operator called in sequence various on-site automatic control programs of the machine, and thereby acted as a supervisor. The supervisory mode of operation has some advantages when the task to be performed is difficult for a human controlling directly.
Handels, H; Busch, C; Encarnação, J; Hahn, C; Kühn, V; Miehe, J; Pöppl, S I; Rinast, E; Rossmanith, C; Seibert, F; Will, A
1997-03-01
The software system KAMEDIN (Kooperatives Arbeiten und MEdizinische Diagnostik auf Innovativen Netzen) is a multimedia telemedicine system for exchange, cooperative diagnostics, and remote analysis of digital medical image data. It provides components for visualisation, processing, and synchronised audio-visual discussion of medical images. Techniques of computer supported cooperative work (CSCW) synchronise user interactions during a teleconference. Visibility of both local and remote cursor on the conference workstations facilitates telepointing and reinforces the conference partner's telepresence. Audio communication during teleconferences is supported by an integrated audio component. Furthermore, brain tissue segmentation with artificial neural networks can be performed on an external supercomputer as a remote image analysis procedure. KAMEDIN is designed as a low cost CSCW tool for ISDN based telecommunication. However it can be used on any TCP/IP supporting network. In a field test, KAMEDIN was installed in 15 clinics and medical departments to validate the systems' usability. The telemedicine system KAMEDIN has been developed, tested, and evaluated within a research project sponsored by German Telekom.
Heart rate measurement based on face video sequence
NASA Astrophysics Data System (ADS)
Xu, Fang; Zhou, Qin-Wu; Wu, Peng; Chen, Xing; Yang, Xiaofeng; Yan, Hong-jian
2015-03-01
This paper proposes a new non-contact heart rate measurement method based on photoplethysmography (PPG) theory. With this method we can measure heart rate remotely with a camera and ambient light. We collected video sequences of subjects, and detected remote PPG signals through video sequences. Remote PPG signals were analyzed with two methods, Blind Source Separation Technology (BSST) and Cross Spectral Power Technology (CSPT). BSST is a commonly used method, and CSPT is used for the first time in the study of remote PPG signals in this paper. Both of the methods can acquire heart rate, but compared with BSST, CSPT has clearer physical meaning, and the computational complexity of CSPT is lower than that of BSST. Our work shows that heart rates detected by CSPT method have good consistency with the heart rates measured by a finger clip oximeter. With good accuracy and low computational complexity, the CSPT method has a good prospect for the application in the field of home medical devices and mobile health devices.
The HEPiX Virtualisation Working Group: Towards a Grid of Clouds
NASA Astrophysics Data System (ADS)
Cass, Tony
2012-12-01
The use of virtual machine images, as for example with Cloud services such as Amazon's Elastic Compute Cloud, is attractive for users as they have a guaranteed execution environment, something that cannot today be provided across sites participating in computing grids such as the Worldwide LHC Computing Grid. However, Grid sites often operate within computer security frameworks which preclude the use of remotely generated images. The HEPiX Virtualisation Working Group was setup with the objective to enable use of remotely generated virtual machine images at Grid sites and, to this end, has introduced the idea of trusted virtual machine images which are guaranteed to be secure and configurable by sites such that security policy commitments can be met. This paper describes the requirements and details of these trusted virtual machine images and presents a model for their use to facilitate the integration of Grid- and Cloud-based computing environments for High Energy Physics.
Research Status and Development Trend of Remote Sensing in China Using Bibliometric Analysis
NASA Astrophysics Data System (ADS)
Zeng, Y.; Zhang, J.; Niu, R.
2015-06-01
Remote sensing was introduced into China in 1970s and then began to flourish. At present, China has developed into a big remote sensing country, and remote sensing is increasingly playing an important role in various fields of national economic construction and social development. Based on China Academic Journals Full-text Database and China Citation Database published by China National Knowledge Infrastructure, this paper analyzed academic characteristics of 963 highly cited papers published by 16 professional and academic journals in the field of surveying and mapping from January 2010 to December 2014 in China, which include hot topics, literature authors, research institutions, and fundations. At the same time, it studied a total of 51,149 keywords published by these 16 journals during the same period. Firstly by keyword selection, keyword normalization, keyword consistency and keyword incorporation, and then by analysis of high frequency keywords, the progress and prospect of China's remote sensing technology in data acquisition, data processing and applications during the past five years were further explored and revealed. It can be seen that: highly cited paper analysis and word frequency analysis is complementary on subject progress analysis; in data acquisition phase, research focus is new civilian remote sensing satellite systems and UAV remote sensing system; research focus of data processing and analysis is multi-source information extraction and classification, laser point cloud data processing, objectoriented high resolution image analysis, SAR data and hyper-spectral image processing, etc.; development trend of remote sensing data processing is quantitative, intelligent, automated, and real-time, and the breadth and depth of remote sensing application is gradually increased; parallel computing, cloud computing and geographic conditions monitoring and census are the new research focuses to be paid attention to.
WinSCP for Windows File Transfers | High-Performance Computing | NREL
WinSCP for Windows File Transfers WinSCP for Windows File Transfers WinSCP for can used to securely transfer files between your local computer running Microsoft Windows and a remote computer running Linux
Virtual network computing: cross-platform remote display and collaboration software.
Konerding, D E
1999-04-01
VNC (Virtual Network Computing) is a computer program written to address the problem of cross-platform remote desktop/application display. VNC uses a client/server model in which an image of the desktop of the server is transmitted to the client and displayed. The client collects mouse and keyboard input from the user and transmits them back to the server. The VNC client and server can run on Windows 95/98/NT, MacOS, and Unix (including Linux) operating systems. VNC is multi-user on Unix machines (any number of servers can be run are unrelated to the primary display of the computer), while it is effectively single-user on Macintosh and Windows machines (only one server can be run, displaying the contents of the primary display of the server). The VNC servers can be configured to allow more than one client to connect at one time, effectively allowing collaboration through the shared desktop. I describe the function of VNC, provide details of installation, describe how it achieves its goal, and evaluate the use of VNC for molecular modelling. VNC is an extremely useful tool for collaboration, instruction, software development, and debugging of graphical programs with remote users.
High-Speed Recording of Test Data on Hard Disks
NASA Technical Reports Server (NTRS)
Lagarde, Paul M., Jr.; Newnan, Bruce
2003-01-01
Disk Recording System (DRS) is a systems-integration computer program for a direct-to-disk (DTD) high-speed data acquisition system (HDAS) that records rocket-engine test data. The HDAS consists partly of equipment originally designed for recording the data on tapes. The tape recorders were replaced with hard-disk drives, necessitating the development of DRS to provide an operating environment that ties two computers, a set of five DTD recorders, and signal-processing circuits from the original tape-recording version of the HDAS into one working system. DRS includes three subsystems: (1) one that generates a graphical user interface (GUI), on one of the computers, that serves as a main control panel; (2) one that generates a GUI, on the other computer, that serves as a remote control panel; and (3) a data-processing subsystem that performs tasks on the DTD recorders according to instructions sent from the main control panel. The software affords capabilities for dynamic configuration to record single or multiple channels from a remote source, remote starting and stopping of the recorders, indexing to prevent overwriting of data, and production of filtered frequency data from an original time-series data file.
LabVIEW Serial Driver Software for an Electronic Load
NASA Technical Reports Server (NTRS)
Scullin, Vincent; Garcia, Christopher
2003-01-01
A LabVIEW-language computer program enables monitoring and control of a Transistor Devices, Inc., Dynaload WCL232 (or equivalent) electronic load via an RS-232 serial communication link between the electronic load and a remote personal computer. (The electronic load can operate at constant voltage, current, power consumption, or resistance.) The program generates a graphical user interface (GUI) at the computer that looks and acts like the front panel of the electronic load. Once the electronic load has been placed in remote-control mode, this program first queries the electronic load for the present values of all its operational and limit settings, and then drops into a cycle in which it reports the instantaneous voltage, current, and power values in displays that resemble those on the electronic load while monitoring the GUI images of pushbuttons for control actions by the user. By means of the pushbutton images and associated prompts, the user can perform such operations as changing limit values, the operating mode, or the set point. The benefit of this software is that it relieves the user of the need to learn one method for operating the electronic load locally and another method for operating it remotely via a personal computer.
Towards a Low-Cost Remote Memory Attestation for the Smart Grid
Yang, Xinyu; He, Xiaofei; Yu, Wei; Lin, Jie; Li, Rui; Yang, Qingyu; Song, Houbing
2015-01-01
In the smart grid, measurement devices may be compromised by adversaries, and their operations could be disrupted by attacks. A number of schemes to efficiently and accurately detect these compromised devices remotely have been proposed. Nonetheless, most of the existing schemes detecting compromised devices depend on the incremental response time in the attestation process, which are sensitive to data transmission delay and lead to high computation and network overhead. To address the issue, in this paper, we propose a low-cost remote memory attestation scheme (LRMA), which can efficiently and accurately detect compromised smart meters considering real-time network delay and achieve low computation and network overhead. In LRMA, the impact of real-time network delay on detecting compromised nodes can be eliminated via investigating the time differences reported from relay nodes. Furthermore, the attestation frequency in LRMA is dynamically adjusted with the compromised probability of each node, and then, the total number of attestations could be reduced while low computation and network overhead can be achieved. Through a combination of extensive theoretical analysis and evaluations, our data demonstrate that our proposed scheme can achieve better detection capacity and lower computation and network overhead in comparison to existing schemes. PMID:26307998
Towards a Low-Cost Remote Memory Attestation for the Smart Grid.
Yang, Xinyu; He, Xiaofei; Yu, Wei; Lin, Jie; Li, Rui; Yang, Qingyu; Song, Houbing
2015-08-21
In the smart grid, measurement devices may be compromised by adversaries, and their operations could be disrupted by attacks. A number of schemes to efficiently and accurately detect these compromised devices remotely have been proposed. Nonetheless, most of the existing schemes detecting compromised devices depend on the incremental response time in the attestation process, which are sensitive to data transmission delay and lead to high computation and network overhead. To address the issue, in this paper, we propose a low-cost remote memory attestation scheme (LRMA), which can efficiently and accurately detect compromised smart meters considering real-time network delay and achieve low computation and network overhead. In LRMA, the impact of real-time network delay on detecting compromised nodes can be eliminated via investigating the time differences reported from relay nodes. Furthermore, the attestation frequency in LRMA is dynamically adjusted with the compromised probability of each node, and then, the total number of attestations could be reduced while low computation and network overhead can be achieved. Through a combination of extensive theoretical analysis and evaluations, our data demonstrate that our proposed scheme can achieve better detection capacity and lower computation and network overhead in comparison to existing schemes.
Cern, Ahuva; Barenholz, Yechezkel; Tropsha, Alexander; Goldblum, Amiram
2014-01-10
Previously we have developed and statistically validated Quantitative Structure Property Relationship (QSPR) models that correlate drugs' structural, physical and chemical properties as well as experimental conditions with the relative efficiency of remote loading of drugs into liposomes (Cern et al., J. Control. Release 160 (2012) 147-157). Herein, these models have been used to virtually screen a large drug database to identify novel candidate molecules for liposomal drug delivery. Computational hits were considered for experimental validation based on their predicted remote loading efficiency as well as additional considerations such as availability, recommended dose and relevance to the disease. Three compounds were selected for experimental testing which were confirmed to be correctly classified by our previously reported QSPR models developed with Iterative Stochastic Elimination (ISE) and k-Nearest Neighbors (kNN) approaches. In addition, 10 new molecules with known liposome remote loading efficiency that were not used by us in QSPR model development were identified in the published literature and employed as an additional model validation set. The external accuracy of the models was found to be as high as 82% or 92%, depending on the model. This study presents the first successful application of QSPR models for the computer-model-driven design of liposomal drugs. © 2013.
Cern, Ahuva; Barenholz, Yechezkel; Tropsha, Alexander; Goldblum, Amiram
2014-01-01
Previously we have developed and statistically validated Quantitative Structure Property Relationship (QSPR) models that correlate drugs’ structural, physical and chemical properties as well as experimental conditions with the relative efficiency of remote loading of drugs into liposomes (Cern et al, Journal of Controlled Release, 160(2012) 14–157). Herein, these models have been used to virtually screen a large drug database to identify novel candidate molecules for liposomal drug delivery. Computational hits were considered for experimental validation based on their predicted remote loading efficiency as well as additional considerations such as availability, recommended dose and relevance to the disease. Three compounds were selected for experimental testing which were confirmed to be correctly classified by our previously reported QSPR models developed with Iterative Stochastic Elimination (ISE) and k-nearest neighbors (kNN) approaches. In addition, 10 new molecules with known liposome remote loading efficiency that were not used in QSPR model development were identified in the published literature and employed as an additional model validation set. The external accuracy of the models was found to be as high as 82% or 92%, depending on the model. This study presents the first successful application of QSPR models for the computer-model-driven design of liposomal drugs. PMID:24184343
Secure Encapsulation and Publication of Biological Services in the Cloud Computing Environment
Zhang, Weizhe; Wang, Xuehui; Lu, Bo; Kim, Tai-hoon
2013-01-01
Secure encapsulation and publication for bioinformatics software products based on web service are presented, and the basic function of biological information is realized in the cloud computing environment. In the encapsulation phase, the workflow and function of bioinformatics software are conducted, the encapsulation interfaces are designed, and the runtime interaction between users and computers is simulated. In the publication phase, the execution and management mechanisms and principles of the GRAM components are analyzed. The functions such as remote user job submission and job status query are implemented by using the GRAM components. The services of bioinformatics software are published to remote users. Finally the basic prototype system of the biological cloud is achieved. PMID:24078906
Secure encapsulation and publication of biological services in the cloud computing environment.
Zhang, Weizhe; Wang, Xuehui; Lu, Bo; Kim, Tai-hoon
2013-01-01
Secure encapsulation and publication for bioinformatics software products based on web service are presented, and the basic function of biological information is realized in the cloud computing environment. In the encapsulation phase, the workflow and function of bioinformatics software are conducted, the encapsulation interfaces are designed, and the runtime interaction between users and computers is simulated. In the publication phase, the execution and management mechanisms and principles of the GRAM components are analyzed. The functions such as remote user job submission and job status query are implemented by using the GRAM components. The services of bioinformatics software are published to remote users. Finally the basic prototype system of the biological cloud is achieved.
Interactive Screen Experiments--Innovative Virtual Laboratories for Distance Learners
ERIC Educational Resources Information Center
Hatherly, P. A.; Jordan, S. E.; Cayless, A.
2009-01-01
The desirability and value of laboratory work for physics students is a well-established principle and issues arise where students are inherently remote from their host institution, as is the case for the UK's Open University. In this paper, we present developments from the Physics Innovations Centre for Excellence in Teaching and Learning…
The IBM PC as an Online Search Machine. Part 6: Uploading.
ERIC Educational Resources Information Center
Kolner, Stuart J.
1986-01-01
This discussion of uploading (the transmission of information from local memory to a remote system) covers benefits, disadvantages, host system features, technical aspects, and four examples of uploading using CROSSTALK. Examples include function keys, searcher assistance with commonly-used terms, the use of hedges and boiler-plate searching, and…
MS Currie at RMS controls during EVA 2
2002-03-05
STS109-E-5625 (5 March 2002) --- Astronaut Nancy J. Currie, mission specialist, controls the Remote Manipulator System (RMS) robotic arm of the Space Shuttle Columbia as two astronauts perform work on the Hubble Space Telescope (HST), temporarily hosted in the shuttle's cargo bay. The image was recorded with a digital still camera.
Remote Adaptive Motor Resistance Training Exercise Apparatus and Method of Use Thereof
NASA Technical Reports Server (NTRS)
Reich, Alton (Inventor); Shaw, James (Inventor)
2017-01-01
The invention comprises a method and/or an apparatus using a computer configured exercise system equipped with an electric motor to provide physical resistance to user motion in conjunction with means for sharing exercise system related data and/or user performance data with a secondary user, such as a medical professional, a physical therapist, a trainer, a computer generated competitor, and/or a human competitor. For example, the exercise system is used with a remote trainer to enhance exercise performance, with a remote medical professional for rehabilitation, and/or with a competitor in a competition, such as in a power/weightlifting competition or in a video game. The exercise system is optionally configured with an intelligent software assistant and knowledge navigator functioning as a personal assistant application.
Structural monitoring for rare events in remote locations
NASA Astrophysics Data System (ADS)
Hale, J. M.
2005-01-01
A structural monitoring system has been developed for use on high value engineering structures, which is particularly suitable for use in remote locations where rare events such as accidental impacts, seismic activity or terrorist attack might otherwise go undetected. The system comprises a low power intelligent on-site data logger and a remote analysis computer that communicate with one another using the internet and mobile telephone technology. The analysis computer also generates e-mail alarms and maintains a web page that displays detected events in near real-time to authorised users. The application of the prototype system to pipeline monitoring is described in which the analysis of detected events is used to differentiate between impacts and pressure surges. The system has been demonstrated successfully and is ready for deployment.
Remote Adaptive Motor Resistance Training Exercise Apparatus and Method of Use Thereof
NASA Technical Reports Server (NTRS)
Shaw, James (Inventor); Reich, Alton (Inventor)
2016-01-01
The invention comprises a method and/or an apparatus using a computer configured exercise system equipped with an electric motor to provide physical resistance to user motion in conjunction with means for sharing exercise system related data and/or user performance data with a secondary user, such as a medical professional, a physical therapist, a trainer, a computer generated competitor, and/or a human competitor. For example, the exercise system is used with a remote trainer to enhance exercise performance, with a remote medical professional for rehabilitation, and/or with a competitor in a competition, such as in a power/weightlifting competition or in a video game. The exercise system is optionally configured with an intelligent software assistant and knowledge navigator functioning as a personal assistant application.
NASA Astrophysics Data System (ADS)
Nandigam, V.; Crosby, C. J.; Baru, C.; Arrowsmith, R.
2009-12-01
LIDAR is an excellent example of the new generation of powerful remote sensing data now available to Earth science researchers. Capable of producing digital elevation models (DEMs) more than an order of magnitude higher resolution than those currently available, LIDAR data allows earth scientists to study the processes that contribute to landscape evolution at resolutions not previously possible, yet essential for their appropriate representation. Along with these high-resolution datasets comes an increase in the volume and complexity of data that the user must efficiently manage and process in order for it to be scientifically useful. Although there are expensive commercial LIDAR software applications available, processing and analysis of these datasets are typically computationally inefficient on the conventional hardware and software that is currently available to most of the Earth science community. We have designed and implemented an Internet-based system, the OpenTopography Portal, that provides integrated access to high-resolution LIDAR data as well as web-based tools for processing of these datasets. By using remote data storage and high performance compute resources, the OpenTopography Portal attempts to simplify data access and standard LIDAR processing tasks for the Earth Science community. The OpenTopography Portal allows users to access massive amounts of raw point cloud LIDAR data as well as a suite of DEM generation tools to enable users to generate custom digital elevation models to best fit their science applications. The Cyberinfrastructure software tools for processing the data are freely available via the portal and conveniently integrated with the data selection in a single user-friendly interface. The ability to run these tools on powerful Cyberinfrastructure resources instead of their own labs provides a huge advantage in terms of performance and compute power. The system also encourages users to explore data processing methods and the variations in algorithm parameters since all of the processing is done remotely and numerous jobs can be submitted in sequence. The web-based software also eliminates the need for users to deal with the hassles and costs associated with software installation and licensing while providing adequate disk space for storage and personal job archival capability. Although currently limited to data access and DEM generation tasks, the OpenTopography system is modular in design and can be modified to accommodate new processing tools as they become available. We are currently exploring implementation of higher-level DEM analysis tasks in OpenTopography, since such processing is often computationally intensive and thus lends itself to utilization of cyberinfrastructure. Products derived from OpenTopography processing are available in a variety of formats ranging from simple Google Earth visualizations of LIDAR-derived hillshades to various GIS-compatible grid formats. To serve community users less interested in data processing, OpenTopography also hosts 1 km^2 digital elevation model tiles as well as Google Earth image overlays for a synoptic view of the data.
Greenland 1979 microwave remote sensing data catalog report, 14-15 October 1979
NASA Technical Reports Server (NTRS)
Hennigar, H. F.; Hirstein, W. S.; Schaffner, S. K.; Delnore, V. E.; Grantham, W. L.
1983-01-01
Microwave remote sensing measurements were cataloged for active and passive instruments in support of the 1979 Greenland Remote Sensing Experiment. Instruments used in this field experiment include the stepped frequency microwave radiometer (4 to 8 GHz) and the airborne microwave scatterometer (14.6 GHz). The microwave signature data are inventoried and cataloged in a user friendly format and are available on 9 track computer compatible tapes upon request.
A Computer Lab that Students Use but Never See
ERIC Educational Resources Information Center
Young, Jeffrey R.
2008-01-01
North Carolina State University may never build another computer lab. Instead the university has installed racks of equipment in windowless rooms where students and professors never go. This article describes a project called the Virtual Computing Lab. Users enter it remotely from their own computers in dormitory rooms or libraries. They get all…
Computer Exercises in Systems and Fields Experiments
ERIC Educational Resources Information Center
Bacon, C. M.; McDougal, J. R.
1971-01-01
Laboratory activities give students an opportunity to interact with computers in modes ranging from remote terminal use in laboratory experimentation to the direct hands-on use of a small digital computer with disk memory and on-line plotter, and finally to the use of a large computer under closed-shop operation. (Author/TS)
Remote control system for high-perfomance computer simulation of crystal growth by the PFC method
NASA Astrophysics Data System (ADS)
Pavlyuk, Evgeny; Starodumov, Ilya; Osipov, Sergei
2017-04-01
Modeling of crystallization process by the phase field crystal method (PFC) - one of the important directions of modern computational materials science. In this paper, the practical side of the computer simulation of the crystallization process by the PFC method is investigated. To solve problems using this method, it is necessary to use high-performance computing clusters, data storage systems and other often expensive complex computer systems. Access to such resources is often limited, unstable and accompanied by various administrative problems. In addition, the variety of software and settings of different computing clusters sometimes does not allow researchers to use unified program code. There is a need to adapt the program code for each configuration of the computer complex. The practical experience of the authors has shown that the creation of a special control system for computing with the possibility of remote use can greatly simplify the implementation of simulations and increase the performance of scientific research. In current paper we show the principal idea of such a system and justify its efficiency.
Remote gaze tracking system for 3D environments.
Congcong Liu; Herrup, Karl; Shi, Bertram E
2017-07-01
Eye tracking systems are typically divided into two categories: remote and mobile. Remote systems, where the eye tracker is located near the object being viewed by the subject, have the advantage of being less intrusive, but are typically used for tracking gaze points on fixed two dimensional (2D) computer screens. Mobile systems such as eye tracking glasses, where the eye tracker are attached to the subject, are more intrusive, but are better suited for cases where subjects are viewing objects in the three dimensional (3D) environment. In this paper, we describe how remote gaze tracking systems developed for 2D computer screens can be used to track gaze points in a 3D environment. The system is non-intrusive. It compensates for small head movements by the user, so that the head need not be stabilized by a chin rest or bite bar. The system maps the 3D gaze points of the user onto 2D images from a scene camera and is also located remotely from the subject. Measurement results from this system indicate that it is able to estimate gaze points in the scene camera to within one degree over a wide range of head positions.
Maia, João P; Alvares, Francisco; Boratyński, Zbyszek; Brito, José C; Leite, João V; Harris, D James
2014-10-01
Parasites play a major role in ecosystems, and understanding of host-parasite interactions is important for predicting parasite transmission dynamics and epidemiology. However, there is still a lack of knowledge about the distribution, diversity, and impact of parasites in wildlife, especially from remote areas. Hepatozoon is a genus of apicomplexan parasites that is transmitted by ingestion of infected arthropod vectors. However, alternative modes of transmission have been identified such as trophic transmission. Using the 18S rRNA gene as a marker, we provide an assessment of Hepatozoon prevalence in six wild canid and two rodent species collected between 2003 and 2012 from remote areas in North Africa. By combining this with other predator-prey systems in a phylogenetic framework, we investigate Hepatozoon transmission dynamics in distinct host taxa. Prevalence was high overall among host species (African jerboa Jaculus jaculus [17/47, 36%], greater Egyptian jerboa Jaculus orientalis [5/7, 71%], side-striped jackal Canis adustus [1/2, 50%], golden jackal Canis aureus [6/32, 18%], pale fox Vulpes pallida [14/28, 50%], Rüppell's fox Vulpes rueppellii [6/11, 55%], red fox Vulpes vulpes [8/16, 50%], and fennec fox Vulpes zerda [7/11, 42%]). Phylogenetic analysis showed further evidence of occasional transmission of Hepatozoon lineages from prey to canid predators, which seems to occur less frequently than in other predator-prey systems such as between snakes and lizards. Due to the complex nature of the Hepatozoon lifecycle (heteroxenous and vector-borne), future studies on these wild host species need to clarify the dynamics of alternative modes of Hepatozoon transmission and identify reservoir and definitive hosts in natural populations. We also detected putative Babesia spp. (Apicomplexa: Piroplasmida) infections in two canid species from this region, V. pallida (1/28) and V. zerda (1/11).
Investigating a holobiont: Microbiota perturbations and transkingdom networks.
Greer, Renee; Dong, Xiaoxi; Morgun, Andrey; Shulzhenko, Natalia
2016-01-01
The scientific community has recently come to appreciate that, rather than existing as independent organisms, multicellular hosts and their microbiota comprise a complex evolving superorganism or metaorganism, termed a holobiont. This point of view leads to a re-evaluation of our understanding of different physiological processes and diseases. In this paper we focus on experimental and computational approaches which, when combined in one study, allowed us to dissect mechanisms (traditionally named host-microbiota interactions) regulating holobiont physiology. Specifically, we discuss several approaches for microbiota perturbation, such as use of antibiotics and germ-free animals, including advantages and potential caveats of their usage. We briefly review computational approaches to characterize the microbiota and, more importantly, methods to infer specific components of microbiota (such as microbes or their genes) affecting host functions. One such approach called transkingdom network analysis has been recently developed and applied in our study. (1) Finally, we also discuss common methods used to validate the computational predictions of host-microbiota interactions using in vitro and in vivo experimental systems.
Using plant canopy temperature to improve irrigated crop management
USDA-ARS?s Scientific Manuscript database
Remotely sensed plant canopy temperature has long been recognized as having potential as a tool for irrigation management. However, a number of barriers have prevented its routine use in practice, such as the spatial and temporal resolution of remote sensing platforms, limitations in computing capac...
The use of multimedia and programmed teaching machines for remote sensing education
NASA Technical Reports Server (NTRS)
Ulliman, J. J.
1980-01-01
The advantages, limitations, and uses of various audio visual equipments and techniques used in various universities for individualized and group instruction in the interpretation and classification of remotely sensed data are considered as well as systems for programmed and computer-assisted instruction.
Can hyperspectral remote sensing detect species specific biochemicals?
USDA-ARS?s Scientific Manuscript database
Discrimination of a few plants scattered among many plants is a goal common to detection of agricultural weeds and invasive species. Detection of clandestinely grown Cannabis sativa L. is in many ways a special case of weed detection. Remote sensing technology provides an automated, computer based,...
NASA Technical Reports Server (NTRS)
Faller, K. H.
1976-01-01
A technique for the detection and measurement of surface feature interfaces in remotely acquired data was developed and evaluated. A computer implementation of this technique was effected to automatically process classified data derived from various sources such as the LANDSAT multispectral scanner and other scanning sensors. The basic elements of the operational theory of the technique are described, followed by the details of the procedure. An example of an application of the technique to the analysis of tidal shoreline length is given with a breakdown of manpower requirements.
Advancements in remote physiological measurement and applications in human-computer interaction
NASA Astrophysics Data System (ADS)
McDuff, Daniel
2017-04-01
Physiological signals are important for tracking health and emotional states. Imaging photoplethysmography (iPPG) is a set of techniques for remotely recovering cardio-pulmonary signals from video of the human body. Advances in iPPG methods over the past decade combined with the ubiquity of digital cameras presents the possibility for many new, lowcost applications of physiological monitoring. This talk will highlight methods for recovering physiological signals, work characterizing the impact of video parameters and hardware on these measurements, and applications of this technology in human-computer interfaces.
Where the Cloud Meets the Commons
ERIC Educational Resources Information Center
Ipri, Tom
2011-01-01
Changes presented by cloud computing--shared computing services, applications, and storage available to end users via the Internet--have the potential to seriously alter how libraries provide services, not only remotely, but also within the physical library, specifically concerning challenges facing the typical desktop computing experience.…
Computer Networking with the Victorian Correspondence School.
ERIC Educational Resources Information Center
Conboy, Ian
During 1985 the Education Department installed two-way radios in 44 remote secondary schools in Victoria, Australia, to improve turn-around time for correspondence assignments. Subsequently, teacher supervisors at Melbourne's Correspondence School sought ways to further augument audio interactivity with computer networking. Computer equipment was…
System enhancements of Mesoscale Analysis and Space Sensor (MASS) computer system
NASA Technical Reports Server (NTRS)
Hickey, J. S.; Karitani, S.
1985-01-01
The interactive information processing for the mesoscale analysis and space sensor (MASS) program is reported. The development and implementation of new spaceborne remote sensing technology to observe and measure atmospheric processes is described. The space measurements and conventional observational data are processed together to gain an improved understanding of the mesoscale structure and dynamical evolution of the atmosphere relative to cloud development and precipitation processes. A Research Computer System consisting of three primary computers was developed (HP-1000F, Perkin-Elmer 3250, and Harris/6) which provides a wide range of capabilities for processing and displaying interactively large volumes of remote sensing data. The development of a MASS data base management and analysis system on the HP-1000F computer and extending these capabilities by integration with the Perkin-Elmer and Harris/6 computers using the MSFC's Apple III microcomputer workstations is described. The objectives are: to design hardware enhancements for computer integration and to provide data conversion and transfer between machines.
NH11B-1726: FrankenRaven: A New Platform for Remote Sensing
NASA Technical Reports Server (NTRS)
Dahlgren, Robert; Fladeland, Matthew M.; Pinsker, Ethan A.; Jasionowicz, John P.; Jones, Lowell L.; Pscheid, Matthew J.
2016-01-01
Small, modular aircraft are an emerging technology with a goal to maximize flexibility and enable multi-mission support. This reports the progress of an unmanned aerial system (UAS) project conducted at the NASA Ames Research Center (ARC) in 2016. This interdisciplinary effort builds upon the success of the 2014 FrankenEye project to apply rapid prototyping techniques to UAS, to develop a variety of platforms to host remote sensing instruments. In 2016, ARC received AeroVironment RQ-11A and RQ-11B Raven UAS from the US Department of the Interior, Office of Aviation Services. These aircraft have electric propulsion, a wingspan of roughly 1.3m, and have demonstrated reliability in challenging environments. The Raven airframe is an ideal foundation to construct more complex aircraft, and student interns using 3D printing were able to graft multiple Raven wings and fuselages into FrankenRaven aircraft. Aeronautical analysis shows that the new configuration has enhanced flight time, payload capacity, and distance compared to the original Raven. The FrankenRaven avionics architecture replaces the mil-spec avionics with COTS technology based upon the 3DR Pixhawk PX4 autopilot with a safety multiplexer for failsafe handoff to 2.4 GHz RC control and 915 MHz telemetry. This project demonstrates how design reuse, rapid prototyping, and modular subcomponents can be leveraged into flexible airborne platforms that can host a variety of remote sensing payloads and even multiple payloads. Modularity advances a new paradigm: mass-customization of aircraft around given payload(s). Multi-fuselage designs are currently under development to host a wide variety of payloads including a zenith-pointing spectrometer, a magnetometer, a multi-spectral camera, and a RGB camera. After airworthiness certification, flight readiness review, and test flights are performed at Crows Landing airfield in central California, field data will be taken at Kilauea volcano in Hawaii and other locations.
FrankenRaven: A New Platform for Remote Sensing
NASA Astrophysics Data System (ADS)
Dahlgren, R. P.; Fladeland, M. M.; Pinsker, E. A.; Jasionowicz, J. P.; Jones, L. L.; Mosser, C. D.; Pscheid, M. J.; Weidow, N. L.; Kelly, P. J.; Kern, C.; Werner, C. A.; Johnson, M. S.
2016-12-01
Small, modular aircraft are an emerging technology with a goal to maximize flexibility and enable multi-mission support. This reports the progress of an unmanned aerial system (UAS) project conducted at the NASA Ames Research Center (ARC) in 2016. This interdisciplinary effort builds upon the success of the 2014 FrankenEye project to apply rapid prototyping techniques to UAS, to develop a variety of platforms to host remote sensing instruments. In 2016, ARC received AeroVironment RQ-11A and RQ-11B Raven UAS from the US Department of the Interior, Office of Aviation Services. These aircraft have electric propulsion, a wingspan of roughly 1.3m, and have demonstrated reliability in challenging environments. The Raven airframe is an ideal foundation to construct more complex aircraft, and student interns using 3D printing were able to graft multiple Raven wings and fuselages into "FrankenRaven" aircraft. Aeronautical analysis shows that the new configuration has enhanced flight time, payload capacity, and distance compared to the original Raven. The FrankenRaven avionics architecture replaces the mil-spec avionics with COTS technology based upon the 3DR Pixhawk PX4 autopilot with a safety multiplexer for failsafe handoff to 2.4 GHz RC control and 915 MHz telemetry. This project demonstrates how design reuse, rapid prototyping, and modular subcomponents can be leveraged into flexible airborne platforms that can host a variety of remote sensing payloads and even multiple payloads. Modularity advances a new paradigm: mass-customization of aircraft around given payload(s). Multi-fuselage designs are currently under development to host a wide variety of payloads including a zenith-pointing spectrometer, a magnetometer, a multi-spectral camera, and a RGB camera. After airworthiness certification, flight readiness review, and test flights are performed at Crows Landing airfield in central California, field data will be taken at Kilauea volcano in Hawaii and other locations.
Online catalog access and distribution of remotely sensed information
NASA Astrophysics Data System (ADS)
Lutton, Stephen M.
1997-09-01
Remote sensing is providing voluminous data and value added information products. Electronic sensors, communication electronics, computer software, hardware, and network communications technology have matured to the point where a distributed infrastructure for remotely sensed information is a reality. The amount of remotely sensed data and information is making distributed infrastructure almost a necessity. This infrastructure provides data collection, archiving, cataloging, browsing, processing, and viewing for applications from scientific research to economic, legal, and national security decision making. The remote sensing field is entering a new exciting stage of commercial growth and expansion into the mainstream of government and business decision making. This paper overviews this new distributed infrastructure and then focuses on describing a software system for on-line catalog access and distribution of remotely sensed information.
NASA Technical Reports Server (NTRS)
Miller, L. D.; Tom, C.; Nualchawee, K.
1977-01-01
A tropical forest area of Northern Thailand provided a test case of the application of the approach in more natural surroundings. Remote sensing imagery subjected to proper computer analysis has been shown to be a very useful means of collecting spatial data for the science of hydrology. Remote sensing products provide direct input to hydrologic models and practical data bases for planning large and small-scale hydrologic developments. Combining the available remote sensing imagery together with available map information in the landscape model provides a basis for substantial improvements in these applications.
Remote access to very large image repositories, a high performance computing perspective
NASA Technical Reports Server (NTRS)
Plesea, Lucian
2005-01-01
The main challenges of using the increasingly large repositories of remote imagery data can be summarized in one word: efficiency. In this paper, a number of concrete problems and the chosen solutions are described, based on the construction of a 5TB global Landsat 7 mosaic.
Remote Labs and Game-Based Learning for Process Control
ERIC Educational Resources Information Center
Zualkernan, Imran A.; Husseini, Ghaleb A.; Loughlin, Kevin F.; Mohebzada, Jamshaid G.; El Gaml, Moataz
2013-01-01
Social networking platforms and computer games represent a natural informal learning environment for the current generation of learners in higher education. This paper explores the use of game-based learning in the context of an undergraduate chemical engineering remote laboratory. Specifically, students are allowed to manipulate chemical…
Update and review of accuracy assessment techniques for remotely sensed data
NASA Technical Reports Server (NTRS)
Congalton, R. G.; Heinen, J. T.; Oderwald, R. G.
1983-01-01
Research performed in the accuracy assessment of remotely sensed data is updated and reviewed. The use of discrete multivariate analysis techniques for the assessment of error matrices, the use of computer simulation for assessing various sampling strategies, and an investigation of spatial autocorrelation techniques are examined.
Adaptive Control Of Remote Manipulator
NASA Technical Reports Server (NTRS)
Seraji, Homayoun
1989-01-01
Robotic control system causes remote manipulator to follow closely reference trajectory in Cartesian reference frame in work space, without resort to computationally intensive mathematical model of robot dynamics and without knowledge of robot and load parameters. System, derived from linear multivariable theory, uses relatively simple feedforward and feedback controllers with model-reference adaptive control.
Audiographics for Distance Education: An Alternative Technology.
ERIC Educational Resources Information Center
Fredrickson, Scott
Audiographics is the merging of microcomputer graphics, telephone communications systems, and teaching strategies into a cost effective method of delivering distance education classes. The teacher creates visual images that are sent to and stored on computers at the remote sites. At the appropriate time the teacher and the remote site assistants…
Remotely Sensed Information and Field Data are both Essential to Assess Biodiversity CONDITION!
NASA Astrophysics Data System (ADS)
Sparrow, B.; Schaefer, M.; Scarth, P.; Phinn, S. R.; Christensen, R.; Lowe, A. J.; O'Neill, S.; Thurgate, N.; Wundke, D.
2015-12-01
Over the past year the TERN Ausplots facility has hosted a process to determine the definition of Biodiversity Condition in an Australian Continental Context, and conducted a wide collaborative process to determine which environmental attributes are required to be measures to accurately inform on biodiversity condition. A major output from this work was the acknowledgement that good quality data from both remotely sensed sources and good quality field collected data are both essential to provide the best information possible on biodiversity condition. This poster details some background to the project, the assesment of which attributes to measure, and if the are sources primarily from field based or remotely sensed measures. It then proceeds to provide three examples of ways in which the combination of data types provides a superior product as output, with one example being provided for the three cornerstone areas of condition: Structure, Function and Composition.
Remote sensing programs and courses in engineering and water resources
NASA Technical Reports Server (NTRS)
Kiefer, R. W.
1981-01-01
The content of typical basic and advanced remote sensing and image interpretation courses are described and typical remote sensing graduate programs of study in civil engineering and in interdisciplinary environmental remote sensing and water resources management programs are outlined. Ideally, graduate programs with an emphasis on remote sensing and image interpretation should be built around a core of five courses: (1) a basic course in fundamentals of remote sensing upon which the more specialized advanced remote sensing courses can build; (2) a course dealing with visual image interpretation; (3) a course dealing with quantitative (computer-based) image interpretation; (4) a basic photogrammetry course; and (5) a basic surveying course. These five courses comprise up to one-half of the course work required for the M.S. degree. The nature of other course work and thesis requirements vary greatly, depending on the department in which the degree is being awarded.
"One-Stop Shopping" for Ocean Remote-Sensing and Model Data
NASA Technical Reports Server (NTRS)
Li, P. Peggy; Vu, Quoc; Chao, Yi; Li, Zhi-Jin; Choi, Jei-Kook
2006-01-01
OurOcean Portal 2.0 (http:// ourocean.jpl.nasa.gov) is a software system designed to enable users to easily gain access to ocean observation data, both remote-sensing and in-situ, configure and run an Ocean Model with observation data assimilated on a remote computer, and visualize both the observation data and the model outputs. At present, the observation data and models focus on the California coastal regions and Prince William Sound in Alaska. This system can be used to perform both real-time and retrospective analyses of remote-sensing data and model outputs. OurOcean Portal 2.0 incorporates state-of-the-art information technologies (IT) such as MySQL database, Java Web Server (Apache/Tomcat), Live Access Server (LAS), interactive graphics with Java Applet at the Client site and MatLab/GMT at the server site, and distributed computing. OurOcean currently serves over 20 real-time or historical ocean data products. The data are served in pre-generated plots or their native data format. For some of the datasets, users can choose different plotting parameters and produce customized graphics. OurOcean also serves 3D Ocean Model outputs generated by ROMS (Regional Ocean Model System) using LAS. The Live Access Server (LAS) software, developed by the Pacific Marine Environmental Laboratory (PMEL) of the National Oceanic and Atmospheric Administration (NOAA), is a configurable Web-server program designed to provide flexible access to geo-referenced scientific data. The model output can be views as plots in horizontal slices, depth profiles or time sequences, or can be downloaded as raw data in different data formats, such as NetCDF, ASCII, Binary, etc. The interactive visualization is provided by graphic software, Ferret, also developed by PMEL. In addition, OurOcean allows users with minimal computing resources to configure and run an Ocean Model with data assimilation on a remote computer. Users may select the forcing input, the data to be assimilated, the simulation period, and the output variables and submit the model to run on a backend parallel computer. When the run is complete, the output will be added to the LAS server for
Web-GIS visualisation of permafrost-related Remote Sensing products for ESA GlobPermafrost
NASA Astrophysics Data System (ADS)
Haas, A.; Heim, B.; Schaefer-Neth, C.; Laboor, S.; Nitze, I.; Grosse, G.; Bartsch, A.; Kaab, A.; Strozzi, T.; Wiesmann, A.; Seifert, F. M.
2016-12-01
The ESA GlobPermafrost (www.globpermafrost.info) provides a remote sensing service for permafrost research and applications. The service comprises of data product generation for various sites and regions as well as specific infrastructure allowing overview and access to datasets. Based on an online user survey conducted within the project, the user community extensively applies GIS software to handle remote sensing-derived datasets and requires preview functionalities before accessing them. In response, we develop the Permafrost Information System PerSys which is conceptualized as an open access geospatial data dissemination and visualization portal. PerSys will allow visualisation of GlobPermafrost raster and vector products such as land cover classifications, Landsat multispectral index trend datasets, lake and wetland extents, InSAR-based land surface deformation maps, rock glacier velocity fields, spatially distributed permafrost model outputs, and land surface temperature datasets. The datasets will be published as WebGIS services relying on OGC-standardized Web Mapping Service (WMS) and Web Feature Service (WFS) technologies for data display and visualization. The WebGIS environment will be hosted at the AWI computing centre where a geodata infrastructure has been implemented comprising of ArcGIS for Server 10.4, PostgreSQL 9.2 and a browser-driven data viewer based on Leaflet (http://leafletjs.com). Independently, we will provide an `Access - Restricted Data Dissemination Service', which will be available to registered users for testing frequently updated versions of project datasets. PerSys will become a core project of the Arctic Permafrost Geospatial Centre (APGC) within the ERC-funded PETA-CARB project (www.awi.de/petacarb). The APGC Data Catalogue will contain all final products of GlobPermafrost, allow in-depth dataset search via keywords, spatial and temporal coverage, data type, etc., and will provide DOI-based links to the datasets archived in the long-term, open access PANGAEA data repository.
A new respiratory monitoring and processing system based on Wii remote: proof of principle.
Peng, Y; Vedam, S; Gao, S; Balter, P
2013-07-01
To create a patient respiratory management system and patient self-practice tool using the Wii remote, a widely available consumer hardware product. The Wii remote (Wiimote) (Nintendo, Redmond, WA) contains an infrared (IR) camera that can track up to four spots whose coordinates are reported to a host computer via Bluetooth. The Wiimote is capable of tracking a fiducial box currently used by a commercial monitoring system [Real-time Position Management(TM) (RPM) system, Varian Associates, Palo Alto, CA], if the correct IR source is used. The authors validated the Wiimote tracking by comparing the amplitude and frequency of signals among those reported by Wiimote with known movements from an inhouse servo-driven respiratory simulator, as well as with those measured using the RPM. The simulator comparison was done using standard sinusoid signals with amplitude of 2.0 cm as well as recorded patient respiratory traces. The RPM comparisons were done by simultaneously recording the RPM reflective box position with the Wiimote and the RPM. Timing was compared between these two systems by using the digital beam-on signal from the CT scanner, for the 4DCT to synchronize these acquisitions. The data acquisition rate from the Wiimote was 100.0 ± 0.4 Hz with a version 2.1 Bluetooth adaptor. The standard deviation of the height of the motion extrema was 0.06 and 1.1 mm when comparing those measured by the Wiimote and the servomotor encoder for standard sinusoid signal and prerecorded patient respiratory signal, respectively. The standard deviation of the amplitude of motion extrema between the Wiimote and RPM was 0.9 mm and the timing difference was 253 ms. The performance of Wiimote shows promise for respiratory monitoring for its faster sampling rate as well as the potential optical and GPU abilities. If used with care it can deliver reasonable spatial and temporal accuracy.
Archer, Charles J.; Blocksome, Michael A.
2012-12-11
Methods, parallel computers, and computer program products are disclosed for remote direct memory access. Embodiments include transmitting, from an origin DMA engine on an origin compute node to a plurality target DMA engines on target compute nodes, a request to send message, the request to send message specifying a data to be transferred from the origin DMA engine to data storage on each target compute node; receiving, by each target DMA engine on each target compute node, the request to send message; preparing, by each target DMA engine, to store data according to the data storage reference and the data length, including assigning a base storage address for the data storage reference; sending, by one or more of the target DMA engines, an acknowledgment message acknowledging that all the target DMA engines are prepared to receive a data transmission from the origin DMA engine; receiving, by the origin DMA engine, the acknowledgement message from the one or more of the target DMA engines; and transferring, by the origin DMA engine, data to data storage on each of the target compute nodes according to the data storage reference using a single direct put operation.
Host computer software specifications for a zero-g payload manhandling simulator
NASA Technical Reports Server (NTRS)
Wilson, S. W.
1986-01-01
The HP PASCAL source code was developed for the Mission Planning and Analysis Division (MPAD) of NASA/JSC, and takes the place of detailed flow charts defining the host computer software specifications for MANHANDLE, a digital/graphical simulator that can be used to analyze the dynamics of onorbit (zero-g) payload manhandling operations. Input and output data for representative test cases are contained.
How the microbiota shapes rheumatic diseases.
Van de Wiele, Tom; Van Praet, Jens T; Marzorati, Massimo; Drennan, Michael B; Elewaut, Dirk
2016-07-01
The human gut harbours a tremendously diverse and abundant microbial community that correlates with, and even modulates, many health-related processes. The mucosal interfaces are particularly active sites of microorganism-host interplay. Growing insight into the characteristic composition and functionality of the mucosal microbiota has revealed that the microbiota is involved in mucosal barrier integrity and immune function. This involvement affects proinflammatory and anti-inflammatory processes not only at the epithelial level, but also at remote sites such as the joints. Here, we review the role of the gut microbiota in shaping local and systemic immune responses and how disturbances in the host-microorganism interplay can potentially affect the development and progression of rheumatic diseases. Increasing our understanding of how to promote host-microorganism homeostasis could therefore reveal novel strategies for the prevention or alleviation of rheumatic disease.
A study of application of remote sensing to river forecasting. Volume 1: Executive summary
NASA Technical Reports Server (NTRS)
1975-01-01
A project is described whose goal was to define, implement and evaluate a pilot demonstration test to show the practicability of applying remotely sensed data to operational river forecasting in gaged or previously ungaged watersheds. A secondary objective was to provide NASA with documentation describing the computer programs that comprise the streamflow forecasting simulation model used. A computer-based simulation model was adapted to a streamflow forecasting application and implemented in an IBM System/360 Model 44 computer, operating in a dedicated mode, with operator interactive control through a Model 2250 keyboard/graphic CRT terminal. The test site whose hydrologic behavior was simulated is a small basin (365 square kilometers) designated Town Creek near Geraldine, Alabama.
The Thirty Meter Telescope Site Testing Robotic Computer System
NASA Astrophysics Data System (ADS)
Riddle, Reed L.; Schöck, M.; Skidmore, W.; Els, S.; Travouillon, T.
2008-03-01
The Thirty Meter Telescope (TMT) project is currently testing five remote sites as candidates for the final location of the telescope. Each site has several instruments, including seeing monitors, weather stations, and turbulence profile measuring systems, each of which is computer controlled. As the sites are remote, they require a control system that can automatically manage the operations of all the varied subsystems, keep the systems safe from damage and recover from errors during operation. The robotic system must also be robust enough to operate without human intervention and when internet connections are lost. It is also critical that a data archiving system diligently records all data as gathered. This is a discussion of the TMT site testing robotic computer system as implemented.
Using remote sensing and machine learning for the spatial modelling of a bluetongue virus vector
NASA Astrophysics Data System (ADS)
Van doninck, J.; Peters, J.; De Baets, B.; Ducheyne, E.; Verhoest, N. E. C.
2012-04-01
Bluetongue is a viral vector-borne disease transmitted between hosts, mostly cattle and small ruminants, by some species of Culicoides midges. Within the Mediterranean basin, C. imicola is the main vector of the bluetongue virus. The spatial distribution of this species is limited by a number of environmental factors, including temperature, soil properties and land cover. The identification of zones at risk of bluetongue outbreaks thus requires detailed information on these environmental factors, as well as appropriate epidemiological modelling techniques. We here give an overview of the environmental factors assumed to be constraining the spatial distribution of C. imicola, as identified in different studies. Subsequently, remote sensing products that can be used as proxies for these environmental constraints are presented. Remote sensing data are then used together with species occurrence data from the Spanish Bluetongue National Surveillance Programme to calibrate a supervised learning model, based on Random Forests, to model the probability of occurrence of the C. imicola midge. The model will then be applied for a pixel-based prediction over the Iberian peninsula using remote sensing products for habitat characterization.
An empirical InSAR-optical fusion approach to mapping vegetation canopy height
Wayne S. Walker; Josef M. Kellndorfer; Elizabeth LaPoint; Michael Hoppus; James Westfall
2007-01-01
Exploiting synergies afforded by a host of recently available national-scale data sets derived from interferometric synthetic aperture radar (InSAR) and passive optical remote sensing, this paper describes the development of a novel empirical approach for the provision of regional- to continental-scale estimates of vegetation canopy height. Supported by data from the...
South Carolina's SC LENDS: Optimizing Libraries, Transforming Lending
ERIC Educational Resources Information Center
Hamby, Rogan; McBride, Ray; Lundberg, Maria
2011-01-01
Since SC LENDS started operating in June 2009, more public libraries have come on board. All of this on the back end connects to a Mozilla-based staff client that has distributions for Mac OS X and Microsoft Windows, using SSL encryption to keep communications secure and private between remote libraries and the servers hosted at a high-end…
ERIC Educational Resources Information Center
Stifle, Jack
The PLATO IV computer-based instructional system consists of a large scale centrally located CDC 6400 computer and a large number of remote student terminals. This is a brief and general description of the proposed input/output hardware necessary to interface the student terminals with the computer's central processing unit (CPU) using available…
Central Computational Facility CCF communications subsystem options
NASA Technical Reports Server (NTRS)
Hennigan, K. B.
1979-01-01
A MITRE study which investigated the communication options available to support both the remaining Central Computational Facility (CCF) computer systems and the proposed U1108 replacements is presented. The facilities utilized to link the remote user terminals with the CCF were analyzed and guidelines to provide more efficient communications were established.
Digital Data Transmission Via CATV.
ERIC Educational Resources Information Center
Stifle, Jack; And Others
A low cost communications network has been designed for use in the PLATO IV computer-assisted instruction system. Over 1,000 remote computer graphic terminals each requiring a 1200 bps channel are to be connected to one centrally located computer. Digital data are distributed to these terminals using standard commercial cable television (CATV)…
2007-09-01
example, an application developed in Sun’s Netbeans [2007] integrated development environment (IDE) uses Swing class object for graphical user... Netbeans Version 5.5.1 [Computer Software]. Santa Clara, CA: Sun Microsystems. Process Modeler Version 7.0 [Computer Software]. Santa Clara, Ca
A Spacelab Expert System for Remote Engineering and Science
NASA Technical Reports Server (NTRS)
Groleau, Nick; Colombano, Silvano; Friedland, Peter (Technical Monitor)
1994-01-01
NASA's space science program is based on strictly pre-planned activities. This approach does not always result in the best science. We describe an existing computer system that enables space science to be conducted in a more reactive manner through advanced automation techniques that have recently been used in SLS-2 October 1993 space shuttle flight. Advanced computing techniques, usually developed in the field of Artificial Intelligence, allow large portions of the scientific investigator's knowledge to be "packaged" in a portable computer to present advice to the astronaut operator. We strongly believe that this technology has wide applicability to other forms of remote science/engineering. In this brief article, we present the technology of remote science/engineering assistance as implemented for the SLS-2 space shuttle flight. We begin with a logical overview of the system (paying particular attention to the implementation details relevant to the use of the embedded knowledge for system reasoning), then describe its use and success in space, and conclude with ideas about possible earth uses of the technology in the life and medical sciences.
NASA Astrophysics Data System (ADS)
Tysowski, Piotr K.; Ling, Xinhua; Lütkenhaus, Norbert; Mosca, Michele
2018-04-01
Quantum key distribution (QKD) is a means of generating keys between a pair of computing hosts that is theoretically secure against cryptanalysis, even by a quantum computer. Although there is much active research into improving the QKD technology itself, there is still significant work to be done to apply engineering methodology and determine how it can be practically built to scale within an enterprise IT environment. Significant challenges exist in building a practical key management service (KMS) for use in a metropolitan network. QKD is generally a point-to-point technique only and is subject to steep performance constraints. The integration of QKD into enterprise-level computing has been researched, to enable quantum-safe communication. A novel method for constructing a KMS is presented that allows arbitrary computing hosts on one site to establish multiple secure communication sessions with the hosts of another site. A key exchange protocol is proposed where symmetric private keys are granted to hosts while satisfying the scalability needs of an enterprise population of users. The KMS operates within a layered architectural style that is able to interoperate with various underlying QKD implementations. Variable levels of security for the host population are enforced through a policy engine. A network layer provides key generation across a network of nodes connected by quantum links. Scheduling and routing functionality allows quantum key material to be relayed across trusted nodes. Optimizations are performed to match the real-time host demand for key material with the capacity afforded by the infrastructure. The result is a flexible and scalable architecture that is suitable for enterprise use and independent of any specific QKD technology.
NASA Technical Reports Server (NTRS)
Mckay, C. W.; Bown, R. L.
1985-01-01
The space station data management system involves networks of computing resources that must work cooperatively and reliably over an indefinite life span. This program requires a long schedule of modular growth and an even longer period of maintenance and operation. The development and operation of space station computing resources will involve a spectrum of systems and software life cycle activities distributed across a variety of hosts, an integration, verification, and validation host with test bed, and distributed targets. The requirement for the early establishment and use of an apporopriate Computer Systems and Software Engineering Support Environment is identified. This environment will support the Research and Development Productivity challenges presented by the space station computing system.
UTOPIA-User-Friendly Tools for Operating Informatics Applications.
Pettifer, S R; Sinnott, J R; Attwood, T K
2004-01-01
Bioinformaticians routinely analyse vast amounts of information held both in large remote databases and in flat data files hosted on local machines. The contemporary toolkit available for this purpose consists of an ad hoc collection of data manipulation tools, scripting languages and visualization systems; these must often be combined in complex and bespoke ways, the result frequently being an unwieldy artefact capable of one specific task, which cannot easily be exploited or extended by other practitioners. Owing to the sizes of current databases and the scale of the analyses necessary, routine bioinformatics tasks are often automated, but many still require the unique experience and intuition of human researchers: this requires tools that support real-time interaction with complex datasets. Many existing tools have poor user interfaces and limited real-time performance when applied to realistically large datasets; much of the user's cognitive capacity is therefore focused on controlling the tool rather than on performing the research. The UTOPIA project is addressing some of these issues by building reusable software components that can be combined to make useful applications in the field of bioinformatics. Expertise in the fields of human computer interaction, high-performance rendering, and distributed systems is being guided by bioinformaticians and end-user biologists to create a toolkit that is both architecturally sound from a computing point of view, and directly addresses end-user and application-developer requirements.
Enabling Research Network Connectivity to Clouds with Virtual Router Technology
NASA Astrophysics Data System (ADS)
Seuster, R.; Casteels, K.; Leavett-Brown, CR; Paterson, M.; Sobie, RJ
2017-10-01
The use of opportunistic cloud resources by HEP experiments has significantly increased over the past few years. Clouds that are owned or managed by the HEP community are connected to the LHCONE network or the research network with global access to HEP computing resources. Private clouds, such as those supported by non-HEP research funds are generally connected to the international research network; however, commercial clouds are either not connected to the research network or only connect to research sites within their national boundaries. Since research network connectivity is a requirement for HEP applications, we need to find a solution that provides a high-speed connection. We are studying a solution with a virtual router that will address the use case when a commercial cloud has research network connectivity in a limited region. In this situation, we host a virtual router in our HEP site and require that all traffic from the commercial site transit through the virtual router. Although this may increase the network path and also the load on the HEP site, it is a workable solution that would enable the use of the remote cloud for low I/O applications. We are exploring some simple open-source solutions. In this paper, we present the results of our studies and how it will benefit our use of private and public clouds for HEP computing.
[Design and application of user managing system of cardiac remote monitoring network].
Chen, Shouqiang; Zhang, Jianmin; Yuan, Feng; Gao, Haiqing
2007-12-01
According to inpatient records, data managing demand of cardiac remote monitoring network and computer, this software was designed with relative database ACCESS. Its interface, operational button and menu were designed in VBA language assistantly. Its design included collective design, amity, practicability and compatibility. Its function consisted of registering, inquiring, statisticing and printing, et al. It could be used to manage users effectively and could be helpful to exerting important action of cardiac remote monitoring network in preventing cardiac-vascular emergency ulteriorly.
1994 ASPRS/ACSM annual convention exposition. Volume 2
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1994-01-01
This report is Volume II of presented papers at the joint 1994 convention of the American Society for Photgrammetry and Remote Sensing and American Congress on Surveying and Mapping. Topic areas covered include the following: Data Base/GPS Issues; Survey Management Issues; Surveying computations; Surveying education; Digital mapping; global change, EOS and NALC issues; GPS issues; Battelle Research in Remote Sensing and in GIS; Advanced Image Processing;GIS Issues; Surveying and Geodesy Issues; water resource issues; Advanced applications of remote sensing; Landsat Pathfinder I.
Interfacing geographic information systems and remote sensing for rural land-use analysis
NASA Technical Reports Server (NTRS)
Nellis, M. Duane; Lulla, Kamlesh; Jensen, John
1990-01-01
Recent advances in computer-based geographic information systems (GISs) are briefly reviewed, with an emphasis on the incorporation of remote-sensing data in GISs for rural applications. Topics addressed include sampling procedures for rural land-use analyses; GIS-based mapping of agricultural land use and productivity; remote sensing of land use and agricultural, forest, rangeland, and water resources; monitoring the dynamics of irrigation agriculture; GIS methods for detecting changes in land use over time; and the development of land-use modeling strategies.
Mehmood, Irfan; Sajjad, Muhammad; Baik, Sung Wook
2014-01-01
Wireless capsule endoscopy (WCE) has great advantages over traditional endoscopy because it is portable and easy to use, especially in remote monitoring health-services. However, during the WCE process, the large amount of captured video data demands a significant deal of computation to analyze and retrieve informative video frames. In order to facilitate efficient WCE data collection and browsing task, we present a resource- and bandwidth-aware WCE video summarization framework that extracts the representative keyframes of the WCE video contents by removing redundant and non-informative frames. For redundancy elimination, we use Jeffrey-divergence between color histograms and inter-frame Boolean series-based correlation of color channels. To remove non-informative frames, multi-fractal texture features are extracted to assist the classification using an ensemble-based classifier. Owing to the limited WCE resources, it is impossible for the WCE system to perform computationally intensive video summarization tasks. To resolve computational challenges, mobile-cloud architecture is incorporated, which provides resizable computing capacities by adaptively offloading video summarization tasks between the client and the cloud server. The qualitative and quantitative results are encouraging and show that the proposed framework saves information transmission cost and bandwidth, as well as the valuable time of data analysts in browsing remote sensing data. PMID:25225874
Cloud CPFP: a shotgun proteomics data analysis pipeline using cloud and high performance computing.
Trudgian, David C; Mirzaei, Hamid
2012-12-07
We have extended the functionality of the Central Proteomics Facilities Pipeline (CPFP) to allow use of remote cloud and high performance computing (HPC) resources for shotgun proteomics data processing. CPFP has been modified to include modular local and remote scheduling for data processing jobs. The pipeline can now be run on a single PC or server, a local cluster, a remote HPC cluster, and/or the Amazon Web Services (AWS) cloud. We provide public images that allow easy deployment of CPFP in its entirety in the AWS cloud. This significantly reduces the effort necessary to use the software, and allows proteomics laboratories to pay for compute time ad hoc, rather than obtaining and maintaining expensive local server clusters. Alternatively the Amazon cloud can be used to increase the throughput of a local installation of CPFP as necessary. We demonstrate that cloud CPFP allows users to process data at higher speed than local installations but with similar cost and lower staff requirements. In addition to the computational improvements, the web interface to CPFP is simplified, and other functionalities are enhanced. The software is under active development at two leading institutions and continues to be released under an open-source license at http://cpfp.sourceforge.net.
Mehmood, Irfan; Sajjad, Muhammad; Baik, Sung Wook
2014-09-15
Wireless capsule endoscopy (WCE) has great advantages over traditional endoscopy because it is portable and easy to use, especially in remote monitoring health-services. However, during the WCE process, the large amount of captured video data demands a significant deal of computation to analyze and retrieve informative video frames. In order to facilitate efficient WCE data collection and browsing task, we present a resource- and bandwidth-aware WCE video summarization framework that extracts the representative keyframes of the WCE video contents by removing redundant and non-informative frames. For redundancy elimination, we use Jeffrey-divergence between color histograms and inter-frame Boolean series-based correlation of color channels. To remove non-informative frames, multi-fractal texture features are extracted to assist the classification using an ensemble-based classifier. Owing to the limited WCE resources, it is impossible for the WCE system to perform computationally intensive video summarization tasks. To resolve computational challenges, mobile-cloud architecture is incorporated, which provides resizable computing capacities by adaptively offloading video summarization tasks between the client and the cloud server. The qualitative and quantitative results are encouraging and show that the proposed framework saves information transmission cost and bandwidth, as well as the valuable time of data analysts in browsing remote sensing data.
2014-06-12
interferometry and polarimetry . In the paper, the model was used to simulate SAR data for Mangrove (tropical) and Nezer (temperate) forests for P-band and...Scattering Model Applied to Radiometry, Interferometry, and Polarimetry at P- and L-Band. IEEE Transactions on Geoscience and Remote Sensing 44(4): 849
Jonathan P. Dandois; Erle C. Ellis
2013-01-01
High spatial resolution three-dimensional (3D) measurements of vegetation by remote sensing are advancing ecological research and environmental management. However, substantial economic and logistical costs limit this application, especially for observing phenological dynamics in ecosystem structure and spectral traits. Here we demonstrate a new aerial remote sensing...
National remote computational flight research facility
NASA Technical Reports Server (NTRS)
Rediess, Herman A.
1989-01-01
The extension of the NASA Ames-Dryden remotely augmented vehicle (RAV) facility to accommodate flight testing of a hypersonic aircraft utilizing the continental United States as a test range is investigated. The development and demonstration of an automated flight test management system (ATMS) that uses expert system technology for flight test planning, scheduling, and execution is documented.
Remote Sensing: The View from Above. Know Your Environment.
ERIC Educational Resources Information Center
Academy of Natural Sciences, Philadelphia, PA.
This publication identifies some of the general concepts of remote sensing and explains the image collection process and computer-generated reconstruction of the data. Monitoring the ecological collapse in coral reefs, weather phenomena like El Nino/La Nina, and U.S. Space Shuttle-based sensing projects are some of the areas for which remote…
On the Integration of Remote Experimentation into Undergraduate Laboratories--Pedagogical Approach
ERIC Educational Resources Information Center
Esche, Sven K.
2005-01-01
This paper presents an Internet-based open approach to laboratory instruction. In this article, the author talks about an open laboratory approach using a multi-user multi-device remote facility. This approach involves both the direct contact with the computer-controlled laboratory setup of interest with the students present in the laboratory…
NASA Astrophysics Data System (ADS)
Krause, H. F.; Deveney, E. F.; Jones, N. L.; Vane, C. R.; Datz, S.; Knudsen, H.; Grafström, P.; Schuch, R.
1997-04-01
Recent atomic physics studies involving ultrarelativistic Pb ions required solid target positioners, scintillators, and a sophisticated data acquisition and control system placed in a remote location at the CERN Super Proton Synchrotron near Geneva, Switzerland. The apparatus, installed in a high-radiation zone underground, had to (i) function for months, (ii) automatically respond to failures such as power outages and particle-induced computer upsets, and (iii) communicate with the outside world via a telephone line. The heart of the apparatus developed was an Apple Macintosh-based CAMAC system that answered the telephone and interpreted and executed remote control commands that (i) sensed and set targets, (ii) controlled voltages and discriminator levels for scintillators, (iii) modified data acquisition hardware logic, (iv) reported control information, and (v) automatically synchronized data acquisition to the CERN spill cycle via a modem signal and transmitted experimental data to a remote computer. No problems were experienced using intercontinental telephone connections at 1200 baud. Our successful "virtual laboratory" approach that uses off-the-shelf electronics is generally adaptable to more conventional bench-type experiments.
Remote Data Retrieval for Bioinformatics Applications: An Agent Migration Approach
Gao, Lei; Dai, Hua; Zhang, Tong-Liang; Chou, Kuo-Chen
2011-01-01
Some of the approaches have been developed to retrieve data automatically from one or multiple remote biological data sources. However, most of them require researchers to remain online and wait for returned results. The latter not only requires highly available network connection, but also may cause the network overload. Moreover, so far none of the existing approaches has been designed to address the following problems when retrieving the remote data in a mobile network environment: (1) the resources of mobile devices are limited; (2) network connection is relatively of low quality; and (3) mobile users are not always online. To address the aforementioned problems, we integrate an agent migration approach with a multi-agent system to overcome the high latency or limited bandwidth problem by moving their computations to the required resources or services. More importantly, the approach is fit for the mobile computing environments. Presented in this paper are also the system architecture, the migration strategy, as well as the security authentication of agent migration. As a demonstration, the remote data retrieval from GenBank was used to illustrate the feasibility of the proposed approach. PMID:21701677
Rapid cable tension estimation using dynamic and mechanical properties
NASA Astrophysics Data System (ADS)
Martínez-Castro, Rosana E.; Jang, Shinae; Christenson, Richard E.
2016-04-01
Main tension elements are critical to the overall stability of cable-supported bridges. A dependable and rapid determination of cable tension is desired to assess the state of a cable-supported bridge and evaluate its operability. A portable smart sensor setup is presented to reduce post-processing time and deployment complexity while reliably determining cable tension using dynamic characteristics extracted from spectral analysis. A self-recording accelerometer is coupled with a single-board microcomputer that communicates wirelessly with a remote host computer. The portable smart sensing device is designed such that additional algorithms, sensors and controlling devices for various monitoring applications can be installed and operated for additional structural assessment. The tension-estimating algorithms are based on taut string theory and expand to consider bending stiffness. The successful combination of cable properties allows the use of a cable's dynamic behavior to determine tension force. The tension-estimating algorithms are experimentally validated on a through-arch steel bridge subject to ambient vibration induced by passing traffic. The tension estimation is determined in well agreement with previously determined tension values for the structure.
Code 672 observational science branch computer networks
NASA Technical Reports Server (NTRS)
Hancock, D. W.; Shirk, H. G.
1988-01-01
In general, networking increases productivity due to the speed of transmission, easy access to remote computers, ability to share files, and increased availability of peripherals. Two different networks within the Observational Science Branch are described in detail.
Administering an epoch initiated for remote memory access
Blocksome, Michael A; Miller, Douglas R
2014-03-18
Methods, systems, and products are disclosed for administering an epoch initiated for remote memory access that include: initiating, by an origin application messaging module on an origin compute node, one or more data transfers to a target compute node for the epoch; initiating, by the origin application messaging module after initiating the data transfers, a closing stage for the epoch, including rejecting any new data transfers after initiating the closing stage for the epoch; determining, by the origin application messaging module, whether the data transfers have completed; and closing, by the origin application messaging module, the epoch if the data transfers have completed.
Calculations of atmospheric refraction for spacecraft remote-sensing applications
NASA Technical Reports Server (NTRS)
Chu, W. P.
1983-01-01
Analytical solutions to the refraction integrals appropriate for ray trajectories along slant paths through the atmosphere are derived in this paper. This type of geometry is commonly encountered in remote-sensing applications utilizing an occultation technique. The solutions are obtained by evaluating higher-order terms from expansion of the refraction integral and are dependent on the vertical temperature distributions. Refraction parameters such as total refraction angles, air masses, and path lengths can be accurately computed. It is also shown that the method can be used for computing refraction parameters in astronomical refraction geometry for large zenith angles.
Administering an epoch initiated for remote memory access
Blocksome, Michael A; Miller, Douglas R
2012-10-23
Methods, systems, and products are disclosed for administering an epoch initiated for remote memory access that include: initiating, by an origin application messaging module on an origin compute node, one or more data transfers to a target compute node for the epoch; initiating, by the origin application messaging module after initiating the data transfers, a closing stage for the epoch, including rejecting any new data transfers after initiating the closing stage for the epoch; determining, by the origin application messaging module, whether the data transfers have completed; and closing, by the origin application messaging module, the epoch if the data transfers have completed.
Administering an epoch initiated for remote memory access
Blocksome, Michael A.; Miller, Douglas R.
2013-01-01
Methods, systems, and products are disclosed for administering an epoch initiated for remote memory access that include: initiating, by an origin application messaging module on an origin compute node, one or more data transfers to a target compute node for the epoch; initiating, by the origin application messaging module after initiating the data transfers, a closing stage for the epoch, including rejecting any new data transfers after initiating the closing stage for the epoch; determining, by the origin application messaging module, whether the data transfers have completed; and closing, by the origin application messaging module, the epoch if the data transfers have completed.
NASA Technical Reports Server (NTRS)
Ustinov, E. A.
1999-01-01
Evaluation of weighting functions in the atmospheric remote sensing is usually the most computer-intensive part of the inversion algorithms. We present an analytic approach to computations of temperature and mixing ratio weighting functions that is based on our previous results but the resulting expressions use the intermediate variables that are generated in computations of observable radiances themselves. Upwelling radiances at the given level in the atmosphere and atmospheric transmittances from space to the given level are combined with local values of the total absorption coefficient and its components due to absorption of atmospheric constituents under study. This makes it possible to evaluate the temperature and mixing ratio weighting functions in parallel with evaluation of radiances. This substantially decreases the computer time required for evaluation of weighting functions. Implications for the nadir and limb viewing geometries are discussed.
Using a Cray Y-MP as an array processor for a RISC Workstation
NASA Technical Reports Server (NTRS)
Lamaster, Hugh; Rogallo, Sarah J.
1992-01-01
As microprocessors increase in power, the economics of centralized computing has changed dramatically. At the beginning of the 1980's, mainframes and super computers were often considered to be cost-effective machines for scalar computing. Today, microprocessor-based RISC (reduced-instruction-set computer) systems have displaced many uses of mainframes and supercomputers. Supercomputers are still cost competitive when processing jobs that require both large memory size and high memory bandwidth. One such application is array processing. Certain numerical operations are appropriate to use in a Remote Procedure Call (RPC)-based environment. Matrix multiplication is an example of an operation that can have a sufficient number of arithmetic operations to amortize the cost of an RPC call. An experiment which demonstrates that matrix multiplication can be executed remotely on a large system to speed the execution over that experienced on a workstation is described.
DBSecSys: a database of Burkholderia mallei secretion systems.
Memišević, Vesna; Kumar, Kamal; Cheng, Li; Zavaljevski, Nela; DeShazer, David; Wallqvist, Anders; Reifman, Jaques
2014-07-16
Bacterial pathogenicity represents a major public health concern worldwide. Secretion systems are a key component of bacterial pathogenicity, as they provide the means for bacterial proteins to penetrate host-cell membranes and insert themselves directly into the host cells' cytosol. Burkholderia mallei is a Gram-negative bacterium that uses multiple secretion systems during its host infection life cycle. To date, the identities of secretion system proteins for B. mallei are not well known, and their pathogenic mechanisms of action and host factors are largely uncharacterized. We present the Database of Burkholderia malleiSecretion Systems (DBSecSys), a compilation of manually curated and computationally predicted bacterial secretion system proteins and their host factors. Currently, DBSecSys contains comprehensive experimentally and computationally derived information about B. mallei strain ATCC 23344. The database includes 143 B. mallei proteins associated with five secretion systems, their 1,635 human and murine interacting targets, and the corresponding 2,400 host-B. mallei interactions. The database also includes information about 10 pathogenic mechanisms of action for B. mallei secretion system proteins inferred from the available literature. Additionally, DBSecSys provides details about 42 virulence attenuation experiments for 27 B. mallei secretion system proteins. Users interact with DBSecSys through a Web interface that allows for data browsing, querying, visualizing, and downloading. DBSecSys provides a comprehensive, systematically organized resource of experimental and computational data associated with B. mallei secretion systems. It provides the unique ability to study secretion systems not only through characterization of their corresponding pathogen proteins, but also through characterization of their host-interacting partners.The database is available at https://applications.bhsai.org/dbsecsys.
Cotes-Ruiz, Iván Tomás; Prado, Rocío P.; García-Galán, Sebastián; Muñoz-Expósito, José Enrique; Ruiz-Reyes, Nicolás
2017-01-01
Nowadays, the growing computational capabilities of Cloud systems rely on the reduction of the consumed power of their data centers to make them sustainable and economically profitable. The efficient management of computing resources is at the heart of any energy-aware data center and of special relevance is the adaptation of its performance to workload. Intensive computing applications in diverse areas of science generate complex workload called workflows, whose successful management in terms of energy saving is still at its beginning. WorkflowSim is currently one of the most advanced simulators for research on workflows processing, offering advanced features such as task clustering and failure policies. In this work, an expected power-aware extension of WorkflowSim is presented. This new tool integrates a power model based on a computing-plus-communication design to allow the optimization of new management strategies in energy saving considering computing, reconfiguration and networks costs as well as quality of service, and it incorporates the preeminent strategy for on host energy saving: Dynamic Voltage Frequency Scaling (DVFS). The simulator is designed to be consistent in different real scenarios and to include a wide repertory of DVFS governors. Results showing the validity of the simulator in terms of resources utilization, frequency and voltage scaling, power, energy and time saving are presented. Also, results achieved by the intra-host DVFS strategy with different governors are compared to those of the data center using a recent and successful DVFS-based inter-host scheduling strategy as overlapped mechanism to the DVFS intra-host technique. PMID:28085932
Cotes-Ruiz, Iván Tomás; Prado, Rocío P; García-Galán, Sebastián; Muñoz-Expósito, José Enrique; Ruiz-Reyes, Nicolás
2017-01-01
Nowadays, the growing computational capabilities of Cloud systems rely on the reduction of the consumed power of their data centers to make them sustainable and economically profitable. The efficient management of computing resources is at the heart of any energy-aware data center and of special relevance is the adaptation of its performance to workload. Intensive computing applications in diverse areas of science generate complex workload called workflows, whose successful management in terms of energy saving is still at its beginning. WorkflowSim is currently one of the most advanced simulators for research on workflows processing, offering advanced features such as task clustering and failure policies. In this work, an expected power-aware extension of WorkflowSim is presented. This new tool integrates a power model based on a computing-plus-communication design to allow the optimization of new management strategies in energy saving considering computing, reconfiguration and networks costs as well as quality of service, and it incorporates the preeminent strategy for on host energy saving: Dynamic Voltage Frequency Scaling (DVFS). The simulator is designed to be consistent in different real scenarios and to include a wide repertory of DVFS governors. Results showing the validity of the simulator in terms of resources utilization, frequency and voltage scaling, power, energy and time saving are presented. Also, results achieved by the intra-host DVFS strategy with different governors are compared to those of the data center using a recent and successful DVFS-based inter-host scheduling strategy as overlapped mechanism to the DVFS intra-host technique.
The Ethics of Cloud Computing.
de Bruin, Boudewijn; Floridi, Luciano
2017-02-01
Cloud computing is rapidly gaining traction in business. It offers businesses online services on demand (such as Gmail, iCloud and Salesforce) and allows them to cut costs on hardware and IT support. This is the first paper in business ethics dealing with this new technology. It analyzes the informational duties of hosting companies that own and operate cloud computing datacentres (e.g., Amazon). It considers the cloud services providers leasing 'space in the cloud' from hosting companies (e.g., Dropbox, Salesforce). And it examines the business and private 'clouders' using these services. The first part of the paper argues that hosting companies, services providers and clouders have mutual informational (epistemic) obligations to provide and seek information about relevant issues such as consumer privacy, reliability of services, data mining and data ownership. The concept of interlucency is developed as an epistemic virtue governing ethically effective communication. The second part considers potential forms of government restrictions on or proscriptions against the development and use of cloud computing technology. Referring to the concept of technology neutrality, it argues that interference with hosting companies and cloud services providers is hardly ever necessary or justified. It is argued, too, however, that businesses using cloud services (e.g., banks, law firms, hospitals etc. storing client data in the cloud) will have to follow rather more stringent regulations.
Grid Computing in K-12 Schools. Soapbox Digest. Volume 3, Number 2, Fall 2004
ERIC Educational Resources Information Center
AEL, 2004
2004-01-01
Grid computing allows large groups of computers (either in a lab, or remote and connected only by the Internet) to extend extra processing power to each individual computer to work on components of a complex request. Grid middleware, recognizing priorities set by systems administrators, allows the grid to identify and use this power without…
NASA Technical Reports Server (NTRS)
Jacob, Joseph; Katz, Daniel; Prince, Thomas; Berriman, Graham; Good, John; Laity, Anastasia
2006-01-01
The final version (3.0) of the Montage software has been released. To recapitulate from previous NASA Tech Briefs articles about Montage: This software generates custom, science-grade mosaics of astronomical images on demand from input files that comply with the Flexible Image Transport System (FITS) standard and contain image data registered on projections that comply with the World Coordinate System (WCS) standards. This software can be executed on single-processor computers, multi-processor computers, and such networks of geographically dispersed computers as the National Science Foundation s TeraGrid or NASA s Information Power Grid. The primary advantage of running Montage in a grid environment is that computations can be done on a remote supercomputer for efficiency. Multiple computers at different sites can be used for different parts of a computation a significant advantage in cases of computations for large mosaics that demand more processor time than is available at any one site. Version 3.0 incorporates several improvements over prior versions. The most significant improvement is that this version is accessible to scientists located anywhere, through operational Web services that provide access to data from several large astronomical surveys and construct mosaics on either local workstations or remote computational grids as needed.
NASA Technical Reports Server (NTRS)
Birnie, R. W.; Stoiber, R. E. (Principal Investigator)
1983-01-01
Computer classification of LANDSAT data was used for forest type mapping in New England. The ability to classify areas of hardwood, softwood, and mixed tree types was assessed along with determining clearcut regions and gypsy moth defoliation. Applications of the information to forest management and locating potential deer yards were investigated. The principal activities concerned with remote sensing of volcanic emissions centered around the development of remote sensors for SO2 and HCl gas, and their use at appropriate volcanic sites. Two major areas were investigated (Masaya, Nicaragua, and St. Helens, Washington) along with several minor ones.
NASA Astrophysics Data System (ADS)
Lv, Shu-Xin; Zhao, Zheng-Wei; Zhou, Ping
2018-01-01
We present a scheme for joint remote implementation of an arbitrary single-qubit operation following some ideas in one-way quantum computation. All the senders share the information of implemented quantum operation and perform corresponding single-qubit measurements according to their information of implemented operation. An arbitrary single-qubit operation can be implemented upon the remote receiver's quantum system if the receiver cooperates with all the senders. Moreover, we study the protocol of multiparty joint remote implementation of an arbitrary single-qubit operation with many senders by using a multiparticle entangled state as the quantum channel.
Shipping Science Worldwide with Open Source Containers
NASA Astrophysics Data System (ADS)
Molineaux, J. P.; McLaughlin, B. D.; Pilone, D.; Plofchan, P. G.; Murphy, K. J.
2014-12-01
Scientific applications often present difficult web-hosting needs. Their compute- and data-intensive nature, as well as an increasing need for high-availability and distribution, combine to create a challenging set of hosting requirements. In the past year, advancements in container-based virtualization and related tooling have offered new lightweight and flexible ways to accommodate diverse applications with all the isolation and portability benefits of traditional virtualization. This session will introduce and demonstrate an open-source, single-interface, Platform-as-a-Serivce (PaaS) that empowers application developers to seamlessly leverage geographically distributed, public and private compute resources to achieve highly-available, performant hosting for scientific applications.
DOE Office of Scientific and Technical Information (OSTI.GOV)
McDermott, Jason E.; Braun, Pascal; Bonneau, Richard A.
Pathogenic infections are a major cause of both human disease and loss of crop yields and animal stocks and thus cause immense damage to the worldwide economy. The significance of infectious diseases is expected to increase in an ever more connected warming world, in which new viral, bacterial and fungal pathogens can find novel hosts and ecologic niches. At the same time, the complex and sophisticated mechanisms by which diverse pathogenic agents evade defense mechanisms and subvert their hosts networks to suit their lifestyle needs is still very incompletely understood especially from a systems perspective [1]. Thus, understanding host-pathogen interactionsmore » is both an important and a scientifically fascinating topic. Recently, technology has offered the opportunity to investigate host-pathogen interactions on a level of detail and scope that offers immense computational and analytical possibilities. Genome sequencing was pioneered on some of these pathogens, and the number of strains and variants of pathogens sequenced to date vastly outnumbers the number of host genomes available. At the same time, for both plant and human hosts more and more data on population level genomic variation becomes available and offers a rich field for analysis into the genetic interactions between host and pathogen.« less
Dynamic provisioning of local and remote compute resources with OpenStack
NASA Astrophysics Data System (ADS)
Giffels, M.; Hauth, T.; Polgart, F.; Quast, G.
2015-12-01
Modern high-energy physics experiments rely on the extensive usage of computing resources, both for the reconstruction of measured events as well as for Monte-Carlo simulation. The Institut fur Experimentelle Kernphysik (EKP) at KIT is participating in both the CMS and Belle experiments with computing and storage resources. In the upcoming years, these requirements are expected to increase due to growing amount of recorded data and the rise in complexity of the simulated events. It is therefore essential to increase the available computing capabilities by tapping into all resource pools. At the EKP institute, powerful desktop machines are available to users. Due to the multi-core nature of modern CPUs, vast amounts of CPU time are not utilized by common desktop usage patterns. Other important providers of compute capabilities are classical HPC data centers at universities or national research centers. Due to the shared nature of these installations, the standardized software stack required by HEP applications cannot be installed. A viable way to overcome this constraint and offer a standardized software environment in a transparent manner is the usage of virtualization technologies. The OpenStack project has become a widely adopted solution to virtualize hardware and offer additional services like storage and virtual machine management. This contribution will report on the incorporation of the institute's desktop machines into a private OpenStack Cloud. The additional compute resources provisioned via the virtual machines have been used for Monte-Carlo simulation and data analysis. Furthermore, a concept to integrate shared, remote HPC centers into regular HEP job workflows will be presented. In this approach, local and remote resources are merged to form a uniform, virtual compute cluster with a single point-of-entry for the user. Evaluations of the performance and stability of this setup and operational experiences will be discussed.
NASA Astrophysics Data System (ADS)
Budiardja, R. D.; Lingerfelt, E. J.; Guidry, M. W.
2003-05-01
Wireless technology implemented with handheld devices has attractive features because of the potential to access large amounts of data and the prospect of on-the-fly computational analysis from a device that can be carried in a shirt pocket. We shall describe applications of such technology to the general paradigm of making digital wireless connections from the field to upload information and queries to network servers, executing (potentially complex) programs and controlling data analysis and/or database operations on fast network computers, and returning real-time information from this analysis to the handheld device in the field. As illustration, we shall describe several client/server programs that we have written for applications in teaching introductory astronomy. For example, one program allows static and dynamic properties of astronomical objects to be accessed in a remote observation laboratory setting using a digital cell phone or PDA. Another implements interactive quizzing over a cell phone or PDA using a 700-question introductory astronomy quiz database, thus permitting students to study for astronomy quizzes in any environment in which they have a few free minutes and a digital cell phone or wireless PDA. Another allows one to control and monitor a computation done on a Beowulf cluster by changing the parameters of the computation remotely and retrieving the result when the computation is done. The presentation will include hands-on demonstrations with real devices. *Managed by UT-Battelle, LLC, for the U.S. Department of Energy under contract DE-AC05-00OR22725.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stepanov, Vyacheslav E.; Potapov, Victor N.; Smirnov, Sergey V.
Decontamination and decommissioning of the research reactors MR (Testing Reactor) and RFT (Reactor of Physics and Technology) has recently been initiated in the National Research Center (NRC) 'Kurchatov institute', Moscow. In the building, neighboring to the reactor, the storage of HLRW is located. The storage is made of monolithic concrete in which steel cells depth 4 m are located. In cells of storage the HLRW packed into cases are placed. These the radioactive waste are also subject to export on long storage in the specialized organization. For characterization of the radioactive waste in cases the remote-controlled collimated spectrometer system wasmore » used. The system consists of a spectrometric collimated gamma-ray detector, a color video camera and a control unit, mounted on a rotator, which are mounted on a tripod with the host computer. For determination of specific activity of radionuclides in cases, it is developed programs of calculation of coefficients of proportionality of specific activity to the corresponding speeds of the account in peaks of full absorption at single specific activity of radionuclides in cases. For determination of these coefficients the mathematical model of spectrometer system based on the Monte-Carlo method was used. Dependences of calibration coefficients for various radionuclides from distance between the detector and a case at various values of the radioactive waste density in cases are given. Measurements of specific activity in cases are taken and are discussed. By results of measurements decisions on the appeal of the radioactive waste being in cases are made. (authors)« less
Basic BASIC; An Introduction to Computer Programming in BASIC Language.
ERIC Educational Resources Information Center
Coan, James S.
With the increasing availability of computer access through remote terminals and time sharing, more and more schools and colleges are able to introduce programing to substantial numbers of students. This book is an attempt to incorporate computer programming, using BASIC language, and the teaching of mathematics. The general approach of the book…
ERIC Educational Resources Information Center
Dillon, Martin; And Others
The Online Computer Library Center Internet Resource project focused on the nature of electronic textual information available through remote access using the Internet and the problems associated with creating machine-readable cataloging (MARC) records for these objects using current USMARC format for computer files and "Anglo-American…
The WRITEACOURSE Language: Programming Manual. Revised Edition.
ERIC Educational Resources Information Center
Zosel, Mary; And Others
WRITEACOURSE is a programing language for man-computer interactions. It was originally designed for writing computer assisted instruction courses, but it can also be used to control a remote terminal in a variety of applications which involve display and editing of characters. It is not suited for applications which use the computer as an…
A Wireless Monitoring Sub-nA Resolution Test Platform for Nanostructure Sensors
Jang, Chi Woong; Byun, Young Tae; Lee, Taikjin; Woo, Deok Ha; Lee, Seok; Jhon, Young Min
2013-01-01
We have constructed a wireless monitoring test platform with a sub-nA resolution signal amplification/processing circuit (SAPC) and a wireless communication network to test the real-time remote monitoring of the signals from carbon nanotube (CNT) sensors. The operation characteristics of the CNT sensors can also be measured by the ISD-VSD curve with the SAPC. The SAPC signals are transmitted to a personal computer by Bluetooth communication and the signals from the computer are transmitted to smart phones by Wi-Fi communication, in such a way that the signals from the sensors can be remotely monitored through a web browser. Successful remote monitoring of signals from a CNT sensor was achieved with the wireless monitoring test platform for detection of 0.15% methanol vapor with 0.5 nA resolution and 7 Hz sampling rate. PMID:23783735
Lustre Distributed Name Space (DNE) Evaluation at the Oak Ridge Leadership Computing Facility (OLCF)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simmons, James S.; Leverman, Dustin B.; Hanley, Jesse A.
This document describes the Lustre Distributed Name Space (DNE) evaluation carried at the Oak Ridge Leadership Computing Facility (OLCF) between 2014 and 2015. DNE is a development project funded by the OpenSFS, to improve Lustre metadata performance and scalability. The development effort has been split into two parts, the first part (DNE P1) providing support for remote directories over remote Lustre Metadata Server (MDS) nodes and Metadata Target (MDT) devices, while the second phase (DNE P2) addressed split directories over multiple remote MDS nodes and MDT devices. The OLCF have been actively evaluating the performance, reliability, and the functionality ofmore » both DNE phases. For these tests, internal OLCF testbed were used. Results are promising and OLCF is planning on a full DNE deployment by mid-2016 timeframe on production systems.« less
Semiautonomous teleoperation system with vision guidance
NASA Astrophysics Data System (ADS)
Yu, Wai; Pretlove, John R. G.
1998-12-01
This paper describes the ongoing research work on developing a telerobotic system in Mechatronic Systems and Robotics Research group at the University of Surrey. As human operators' manual control of remote robots always suffer from reduced performance and difficulties in perceiving information from the remote site, a system with a certain level of intelligence and autonomy will help to solve some of these problems. Thus, this system has been developed for this purpose. It also serves as an experimental platform to test the idea of using the combination of human and computer intelligence in teleoperation and finding out the optimum balance between them. The system consists of a Polhemus- based input device, a computer vision sub-system and a graphical user interface which communicates the operator with the remote robot. The system description is given in this paper as well as the preliminary experimental results of the system evaluation.
Increasing Access and Usability of Remote Sensing Data: The NASA Protected Area Archive
NASA Technical Reports Server (NTRS)
Geller, Gary N.
2004-01-01
Although remote sensing data are now widely available, much of it at low or no-cost, many managers of protected conservation areas do not have the expertise or tools to view or analyze it. Thus access to it by the protected area management community is effectively blocked. The Protected Area Archive will increase access to remote sensing data by creating collections of satellite images of protected areas and packaging them with simple-to-use visualization and analytical tools. The user can easily locate the area and image of interest on a map, then display, roam, and zoom the image. A set of simple tools will be provided so the user can explore the data and employ it to assist in management and monitoring of their area. The 'Phase 1 ' version requires only a Windows-based computer and basic computer skills, and may be of particular help to protected area managers in developing countries.
Distributive, Non-destructive Real-time System and Method for Snowpack Monitoring
NASA Technical Reports Server (NTRS)
Frolik, Jeff (Inventor); Skalka, Christian (Inventor)
2013-01-01
A ground-based system that provides quasi real-time measurement and collection of snow-water equivalent (SWE) data in remote settings is provided. The disclosed invention is significantly less expensive and easier to deploy than current methods and less susceptible to terrain and snow bridging effects. Embodiments of the invention include remote data recovery solutions. Compared to current infrastructure using existing SWE technology, the disclosed invention allows more SWE sites to be installed for similar cost and effort, in a greater variety of terrain; thus, enabling data collection at improved spatial resolutions. The invention integrates a novel computational architecture with new sensor technologies. The invention's computational architecture is based on wireless sensor networks, comprised of programmable, low-cost, low-powered nodes capable of sophisticated sensor control and remote data communication. The invention also includes measuring attenuation of electromagnetic radiation, an approach that is immune to snow bridging and significantly reduces sensor footprints.
NASA Technical Reports Server (NTRS)
1974-01-01
The present work gathers together numerous papers describing the use of remote sensing technology for mapping, monitoring, and management of earth resources and man's environment. Studies using various types of sensing equipment are described, including multispectral scanners, radar imagery, spectrometers, lidar, and aerial photography, and both manual and computer-aided data processing techniques are described. Some of the topics covered include: estimation of population density in Tokyo districts from ERTS-1 data, a clustering algorithm for unsupervised crop classification, passive microwave sensing of moist soils, interactive computer processing for land use planning, the use of remote sensing to delineate floodplains, moisture detection from Skylab, scanning thermal plumes, electrically scanning microwave radiometers, oil slick detection by X-band synthetic aperture radar, and the use of space photos for search of oil and gas fields. Individual items are announced in this issue.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Adams, Leslie G.; Khare, Sangeeta; Lawhon, Sara D.
The aim of research on infectious diseases is their prevention, and brucellosis and salmonellosis as such are classic examples of worldwide zoonoses for application of a systems biology approach for enhanced rational vaccine development. When used optimally, vaccines prevent disease manifestations, reduce transmission of disease, decrease the need for pharmaceutical intervention, and improve the health and welfare of animals, as well as indirectly protecting against zoonotic diseases of people. Advances in the last decade or so using comprehensive systems biology approaches linking genomics, proteomics, bioinformatics, and biotechnology with immunology, pathogenesis and vaccine formulation and delivery are expected to enable enhancedmore » approaches to vaccine development. The goal of this paper is to evaluate the role of computational systems biology analysis of host:pathogen interactions (the interactome) as a tool for enhanced rational design of vaccines. Systems biology is bringing a new, more robust approach to veterinary vaccine design based upon a deeper understanding of the host pathogen interactions and its impact on the host's molecular network of the immune system. A computational systems biology method was utilized to create interactome models of the host responses to Brucella melitensis (BMEL), Mycobacterium avium paratuberculosis (MAP), Salmonella enterica Typhimurium (STM), and a Salmonella mutant (isogenic *sipA, sopABDE2) and linked to the basis for rational development of vaccines for brucellosis and salmonellosis as reviewed by Adams et al. and Ficht et al. [1,2]. A bovine ligated ileal loop biological model was established to capture the host gene expression response at multiple time points post infection. New methods based on Dynamic Bayesian Network (DBN) machine learning were employed to conduct a comparative pathogenicity analysis of 219 signaling and metabolic pathways and 1620 gene ontology (GO) categories that defined the host's biosignatures to each infectious condition. Through this DBN computational approach, the method identified significantly perturbed pathways and GO category groups of genes that define the pathogenicity signatures of the infectious agent. Our preliminary results provide deeper understanding of the overall complexity of host innate immune response as well as the identification of host gene perturbations that defines a unique host temporal biosignature response to each pathogen. The application of advanced computational methods for developing interactome models based on DBNs has proven to be instrumental in elucidating novel host responses and improved functional biological insight into the host defensive mechanisms. Evaluating the unique differences in pathway and GO perturbations across pathogen conditions allowed the identification of plausible host pathogen interaction mechanisms. Accordingly, a systems biology approach to study molecular pathway gene expression profiles of host cellular responses to microbial pathogens holds great promise as a methodology to identify, model and predict the overall dynamics of the host pathogen interactome. Thus, we propose that such an approach has immediate application to the rational design of brucellosis and salmonellosis vaccines.« less
Adams, L Garry; Khare, Sangeeta; Lawhon, Sara D; Rossetti, Carlos A; Lewin, Harris A; Lipton, Mary S; Turse, Joshua E; Wylie, Dennis C; Bai, Yu; Drake, Kenneth L
2011-09-22
The aim of research on infectious diseases is their prevention, and brucellosis and salmonellosis as such are classic examples of worldwide zoonoses for application of a systems biology approach for enhanced rational vaccine development. When used optimally, vaccines prevent disease manifestations, reduce transmission of disease, decrease the need for pharmaceutical intervention, and improve the health and welfare of animals, as well as indirectly protecting against zoonotic diseases of people. Advances in the last decade or so using comprehensive systems biology approaches linking genomics, proteomics, bioinformatics, and biotechnology with immunology, pathogenesis and vaccine formulation and delivery are expected to enable enhanced approaches to vaccine development. The goal of this paper is to evaluate the role of computational systems biology analysis of host:pathogen interactions (the interactome) as a tool for enhanced rational design of vaccines. Systems biology is bringing a new, more robust approach to veterinary vaccine design based upon a deeper understanding of the host-pathogen interactions and its impact on the host's molecular network of the immune system. A computational systems biology method was utilized to create interactome models of the host responses to Brucella melitensis (BMEL), Mycobacterium avium paratuberculosis (MAP), Salmonella enterica Typhimurium (STM), and a Salmonella mutant (isogenic ΔsipA, sopABDE2) and linked to the basis for rational development of vaccines for brucellosis and salmonellosis as reviewed by Adams et al. and Ficht et al. [1,2]. A bovine ligated ileal loop biological model was established to capture the host gene expression response at multiple time points post infection. New methods based on Dynamic Bayesian Network (DBN) machine learning were employed to conduct a comparative pathogenicity analysis of 219 signaling and metabolic pathways and 1620 gene ontology (GO) categories that defined the host's biosignatures to each infectious condition. Through this DBN computational approach, the method identified significantly perturbed pathways and GO category groups of genes that define the pathogenicity signatures of the infectious agent. Our preliminary results provide deeper understanding of the overall complexity of host innate immune response as well as the identification of host gene perturbations that defines a unique host temporal biosignature response to each pathogen. The application of advanced computational methods for developing interactome models based on DBNs has proven to be instrumental in elucidating novel host responses and improved functional biological insight into the host defensive mechanisms. Evaluating the unique differences in pathway and GO perturbations across pathogen conditions allowed the identification of plausible host-pathogen interaction mechanisms. Accordingly, a systems biology approach to study molecular pathway gene expression profiles of host cellular responses to microbial pathogens holds great promise as a methodology to identify, model and predict the overall dynamics of the host-pathogen interactome. Thus, we propose that such an approach has immediate application to the rational design of brucellosis and salmonellosis vaccines. Copyright © 2011 Elsevier Ltd. All rights reserved.
Toussaint, Sandy; Mak, Donna B
2010-01-01
In 2006 the Kimberley Remote Area Health Placement Program (hereinafter the 'Program') was established at the University of Notre Dame's School of Medicine (Fremantle campus, Western Australia). The Program was developed as one of the strategies to achieve the School of Medicine's mission to graduate knowledgeable, skillful, dutiful and ethical doctors who will want to work in Australian areas of unmet need. The Program aims to immerse medical students in non-clinical settings to provide them with opportunities to learn life skills required for remote area living, and to introduce them to the myriad of socio-cultural, geographic, climatic and economic factors that impact on the health and wellbeing of remote area residents. To meet these objectives, the School organizes for students to live with, and do useful non-clinical work for, a host community or organization for up to one week. In 2008 the Program was evaluated to explore and assess its immediate and potential future benefits and limitations as perceived by Kimberley residents. This paper reports on the evaluation's findings via Kimberley-based narratives and raises some issues that are essential to training and retaining a 'bloody good doctor...' in a remote Australian setting. Using a mix of qualitative, ethnographic methods, the Program was evaluated by an independent researcher during four weeks of field research in late 2008. The methods included a survey, structured and unstructured interviewing and participant observation to elicit data. Thirty-three formal interviews of at least one hour's duration were conducted. Data were also collected via 15 informal discussions. Both formal and conservational interactions occurred in a range of town-based and more remote settings. The majority of persons consulted generally highlighted the Program's benefits. The reasons for this positive evaluation varied, but a common thread was that exposure to the Kimberley introduced the students to local life, a quality that had the potential to result in a medical student returning as a qualified doctor. The Program was seen as beneficial because it provided a structured, constructive means for prospective doctors to appreciate the assets rather than the deficits of remote area living. Another positive implication was that the Program equipped future doctors (regardless of their eventual work location) to treat a person from the 'bush'. It also encouraged students to think and act cross-culturally. An important immediate benefit was that the Program offered human resource support to the host organization at a busy time of the year. In only three of the 33 formal interviews was a negative or ambivalent attitude toward the Program expressed. However, a common concern was the brevity of time students served with their host organization. The data collected revealed that Kimberley people believed that the Program held strong potential for the successful recruitment of doctors prepared to make a long-term commitment to the region. Never far from their minds, conversations and activities was the idea that the effort they put into accommodating, supporting and guiding the students was worth it because 'If only one good doctor comes from the Program, then that's a good thing. A good doctor would have a great impact - the implications are immeasurable...'.
Technology and the Modern Library.
ERIC Educational Resources Information Center
Boss, Richard W.
1984-01-01
Overview of the impact of information technology on libraries highlights turnkey vendors, bibliographic utilities, commercial suppliers of records, state and regional networks, computer-to-computer linkages, remote database searching, terminals and microcomputers, building local databases, delivery of information, digital telefacsimile,…
1994-03-01
Specification and Network Time Protocol(NTP) over the Implementation. RFC-o 119, Network OSI Remote Operations Service. RFC- Working Group, September...approximately ISIS implements a powerful model of 94% of the total computation time. distributed computation known as modelo Timing results are
The Radio Frequency Health Node Wireless Sensor System
NASA Technical Reports Server (NTRS)
Valencia, J. Emilio; Stanley, Priscilla C.; Mackey, Paul J.
2009-01-01
The Radio Frequency Health Node (RFHN) wireless sensor system differs from other wireless sensor systems in ways originally intended to enhance utility as an instrumentation system for a spacecraft. The RFHN can also be adapted to use in terrestrial applications in which there are requirements for operational flexibility and integrability into higher-level instrumentation and data acquisition systems. As shown in the figure, the heart of the system is the RFHN, which is a unit that passes commands and data between (1) one or more commercially available wireless sensor units (optionally, also including wired sensor units) and (2) command and data interfaces with a local control computer that may be part of the spacecraft or other engineering system in which the wireless sensor system is installed. In turn, the local control computer can be in radio or wire communication with a remote control computer that may be part of a higher-level system. The remote control computer, acting via the local control computer and the RFHN, cannot only monitor readout data from the sensor units but can also remotely configure (program or reprogram) the RFHN and the sensor units during operation. In a spacecraft application, the RFHN and the sensor units can also be configured more nearly directly, prior to launch, via a serial interface that includes an umbilical cable between the spacecraft and ground support equipment. In either case, the RFHN wireless sensor system has the flexibility to be configured, as required, with different numbers and types of sensors for different applications. The RFHN can be used to effect realtime transfer of data from, and commands to, the wireless sensor units. It can also store data for later retrieval by an external computer. The RFHN communicates with the wireless sensor units via a radio transceiver module. The modular design of the RFHN makes it possible to add radio transceiver modules as needed to accommodate additional sets of wireless sensor units. The RFHN includes a core module that performs generic computer functions, including management of power and input, output, processing, and storage of data. In a typical application, the processing capabilities in the RFHN are utilized to perform preprocessing, trending, and fusion of sensor data. The core module also serves as the unit through which the remote control computer configures the sensor units and the rest of the RFHN.
NASA Technical Reports Server (NTRS)
Bishop, Ann P.; Pinelli, Thomas E.
1994-01-01
This paper presents selected results from an empirical investigation into the use of computer networks in aerospace engineering. Such networks allow aerospace engineers to communicate with people and access remote resources through electronic mail, file transfer, and remote log-in. The study drew its subjects from private sector, government and academic organizations in the U.S. aerospace industry. Data presented here were gathered in a mail survey, conducted in Spring 1993, that was distributed to aerospace engineers performing a wide variety of jobs. Results from the mail survey provide a snapshot of the current use of computer networks in the aerospace industry, suggest factors associated with the use of networks, and identify perceived impacts of networks on aerospace engineering work and communication.
The Thirty Meter Telescope site testing robotic computer system
NASA Astrophysics Data System (ADS)
Riddle, Reed L.; Schöck, Matthias; Skidmore, Warren
2006-06-01
The Thirty Meter Telescope (TMT) project is currently testing six remote sites as candidates for the final location of the telescope. Each site has several instruments, including seeing monitors, weather stations, and turbulence profile measuring systems, each of which is computer controlled. As the sites are remote (usually hours from the nearest town), they requires a system that can control the operations of all the varied subsystems, keep the systems safe from damage and recover from errors during operation. The robotic system must also be robust enough to operate without human intervention and when internet connections are lost. It is also critical that a data archiving system diligently records all data as gathered. This paper is a discussion of the TMT site testing robotic computer system as implemented.
Towards ubiquitous access of computer-assisted surgery systems.
Liu, Hui; Lufei, Hanping; Shi, Weishong; Chaudhary, Vipin
2006-01-01
Traditional stand-alone computer-assisted surgery (CAS) systems impede the ubiquitous and simultaneous access by multiple users. With advances in computing and networking technologies, ubiquitous access to CAS systems becomes possible and promising. Based on our preliminary work, CASMIL, a stand-alone CAS server developed at Wayne State University, we propose a novel mobile CAS system, UbiCAS, which allows surgeons to retrieve, review and interpret multimodal medical images, and to perform some critical neurosurgical procedures on heterogeneous devices from anywhere at anytime. Furthermore, various optimization techniques, including caching, prefetching, pseudo-streaming-model, and compression, are used to guarantee the QoS of the UbiCAS system. UbiCAS enables doctors at remote locations to actively participate remote surgeries, share patient information in real time before, during, and after the surgery.
Idle waves in high-performance computing
NASA Astrophysics Data System (ADS)
Markidis, Stefano; Vencels, Juris; Peng, Ivy Bo; Akhmetova, Dana; Laure, Erwin; Henri, Pierre
2015-01-01
The vast majority of parallel scientific applications distributes computation among processes that are in a busy state when computing and in an idle state when waiting for information from other processes. We identify the propagation of idle waves through processes in scientific applications with a local information exchange between the two processes. Idle waves are nondispersive and have a phase velocity inversely proportional to the average busy time. The physical mechanism enabling the propagation of idle waves is the local synchronization between two processes due to remote data dependency. This study provides a description of the large number of processes in parallel scientific applications as a continuous medium. This work also is a step towards an understanding of how localized idle periods can affect remote processes, leading to the degradation of global performance in parallel scientific applications.
In Galápagos … and Uncomfortable with Evolution
ERIC Educational Resources Information Center
Cotner, Sehoya; Graczyk, Hannah; Rodríguez Garcia, José Luis; Moore, Randy
2016-01-01
In June 2013, the third World Evolution Summit convened on San Cristóbal, hosting scientists from around the world (Paz-y-Miño-C and Espinosa 2013)--neither the first nor likely the last gathering of biologists on these remote islands. Clearly, both locals and an international audience perceive Galápagos as figuring prominently in discourse about…
Flexible feature interface for multimedia sources
Coffland, Douglas R [Livermore, CA
2009-06-09
A flexible feature interface for multimedia sources system that includes a single interface for the addition of features and functions to multimedia sources and for accessing those features and functions from remote hosts. The interface utilizes the export statement: export "C" D11Export void FunctionName(int argc, char ** argv,char * result, SecureSession *ctrl) or the binary equivalent of the export statement.
Creating a Responsive Website: PPLD Friends--A Case Study
ERIC Educational Resources Information Center
Franklyn, Virginia
2013-01-01
The Friends of the Pikes Peak Library District (PPLD) had a legacy website that was remotely hosted and maintained by a volunteer. They wanted a more significant presence on PPLD.org and an easier interface with which to update information. The developers decided to create a new Friends site in Drupal, using RWD, CSS3, and HTML5. The plan was to…
Institute for Advanced Education in Geospatial Sciences Educating the Next Generation of Scientists
NASA Technical Reports Server (NTRS)
Lawhead, Pamela; Johnson, Jay
2003-01-01
The project, as stated earlier is sponsored by NASA and is located at the University of Mississippi in Oxford, MS. It has two principal investigators with one of them, Pam Lawhead, serving as the Director of the Institute. The goal of the project is to create fifty online courses in Remote Sensing over the five year life of the project. Each year ten courses are put out for bid and the best ten submissions are accepted. This request for proposals insures that the course creators are content experts. Equivalence of product drives the online hosting of the courses. That is, we want the online presentation and delivery of each course to be as multi-media intensive as is effective. The goal is not to replace existing courses but, to provide courses created by content experts to as many colleges and universities as possible. This effort to create and host online courses has as its final goal the creation of a very large college educated workforce prepared to use the vast stores of information gathers by NASA and other remote sensing industries to enhance life on this planet.
Remotely Accessible Testbed for Software Defined Radio Development
NASA Technical Reports Server (NTRS)
Lux, James P.; Lang, Minh; Peters, Kenneth J.; Taylor, Gregory H.
2012-01-01
Previous development testbeds have assumed that the developer was physically present in front of the hardware being used. No provision for remote operation of basic functions (power on/off or reset) was made, because the developer/operator was sitting in front of the hardware, and could just push the button manually. In this innovation, a completely remotely accessible testbed has been created, with all diagnostic equipment and tools set up for remote access, and using standardized interfaces so that failed equipment can be quickly replaced. In this testbed, over 95% of the operating hours were used for testing without the developer being physically present. The testbed includes a pair of personal computers, one running Linux and one running Windows. A variety of peripherals is connected via Ethernet and USB (universal serial bus) interfaces. A private internal Ethernet is used to connect to test instruments and other devices, so that the sole connection to the outside world is via the two PCs. An important design consideration was that all of the instruments and interfaces used stable, long-lived industry standards, such as Ethernet, USB, and GPIB (general purpose interface bus). There are no plug-in cards for the two PCs, so there are no problems with finding replacement computers with matching interfaces, device drivers, and installation. The only thing unique to the two PCs is the locally developed software, which is not specific to computer or operating system version. If a device (including one of the computers) were to fail or become unavailable (e.g., a test instrument needed to be recalibrated), replacing it is a straightforward process with a standard, off-the-shelf device.
An Electronic Pressure Profile Display system for aeronautic test facilities
NASA Technical Reports Server (NTRS)
Woike, Mark R.
1990-01-01
The NASA Lewis Research Center has installed an Electronic Pressure Profile Display system. This system provides for the real-time display of pressure readings on high resolution graphics monitors. The Electronic Pressure Profile Display system will replace manometer banks currently used in aeronautic test facilities. The Electronic Pressure Profile Display system consists of an industrial type Digital Pressure Transmitter (DPI) unit which interfaces with a host computer. The host computer collects the pressure data from the DPI unit, converts it into engineering units, and displays the readings on a high resolution graphics monitor in bar graph format. Software was developed to accomplish the above tasks and also draw facility diagrams as background information on the displays. Data transfer between host computer and DPT unit is done with serial communications. Up to 64 channels are displayed with one second update time. This paper describes the system configuration, its features, and its advantages over existing systems.
An electronic pressure profile display system for aeronautic test facilities
NASA Technical Reports Server (NTRS)
Woike, Mark R.
1990-01-01
The NASA Lewis Research Center has installed an Electronic Pressure Profile Display system. This system provides for the real-time display of pressure readings on high resolution graphics monitors. The Electronic Pressure Profile Display system will replace manometer banks currently used in aeronautic test facilities. The Electronic Pressure Profile Display system consists of an industrial type Digital Pressure Transmitter (DPT) unit which interfaces with a host computer. The host computer collects the pressure data from the DPT unit, converts it into engineering units, and displays the readings on a high resolution graphics monitor in bar graph format. Software was developed to accomplish the above tasks and also draw facility diagrams as background information on the displays. Data transfer between host computer and DPT unit is done with serial communications. Up to 64 channels are displayed with one second update time. This paper describes the system configuration, its features, and its advantages over existing systems.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rahman, Nur Aira Abd, E-mail: nur-aira@nuclearmalaysia.gov.my; Yussup, Nolida; Ibrahim, Maslina Bt. Mohd
2015-04-29
A DAQ (data acquisition) software called RPTv2.0 has been developed for Radioactive Particle Tracking System in Malaysian Nuclear Agency. RPTv2.0 that features scanning control GUI, data acquisition from 12-channel counter via RS-232 interface, and multichannel analyzer (MCA). This software is fully developed on National Instruments Labview 8.6 platform. Ludlum Model 4612 Counter is used to count the signals from the scintillation detectors while a host computer is used to send control parameters, acquire and display data, and compute results. Each detector channel consists of independent high voltage control, threshold or sensitivity value and window settings. The counter is configured withmore » a host board and twelve slave boards. The host board collects the counts from each slave board and communicates with the computer via RS-232 data interface.« less
Species interactions in occurrence data for a community of tick-transmitted pathogens
Estrada-Peña, Agustín; de la Fuente, José
2016-01-01
Interactions between tick species, their realized range of hosts, the pathogens they carry and transmit, and the geographic distribution of species in the Western Palearctic were determined based on evidence published between 1970–2014. These relationships were linked to remotely sensed features of temperature and vegetation and used to extract the network of interactions among the organisms. The resulting datasets focused on niche overlap among ticks and hosts, species interactions, and the fraction of the environmental niche in which tick-borne pathogens may circulate as a result of interactions and overlapping environmental traits. The resulting datasets provide a valuable resource for researchers interested in tick-borne pathogens, as they conciliate the abiotic and biotic sides of their niche, allowing exploration of the importance of each host species acting as a vertebrate reservoir in the circulation of tick-transmitted pathogens in the environmental niche. PMID:27479213
ERIC Educational Resources Information Center
Herrera, Oriel A.; Fuller, David A.
2011-01-01
Remote experimentation laboratories (REL) are systems based on real equipment that allow students to carry out a laboratory practice through the Internet on the computer. In engineering, there have been numerous initiatives to implement REL over recent years, given the fundamental role of laboratory activities. However, in the past efforts have…
Evaluating Intention to Use Remote Robotics Experimentation in Programming Courses
ERIC Educational Resources Information Center
Cheng, Pericles L.
2017-01-01
The Digital Agenda for Europe (2015) states that there will be 825,000 unfilled vacancies for Information and Communications Technology by 2020. This lack of IT professionals stems from the small number of students graduating in computer science. To retain more students in the field, teachers can use remote robotic experiments to explain difficult…
A Simple Solution to Providing Remote Access to CD-ROM.
ERIC Educational Resources Information Center
Garnham, Carla T.; Brodie, Kent
1990-01-01
A pilot project at the Medical College of Wisconsin illustrates how even small computing organizations with limited financial and staff resources can provide remote access to CD-ROM (Compact Disc-Read-Only-Memory) databases, and that providing such convenient access to a vast array of useful information can greatly benefit faculty and students.…
ERIC Educational Resources Information Center
Turcotte, Sandrine; Hamel, Christine
2016-01-01
This study addressed computer-supported collaborative scientific inquiries in remote networked schools (Quebec, Canada). Three dyads of Grade 5-6 classrooms from remote locations across the province collaborated using the knowledge-building tool Knowledge Forum. Customized scaffold supports embedded in the online tool were used to support student…
NASA Astrophysics Data System (ADS)
Linsky, Jeffrey
2017-08-01
We propose to compute state-of-the-art model atmospheres (photospheres, chromospheres, transition regions and coronae) of the 4 K and 7 M exoplanet host stars observed by HST in the MUSCLES Treasury Survey, the nearest host star Proxima Centauri, and TRAPPIST-1. Our semi-empirical models will fit theunique high-resolution panchromatic (X-ray to infrared) spectra of these stars in the MAST High-Level Science Products archive consisting of COS and STIS UV spectra and near-simultaneous Chandra, XMM-Newton, and ground-based observations. We will compute models with the fully tested SSRPM computer software incorporating 52 atoms and ions in full non-LTE (435,986 spectral lines) and the 20 most-abundant diatomic molecules (about 2 million lines). This code has successfully fit the panchromatic spectrum of the M1.5 V exoplanet host star GJ 832 (Fontenla et al. 2016), the first M star with such a detailed model, and solar spectra. Our models will (1) predict the unobservable extreme-UV spectra, (2) determine radiative energy losses and balancing heating rates throughout these atmospheres, (3) compute a stellar irradiance library needed to describe the radiation environment of potentially habitable exoplanets to be studied by TESS and JWST, and (4) in the long post-HST era when UV observations will not be possible, the stellar irradiance library will be a powerful tool for predicting the panchromatic spectra of host stars that have only limited spectral coverage, in particular no UV spectra. The stellar models and spectral irradiance library will be placed quickly in MAST.
Ikhana: A NASA UAS Supporting Long Duration Earth Science Missions
NASA Technical Reports Server (NTRS)
Cobleigh, Brent R.
2006-01-01
NASA's Ikhana unmanned aerial vehicle (UAV) is a General Atomics MQ-9 Predator-B modified to support the conduct of Earth science missions for the NASA Science Mission Directorate through partnerships, other government agencies and universities. Ikhana, a Native American word meaning 'intelligence', can carry over 2000 lbs of atmospheric and remote sensing instruments in the payload bay and external pods. The aircraft is capable of mission durations in excess of 24 hours at altitudes above 40,000 ft. Redundant flight control, avionics, power, and network systems increase the system reliability and allow easier access to public airspace. The aircraft is remotely piloted from a mobile ground control station (GCS) using both C-band line-of-sight and Ku-band over-the-horizon satellite datalinks. NASA's GCS has been modified to support on-site science monitoring, or the downlink data can be networked to remote sites. All ground support systems are designed to be deployable to support global Eart science investigations. On-board support capabilities include an instrumentation system and an Airborne Research Test System (ARTS). The ARTS can host research algorithms that will autonomously command and control on-board sensors, perform sensor health monitoring, conduct data analysis, and request changes to the flight plan to maximize data collection. The ARTS also has the ability to host algorithms that will autonomously control the aircraft trajectory based on sensor needs, (e.g. precision trajectory for repeat pass interferometry) or to optimize mission objectives (e.g. search for specific atmospheric conditions). Standard on-board networks will collect science data for recording and for inclusion in the aircraft's high bandwidth downlink. The Ikhana project will complete GCS development, science support systems integration, external pod integration and flight clearance, and operations crew training in early 2007. A large-area remote sensing mission is currently scheduled for the Summer 2007.
Processing implicit control: evidence from reading times
McCourt, Michael; Green, Jeffrey J.; Lau, Ellen; Williams, Alexander
2015-01-01
Sentences such as “The ship was sunk to collect the insurance” exhibit an unusual form of anaphora, implicit control, where neither anaphor nor antecedent is audible. The non-finite reason clause has an understood subject, PRO, that is anaphoric; here it may be understood as naming the agent of the event of the host clause. Yet since the host is a short passive, this agent is realized by no audible dependent. The putative antecedent to PRO is therefore implicit, which it normally cannot be. What sorts of representations subserve the comprehension of this dependency? Here we present four self-paced reading time studies directed at this question. Previous work showed no processing cost for implicit vs. explicit control, and took this to support the view that PRO is linked syntactically to a silent argument in the passive. We challenge this conclusion by reporting that we also find no processing cost for remote implicit control, as in: “The ship was sunk. The reason was to collect the insurance.” Here the dependency crosses two independent sentences, and so cannot, we argue, be mediated by syntax. Our Experiments 1–4 examined the processing of both implicit (short passive) and explicit (active or long passive) control in both local and remote configurations. Experiments 3 and 4 added either “3 days ago” or “just in order” to the local conditions, to control for the distance between the passive and infinitival verbs, and for the predictability of the reason clause, respectively. We replicate the finding that implicit control does not impose an additional processing cost. But critically we show that remote control does not impose a processing cost either. Reading times at the reason clause were never slower when control was remote. In fact they were always faster. Thus, efficient processing of local implicit control cannot show that implicit control is mediated by syntax; nor, in turn, that there is a silent but grammatically active argument in passives. PMID:26579016
Ubiquitous Accessibility for People with Visual Impairments: Are We There Yet?
Billah, Syed Masum; Ashok, Vikas; Porter, Donald E.; Ramakrishnan, IV
2017-01-01
Ubiquitous access is an increasingly common vision of computing, wherein users can interact with any computing device or service from anywhere, at any time. In the era of personal computing, users with visual impairments required special-purpose, assistive technologies, such as screen readers, to interact with computers. This paper investigates whether technologies like screen readers have kept pace with, or have created a barrier to, the trend toward ubiquitous access, with a specific focus on desktop computing as this is still the primary way computers are used in education and employment. Towards that, the paper presents a user study with 21 visually-impaired participants, specifically involving the switching of screen readers within and across different computing platforms, and the use of screen readers in remote access scenarios. Among the findings, the study shows that, even for remote desktop access—an early forerunner of true ubiquitous access—screen readers are too limited, if not unusable. The study also identifies several accessibility needs, such as uniformity of navigational experience across devices, and recommends potential solutions. In summary, assistive technologies have not made the jump into the era of ubiquitous access, and multiple, inconsistent screen readers create new practical problems for users with visual impairments. PMID:28782061
Ubiquitous Accessibility for People with Visual Impairments: Are We There Yet?
Billah, Syed Masum; Ashok, Vikas; Porter, Donald E; Ramakrishnan, I V
2017-05-01
Ubiquitous access is an increasingly common vision of computing, wherein users can interact with any computing device or service from anywhere, at any time. In the era of personal computing, users with visual impairments required special-purpose, assistive technologies, such as screen readers, to interact with computers. This paper investigates whether technologies like screen readers have kept pace with, or have created a barrier to, the trend toward ubiquitous access, with a specific focus on desktop computing as this is still the primary way computers are used in education and employment. Towards that, the paper presents a user study with 21 visually-impaired participants, specifically involving the switching of screen readers within and across different computing platforms, and the use of screen readers in remote access scenarios. Among the findings, the study shows that, even for remote desktop access-an early forerunner of true ubiquitous access-screen readers are too limited, if not unusable. The study also identifies several accessibility needs, such as uniformity of navigational experience across devices, and recommends potential solutions. In summary, assistive technologies have not made the jump into the era of ubiquitous access, and multiple, inconsistent screen readers create new practical problems for users with visual impairments.
NASA Technical Reports Server (NTRS)
Kumar, L.
1978-01-01
A computer program is described for calculating the flexibility coefficients as arm design changes are made for the remote manipulator system. The coefficients obtained are required as input for a second program which reduces the number of payload deployment and retrieval system simulation runs required to simulate the various remote manipulator system maneuvers. The second program calculates end effector flexibility and joint flexibility terms for the torque model of each joint for any arbitrary configurations. The listing of both programs is included in the appendix.
U. S. GEOLOGICAL SURVEY LAND REMOTE SENSING ACTIVITIES.
Frederick, Doyle G.
1983-01-01
USGS uses all types of remotely sensed data, in combination with other sources of data, to support geologic analyses, hydrologic assessments, land cover mapping, image mapping, and applications research. Survey scientists use all types of remotely sensed data with ground verifications and digital topographic and cartographic data. A considerable amount of research is being done by Survey scientists on developing automated geographic information systems that can handle a wide variety of digital data. The Survey is also investigating the use of microprocessor computer systems for accessing, displaying, and analyzing digital data.
NASA Technical Reports Server (NTRS)
Sellers, J. F.
1973-01-01
The transient performance of two concepts for control of vertical takeoff aircraft remote lift fans is analyzed and discussed. Both concepts employ flow transfer between pairs of lift fans located in separate parts of the aircraft in order to obtain attitude control moments for hover and low-speed flight. The results presented are from a digital computer, dynamic analysis of the YJ97/LF460 remote drive turbofan. The transient responses of the two systems are presented for step demands in lift and moment.
Flight-test experience in digital control of a remotely piloted vehicle.
NASA Technical Reports Server (NTRS)
Edwards, J. W.
1972-01-01
The development of a remotely piloted vehicle system consisting of a remote pilot cockpit and a ground-based digital computer coupled to the aircraft through telemetry data links is described. The feedback control laws are implemented in a FORTRAN program. Flight-test experience involving high feedback gain limits for attitude and attitude rate feedback variables, filtering of sampled data, and system operation during intermittent telemetry data link loss is discussed. Comparisons of closed-loop flight tests with analytical calculations, and pilot comments on system operation are included.
An Efficient Authenticated Key Transfer Scheme in Client-Server Networks
NASA Astrophysics Data System (ADS)
Shi, Runhua; Zhang, Shun
2017-10-01
In this paper, we presented a novel authenticated key transfer scheme in client-server networks, which can achieve two secure goals of remote user authentication and the session key establishment between the remote user and the server. Especially, the proposed scheme can subtly provide two fully different authentications: identity-base authentication and anonymous authentication, while the remote user only holds a private key. Furthermore, our scheme only needs to transmit 1-round messages from the remote user to the server, thus it is very efficient in communication complexity. In addition, the most time-consuming computation in our scheme is elliptic curve scalar point multiplication, so it is also feasible even for mobile devices.
Study on Remote Monitoring System of Crossing and Spanning Tangent Tower
NASA Astrophysics Data System (ADS)
Chen, Da-bing; Zhang, Nai-long; Zhang, Meng-ge; Wang, Ze-hua; Zhang, Yan
2017-05-01
In order to grasp the vibration state of overhead transmission line and ensure the operational security of transmission line, the remote monitoring system of crossing and spanning tangent tower was studied. By use of this system, the displacement, velocity and acceleration of the tower, and the local weather data are collected automatically, displayed on computer of remote monitoring centre through wireless network, real-time collection and transmission of vibration signals are realized. The applying results show that the system is excellent in reliability and accuracy and so on. The system can be used to remote monitoring of transmission tower of UHV power transmission lines and in large spanning areas.
uPy: a ubiquitous computer graphics Python API with Biological Modeling Applications
Autin, L.; Johnson, G.; Hake, J.; Olson, A.; Sanner, M.
2015-01-01
In this paper we describe uPy, an extension module for the Python programming language that provides a uniform abstraction of the APIs of several 3D computer graphics programs called hosts, including: Blender, Maya, Cinema4D, and DejaVu. A plugin written with uPy is a unique piece of code that will run in all uPy-supported hosts. We demonstrate the creation of complex plug-ins for molecular/cellular modeling and visualization and discuss how uPy can more generally simplify programming for many types of projects (not solely science applications) intended for multi-host distribution. uPy is available at http://upy.scripps.edu PMID:24806987
NASA Astrophysics Data System (ADS)
Washington-Allen, R. A.; Fatoyinbo, T. E.; Ribeiro, N. S.; Shugart, H. H.; Therrell, M. D.; Vaz, K. T.; von Schill, L.
2006-12-01
A workshop titled: Environmental Remote Sensing for Natural Resources Management was held from June 12 23, 2006 at Eduardo Mondlane University in Maputo Mozambique. The workshop was initiated through an invitation and pre-course evaluation form to interested NGOs, universities, and government organizations. The purpose of the workshop was to provide training to interested professionals, graduate students, faculty and researchers at Mozambican institutions on the research and practical uses of remote sensing for natural resource management. The course had 24 participants who were predominantly professionals in remote sensing and GIS from various NGOs, governmental and academic institutions in Mozambique. The course taught remote sensing from an ecological perspective, specifically the course focused on the application of new remote sensing technology [the Shuttle Radar Topography Mission (SRTM) C-band radar data] to carbon accounting research in Miombo woodlands and Mangrove forests. The 2-week course was free to participants and consisted of lectures, laboratories, and a field trip to the mangrove forests of Inhaca Island, Maputo. The field trip consisted of training in the use of forest inventory techniques in support of remote sensing studies. Specifically, the field workshop centered on use of Global Positioning Systems (GPS) and collection of forest inventory data on tree height, structure [leaf area index (LAI)], and productivity. Productivity studies were enhanced with the teaching of introductory dendrochronology including sample collection of tree rings from four different mangrove species. Students were provided with all course materials including a DVD that contained satellite data (e.g., Landsat and SRTM imagery), ancillary data, lectures, exercises, and remote sensing publications used in the course including a CD from the Environmental Protection Agency's Environmental Photographic Interpretation Center's (EPA-EPIC) program to teach remote sensing and data CDs from NASA's SAFARI 2000 field campaign. Nineteen participants evaluated the effectiveness of the course in regards to the course lectures, instructors, and the field trip. Future workshops should focus more on the individual projects that students are engaged with in their jobs, replace the laboratories computers with workstations geared towards computer intensive image processing software, and the purchase of field remote sensing instrumentation for practical exercises.
NASA Technical Reports Server (NTRS)
Anuta, P. E.
1975-01-01
Least squares approximation techniques were developed for use in computer aided correction of spatial image distortions for registration of multitemporal remote sensor imagery. Polynomials were first used to define image distortion over the entire two dimensional image space. Spline functions were then investigated to determine if the combination of lower order polynomials could approximate a higher order distortion with less computational difficulty. Algorithms for generating approximating functions were developed and applied to the description of image distortion in aircraft multispectral scanner imagery. Other applications of the techniques were suggested for earth resources data processing areas other than geometric distortion representation.
Martin, Suzanne; Armstrong, Elaine; Thomson, Eileen; Vargiu, Eloisa; Solà, Marc; Dauwalder, Stefan; Miralles, Felip; Daly Lynn, Jean
2017-07-14
Cognitive rehabilitation is established as a core intervention within rehabilitation programs following a traumatic brain injury (TBI). Digitally enabled assistive technologies offer opportunities for clinicians to increase remote access to rehabilitation supporting transition into home. Brain Computer Interface (BCI) systems can harness the residual abilities of individuals with limited function to gain control over computers through their brain waves. This paper presents an online cognitive rehabilitation application developed with therapists, to work remotely with people who have TBI, who will use BCI at home to engage in the therapy. A qualitative research study was completed with people who are community dwellers post brain injury (end users), and a cohort of therapists involved in cognitive rehabilitation. A user-centered approach over three phases in the development, design and feasibility testing of this cognitive rehabilitation application included two tasks (Find-a-Category and a Memory Card task). The therapist could remotely prescribe activity with different levels of difficulty. The service user had a home interface which would present the therapy activities. This novel work was achieved by an international consortium of academics, business partners and service users.
Energy Conservation Using Dynamic Voltage Frequency Scaling for Computational Cloud
Florence, A. Paulin; Shanthi, V.; Simon, C. B. Sunil
2016-01-01
Cloud computing is a new technology which supports resource sharing on a “Pay as you go” basis around the world. It provides various services such as SaaS, IaaS, and PaaS. Computation is a part of IaaS and the entire computational requests are to be served efficiently with optimal power utilization in the cloud. Recently, various algorithms are developed to reduce power consumption and even Dynamic Voltage and Frequency Scaling (DVFS) scheme is also used in this perspective. In this paper we have devised methodology which analyzes the behavior of the given cloud request and identifies the associated type of algorithm. Once the type of algorithm is identified, using their asymptotic notations, its time complexity is calculated. Using best fit strategy the appropriate host is identified and the incoming job is allocated to the victimized host. Using the measured time complexity the required clock frequency of the host is measured. According to that CPU frequency is scaled up or down using DVFS scheme, enabling energy to be saved up to 55% of total Watts consumption. PMID:27239551
Energy Conservation Using Dynamic Voltage Frequency Scaling for Computational Cloud.
Florence, A Paulin; Shanthi, V; Simon, C B Sunil
2016-01-01
Cloud computing is a new technology which supports resource sharing on a "Pay as you go" basis around the world. It provides various services such as SaaS, IaaS, and PaaS. Computation is a part of IaaS and the entire computational requests are to be served efficiently with optimal power utilization in the cloud. Recently, various algorithms are developed to reduce power consumption and even Dynamic Voltage and Frequency Scaling (DVFS) scheme is also used in this perspective. In this paper we have devised methodology which analyzes the behavior of the given cloud request and identifies the associated type of algorithm. Once the type of algorithm is identified, using their asymptotic notations, its time complexity is calculated. Using best fit strategy the appropriate host is identified and the incoming job is allocated to the victimized host. Using the measured time complexity the required clock frequency of the host is measured. According to that CPU frequency is scaled up or down using DVFS scheme, enabling energy to be saved up to 55% of total Watts consumption.
NASA Astrophysics Data System (ADS)
Zimmermann, Robert; Brandmeier, Melanie; Andreani, Louis; Gloaguen, Richard
2015-04-01
Remote sensing data can provide valuable information about ore deposits and their alteration zones at surface level. High spectral and spatial resolution of the data is essential for detailed mapping of mineral abundances and related structures. Carbonatites are well known for hosting economic enrichments in REE, Ta, Nb and P (Jones et al. 2013). These make them a preferential target for exploration for those critical elements. In this study we show how combining geomorphic, textural and spectral data improves classification result. We selected a site with a well-known occurrence in northern Namibia: the Epembe dyke. For analysis LANDSAT 8, SRTM and airborne hyperspectral (HyMap) data were chosen. The overlapping data allows a multi-scale and multi-resolution approach. Results from data analysis were validated during fieldwork in 2014. Data was corrected for atmospherical and geometrical effects. Image classification, mineral mapping and tectonic geomorphology allow a refinement of the geological map by lithological mapping in a second step. Detailed mineral abundance maps were computed using spectral unmixing techniques. These techniques are well suited to map abundances of carbonate minerals, but not to discriminate the carbonatite itself from surrounding rocks with similar spectral signatures. Thus, geometric indices were calculated using tectonic geomorphology and textures. For this purpose the TecDEM-toolbox (SHAHZAD & GLOAGUEN 2011) was applied to the SRTM-data for geomorphic analysis. Textural indices (e.g. uniformity, entropy, angular second moment) were derived from HyMap and SRTM by a grey-level co-occurrence matrix (CLAUSI 2002). The carbonatite in the study area is ridge-forming and shows a narrow linear feature in the textural bands. Spectral and geometric information were combined using kohonen Self-Organizing Maps (SOM) for unsupervised clustering. The resulting class spectra were visually compared and interpreted. Classes with similar signatures were merged according to geological context. The major conclusions are: 1. Carbonate minerals can be mapped using spectral unmixing techniques. 2. Carbonatites are associated with specific geometric pattern 3. The combination of spectral and geometric information improves classification result and reduces misclassification. References Clausi, D. A. (2002): An analysis of co-occurrence texture statistics as a function of grey-level quantization. - Canadian Journal of Remote Sensing, 28 (1), 45-62 Jones, A. P., Genge, M. and Carmody, L (2013): Carbonate Melts and Carbonatites. - Reviews in Mineralogy & Geochemistry, 75, 289-322 Shahzad, F. & Gloaguen, R. (2011): TecDEM: A MATLAB based toolbox for tectonic geomorphology, Part 2: Surface dynamics and basin analysis. - Computers and Geosciences, 37 (2), 261-271
Models of microbiome evolution incorporating host and microbial selection.
Zeng, Qinglong; Wu, Steven; Sukumaran, Jeet; Rodrigo, Allen
2017-09-25
Numerous empirical studies suggest that hosts and microbes exert reciprocal selective effects on their ecological partners. Nonetheless, we still lack an explicit framework to model the dynamics of both hosts and microbes under selection. In a previous study, we developed an agent-based forward-time computational framework to simulate the neutral evolution of host-associated microbial communities in a constant-sized, unstructured population of hosts. These neutral models allowed offspring to sample microbes randomly from parents and/or from the environment. Additionally, the environmental pool of available microbes was constituted by fixed and persistent microbial OTUs and by contributions from host individuals in the preceding generation. In this paper, we extend our neutral models to allow selection to operate on both hosts and microbes. We do this by constructing a phenome for each microbial OTU consisting of a sample of traits that influence host and microbial fitnesses independently. Microbial traits can influence the fitness of hosts ("host selection") and the fitness of microbes ("trait-mediated microbial selection"). Additionally, the fitness effects of traits on microbes can be modified by their hosts ("host-mediated microbial selection"). We simulate the effects of these three types of selection, individually or in combination, on microbiome diversities and the fitnesses of hosts and microbes over several thousand generations of hosts. We show that microbiome diversity is strongly influenced by selection acting on microbes. Selection acting on hosts only influences microbiome diversity when there is near-complete direct or indirect parental contribution to the microbiomes of offspring. Unsurprisingly, microbial fitness increases under microbial selection. Interestingly, when host selection operates, host fitness only increases under two conditions: (1) when there is a strong parental contribution to microbial communities or (2) in the absence of a strong parental contribution, when host-mediated selection acts on microbes concomitantly. We present a computational framework that integrates different selective processes acting on the evolution of microbiomes. Our framework demonstrates that selection acting on microbes can have a strong effect on microbial diversities and fitnesses, whereas selection on hosts can have weaker outcomes.
Hangout with CERN: a direct conversation with the public
NASA Astrophysics Data System (ADS)
Rao, Achintya; Goldfarb, Steven; Kahle, Kate
2016-04-01
Hangout with CERN refers to a weekly, half-hour-long, topical webcast hosted at CERN. The aim of the programme is threefold: (i) to provide a virtual tour of various locations and facilities at CERN, (ii) to discuss the latest scientific results from the laboratory, and, most importantly, (iii) to engage in conversation with the public and answer their questions. For each ;episode;, scientists gather around webcam-enabled computers at CERN and partner institutes/universities, connecting to one another using the Google+ social network's ;Hangouts; tool. The show is structured as a conversation mediated by a host, usually a scientist, and viewers can ask questions to the experts in real time through a Twitter hashtag or YouTube comments. The history of Hangout with CERN can be traced back to ICHEP 2012, where several physicists crowded in front of a laptop connected to Google+, using a ;Hangout On Air; webcast to explain to the world the importance of the discovery of the Higgs-like boson, announced just two days before at the same conference. Hangout with CERN has also drawn inspiration from two existing outreach endeavours: (i) ATLAS Virtual Visits, which connected remote visitors with scientists in the ATLAS Control Room via video conference, and (ii) the Large Hangout Collider, in which CMS scientists gave underground tours via Hangouts to groups of schools and members of the public around the world. In this paper, we discuss the role of Hangout with CERN as a bi-directional outreach medium and an opportunity to train scientists in effective communication.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gagne, Roderick B.; Hogan, J. Derek; McIntyre, Peter B.
1. Co-introductions of non-native parasites with non-native hosts can be a major driver of disease emergence in native species, but the conditions that promote the establishment and spread of nonnative parasites remain poorly understood. Here, we characterise the infection of a native host species by a non-native parasite relative to the distribution and density of the original non-native host species and a suite of organismal and environmental factors that have been associated with parasitism, but not commonly considered within a single system. 2. We examined the native Hawaiian goby Awaous stamineus across 23 catchments on five islands for infection bymore » the non-native nematode parasite Camallanus cotti. We used model selection to test whether parasite infection was associated with the genetic diversity, size and population density of native hosts, the distribution and density of non-native hosts, land use and water quality. 3. We found that the distribution of non-native C. cotti parasites has become decoupled from the non-native hosts that were primary vectors of introduction to the Hawaiian Islands. Although no single intrinsic or extrinsic factor was identified that best explains parasitism of A. stamineus by C. cotti, native host size, population density and water quality were consistently identified as influencing parasite intensity and prevalence. 4. The introduction of non-native species can indirectly influence native species through infection of co-introduced parasites. Here, we show that the effects of enemy addition can extend beyond the range of non-native hosts through the independent spread of non-native parasites. This suggests that control of non-native hosts is not sufficient to halt the spread of introduced parasites. Furthermore, designing importation regulations to prevent host parasite co-introductions can promote native species conservation, even in remote areas that may not seem susceptible to human influence.« less
Remote health monitoring system for detecting cardiac disorders.
Bansal, Ayush; Kumar, Sunil; Bajpai, Anurag; Tiwari, Vijay N; Nayak, Mithun; Venkatesan, Shankar; Narayanan, Rangavittal
2015-12-01
Remote health monitoring system with clinical decision support system as a key component could potentially quicken the response of medical specialists to critical health emergencies experienced by their patients. A monitoring system, specifically designed for cardiac care with electrocardiogram (ECG) signal analysis as the core diagnostic technique, could play a vital role in early detection of a wide range of cardiac ailments, from a simple arrhythmia to life threatening conditions such as myocardial infarction. The system that the authors have developed consists of three major components, namely, (a) mobile gateway, deployed on patient's mobile device, that receives 12-lead ECG signals from any ECG sensor, (b) remote server component that hosts algorithms for accurate annotation and analysis of the ECG signal and (c) point of care device of the doctor to receive a diagnostic report from the server based on the analysis of ECG signals. In the present study, their focus has been toward developing a system capable of detecting critical cardiac events well in advance using an advanced remote monitoring system. A system of this kind is expected to have applications ranging from tracking wellness/fitness to detection of symptoms leading to fatal cardiac events.
A system for distributed intrusion detection
DOE Office of Scientific and Technical Information (OSTI.GOV)
Snapp, S.R.; Brentano, J.; Dias, G.V.
1991-01-01
The study of providing security in computer networks is a rapidly growing area of interest because the network is the medium over which most attacks or intrusions on computer systems are launched. One approach to solving this problem is the intrusion-detection concept, whose basic premise is that not only abandoning the existing and huge infrastructure of possibly-insecure computer and network systems is impossible, but also replacing them by totally-secure systems may not be feasible or cost effective. Previous work on intrusion-detection systems were performed on stand-alone hosts and on a broadcast local area network (LAN) environment. The focus of ourmore » present research is to extend our network intrusion-detection concept from the LAN environment to arbitarily wider areas with the network topology being arbitrary as well. The generalized distributed environment is heterogeneous, i.e., the network nodes can be hosts or servers from different vendors, or some of them could be LAN managers, like our previous work, a network security monitor (NSM), as well. The proposed architecture for this distributed intrusion-detection system consists of the following components: a host manager in each host; a LAN manager for monitoring each LAN in the system; and a central manager which is placed at a single secure location and which receives reports from various host and LAN managers to process these reports, correlate them, and detect intrusions. 11 refs., 2 figs.« less
Learning Methods of Remote Sensing In the 2013 Curriculum of Secondary School
NASA Astrophysics Data System (ADS)
Lili Somantri, Nandi
2016-11-01
The new remote sensing material included in the subjects of geography in the curriculum of 1994. For geography teachers generation of 90s and over who in college do not get the material remote sensing, for teaching is a tough matter. Most teachers only give a theoretical matter, and do not carry out practical reasons in the lack of facilities and infrastructure of computer laboratories. Therefore, in this paper studies the importance about the method or manner of teaching remote sensing material in schools. The purpose of this paper is 1) to explain the position of remote sensing material in the study of geography, 2) analyze the Geography Curriculum 2013 Subjects related to remote sensing material, 3) describes a method of teaching remote sensing material in schools. The method used in this paper is a descriptive analytical study supported by the literature. The conclusion of this paper that the position of remote sensing in the study of geography is a method or a way to obtain spatial data earth's surface. In the 2013 curriculum remote sensing material has been applied to the study of land use and transportation. Remote sensing methods of teaching must go through a practicum, which starts from the introduction of the theory of remote sensing, data extraction phase of remote sensing imagery to produce maps, both visually and digitally, field surveys, interpretation of test accuracy, and improved maps.
Circus: A Replicated Procedure Call Facility
1984-08-01
Computer Science Laboratory, Xerox PARC, July 1082 . [24) Bruce Ja.y Nelson. Remote Procedure Ctdl. Ph.D. dissertation, Computer Science Department...t. Ph.D. dissertation, Computer Science Division, University of California, Berkeley, Xerox PARC report number CSIF 82-7, December 1082 . [30...Tandem Computers Inc. GUARDIAN Opet’ating Sy•tem Programming Mt~nulll, Volumu 1 11nd 2. C upertino, California, 1082 . [31) R. H. Thoma.s. A majority
Computer networks for remote laboratories in physics and engineering
NASA Technical Reports Server (NTRS)
Starks, Scott; Elizandro, David; Leiner, Barry M.; Wiskerchen, Michael
1988-01-01
This paper addresses a relatively new approach to scientific research, telescience, which is the conduct of scientific operations in locations remote from the site of central experimental activity. A testbed based on the concepts of telescience is being developed to ultimately enable scientific researchers on earth to conduct experiments onboard the Space Station. This system along with background materials are discussed.
J. McKean; D. Tonina; C. Bohn; C. W. Wright
2014-01-01
New remote sensing technologies and improved computer performance now allow numerical flow modeling over large stream domains. However, there has been limited testing of whether channel topography can be remotely mapped with accuracy necessary for such modeling. We assessed the ability of the Experimental Advanced Airborne Research Lidar, to support a multi-dimensional...
Multivariate Density Estimation and Remote Sensing
NASA Technical Reports Server (NTRS)
Scott, D. W.
1983-01-01
Current efforts to develop methods and computer algorithms to effectively represent multivariate data commonly encountered in remote sensing applications are described. While this may involve scatter diagrams, multivariate representations of nonparametric probability density estimates are emphasized. The density function provides a useful graphical tool for looking at data and a useful theoretical tool for classification. This approach is called a thunderstorm data analysis.
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1997-05-01
The Interactive Computer-Enhanced Remote Viewing System (ICERVS) is a software tool for complex three-dimensional (3-D) visualization and modeling. Its primary purpose is to facilitate the use of robotic and telerobotic systems in remote and/or hazardous environments, where spatial information is provided by 3-D mapping sensors. ICERVS provides a robust, interactive system for viewing sensor data in 3-D and combines this with interactive geometric modeling capabilities that allow an operator to construct CAD models to match the remote environment. Part I of this report traces the development of ICERVS through three evolutionary phases: (1) development of first-generation software to render orthogonalmore » view displays and wireframe models; (2) expansion of this software to include interactive viewpoint control, surface-shaded graphics, material (scalar and nonscalar) property data, cut/slice planes, color and visibility mapping, and generalized object models; (3) demonstration of ICERVS as a tool for the remediation of underground storage tanks (USTs) and the dismantlement of contaminated processing facilities. Part II of this report details the software design of ICERVS, with particular emphasis on its object-oriented architecture and user interface.« less
Capturing remote mixing due to internal tides using multi-scale modeling tool: SOMAR-LES
NASA Astrophysics Data System (ADS)
Santilli, Edward; Chalamalla, Vamsi; Scotti, Alberto; Sarkar, Sutanu
2016-11-01
Internal tides that are generated during the interaction of an oscillating barotropic tide with the bottom bathymetry dissipate only a fraction of their energy near the generation region. The rest is radiated away in the form of low- high-mode internal tides. These internal tides dissipate energy at remote locations when they interact with the upper ocean pycnocline, continental slope, and large scale eddies. Capturing the wide range of length and time scales involved during the life-cycle of internal tides is computationally very expensive. A recently developed multi-scale modeling tool called SOMAR-LES combines the adaptive grid refinement features of SOMAR with the turbulence modeling features of a Large Eddy Simulation (LES) to capture multi-scale processes at a reduced computational cost. Numerical simulations of internal tide generation at idealized bottom bathymetries are performed to demonstrate this multi-scale modeling technique. Although each of the remote mixing phenomena have been considered independently in previous studies, this work aims to capture remote mixing processes during the life cycle of an internal tide in more realistic settings, by allowing multi-level (coarse and fine) grids to co-exist and exchange information during the time stepping process.
Calculating binding free energies of host-guest systems using the AMOEBA polarizable force field.
Bell, David R; Qi, Rui; Jing, Zhifeng; Xiang, Jin Yu; Mejias, Christopher; Schnieders, Michael J; Ponder, Jay W; Ren, Pengyu
2016-11-09
Molecular recognition is of paramount interest in many applications. Here we investigate a series of host-guest systems previously used in the SAMPL4 blind challenge by using molecular simulations and the AMOEBA polarizable force field. The free energy results computed by Bennett's acceptance ratio (BAR) method using the AMOEBA polarizable force field ranked favorably among the entries submitted to the SAMPL4 host-guest competition [Muddana, et al., J. Comput.-Aided Mol. Des., 2014, 28, 305-317]. In this work we conduct an in-depth analysis of the AMOEBA force field host-guest binding thermodynamics by using both BAR and the orthogonal space random walk (OSRW) methods. The binding entropy-enthalpy contributions are analyzed for each host-guest system. For systems of inordinate binding entropy-enthalpy values, we further examine the hydrogen bonding patterns and configurational entropy contribution. The binding mechanism of this series of host-guest systems varies from ligand to ligand, driven by enthalpy and/or entropy changes. Convergence of BAR and OSRW binding free energy methods is discussed. Ultimately, this work illustrates the value of molecular modelling and advanced force fields for the exploration and interpretation of binding thermodynamics.
NASA Astrophysics Data System (ADS)
Coburn, C. A.; Qin, Y.; Zhang, J.; Staenz, K.
2015-12-01
Food security is one of the most pressing issues facing humankind. Recent estimates predict that over one billion people don't have enough food to meet their basic nutritional needs. The ability of remote sensing tools to monitor and model crop production and predict crop yield is essential for providing governments and farmers with vital information to ensure food security. Google Earth Engine (GEE) is a cloud computing platform, which integrates storage and processing algorithms for massive remotely sensed imagery and vector data sets. By providing the capabilities of storing and analyzing the data sets, it provides an ideal platform for the development of advanced analytic tools for extracting key variables used in regional and national food security systems. With the high performance computing and storing capabilities of GEE, a cloud-computing based system for near real-time crop land monitoring was developed using multi-source remotely sensed data over large areas. The system is able to process and visualize the MODIS time series NDVI profile in conjunction with Landsat 8 image segmentation for crop monitoring. With multi-temporal Landsat 8 imagery, the crop fields are extracted using the image segmentation algorithm developed by Baatz et al.[1]. The MODIS time series NDVI data are modeled by TIMESAT [2], a software package developed for analyzing time series of satellite data. The seasonality of MODIS time series data, for example, the start date of the growing season, length of growing season, and NDVI peak at a field-level are obtained for evaluating the crop-growth conditions. The system fuses MODIS time series NDVI data and Landsat 8 imagery to provide information of near real-time crop-growth conditions through the visualization of MODIS NDVI time series and comparison of multi-year NDVI profiles. Stakeholders, i.e., farmers and government officers, are able to obtain crop-growth information at crop-field level online. This unique utilization of GEE in combination with advanced analytic and extraction techniques provides a vital remote sensing tool for decision makers and scientists with a high-degree of flexibility to adapt to different uses.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-09-04
...--Intersection of Cloud Computing and Mobility Forum and Workshop AGENCY: National Institute of Standards and.../intersection-of-cloud-and-mobility.cfm . SUPPLEMENTARY INFORMATION: NIST hosted six prior Cloud Computing Forum... interoperability, portability, and security, discuss the Federal Government's experience with cloud computing...
Computer-Controlled HVAC -- at Low Cost
ERIC Educational Resources Information Center
American School and University, 1974
1974-01-01
By tying into a computerized building-automation network, Schaumburg High School, Illinois, slashed its energy consumption by one-third. The remotely connected computer controls the mechanical system for the high school as well as other buildings in the community, with the cost being shared by all. (Author)
Computer program determines performance efficiency of remote measuring systems
NASA Technical Reports Server (NTRS)
Merewether, E. K.
1966-01-01
Computer programs control and evaluate instrumentation system performance for numerous rocket engine test facilities and prescribe calibration and maintenance techniques to maintain the systems within process specifications. Similar programs can be written for other test equipment in an industry such as the petrochemical industry.
A daily living activity remote monitoring system for solitary elderly people.
Maki, Hiromichi; Ogawa, Hidekuni; Matsuoka, Shingo; Yonezawa, Yoshiharu; Caldwell, W Morton
2011-01-01
A daily living activity remote monitoring system has been developed for supporting solitary elderly people. The monitoring system consists of a tri-axis accelerometer, six low-power active filters, a low-power 8-bit microcontroller (MC), a 1GB SD memory card (SDMC) and a 2.4 GHz low transmitting power mobile phone (PHS). The tri-axis accelerometer attached to the subject's chest can simultaneously measure dynamic and static acceleration forces produced by heart sound, respiration, posture and behavior. The heart rate, respiration rate, activity, posture and behavior are detected from the dynamic and static acceleration forces. These data are stored in the SD. The MC sends the data to the server computer every hour. The server computer stores the data and makes a graphic chart from the data. When the caregiver calls from his/her mobile phone to the server computer, the server computer sends the graphical chart via the PHS. The caregiver's mobile phone displays the chart to the monitor graphically.
Parallel Rendering of Large Time-Varying Volume Data
NASA Technical Reports Server (NTRS)
Garbutt, Alexander E.
2005-01-01
Interactive visualization of large time-varying 3D volume datasets has been and still is a great challenge to the modem computational world. It stretches the limits of the memory capacity, the disk space, the network bandwidth and the CPU speed of a conventional computer. In this SURF project, we propose to develop a parallel volume rendering program on SGI's Prism, a cluster computer equipped with state-of-the-art graphic hardware. The proposed program combines both parallel computing and hardware rendering in order to achieve an interactive rendering rate. We use 3D texture mapping and a hardware shader to implement 3D volume rendering on each workstation. We use SGI's VisServer to enable remote rendering using Prism's graphic hardware. And last, we will integrate this new program with ParVox, a parallel distributed visualization system developed at JPL. At the end of the project, we Will demonstrate remote interactive visualization using this new hardware volume renderer on JPL's Prism System using a time-varying dataset from selected JPL applications.
Network Penetration Testing and Research
NASA Technical Reports Server (NTRS)
Murphy, Brandon F.
2013-01-01
This paper will focus the on research and testing done on penetrating a network for security purposes. This research will provide the IT security office new methods of attacks across and against a company's network as well as introduce them to new platforms and software that can be used to better assist with protecting against such attacks. Throughout this paper testing and research has been done on two different Linux based operating systems, for attacking and compromising a Windows based host computer. Backtrack 5 and BlackBuntu (Linux based penetration testing operating systems) are two different "attacker'' computers that will attempt to plant viruses and or NASA USRP - Internship Final Report exploits on a host Windows 7 operating system, as well as try to retrieve information from the host. On each Linux OS (Backtrack 5 and BlackBuntu) there is penetration testing software which provides the necessary tools to create exploits that can compromise a windows system as well as other operating systems. This paper will focus on two main methods of deploying exploits 1 onto a host computer in order to retrieve information from a compromised system. One method of deployment for an exploit that was tested is known as a "social engineering" exploit. This type of method requires interaction from unsuspecting user. With this user interaction, a deployed exploit may allow a malicious user to gain access to the unsuspecting user's computer as well as the network that such computer is connected to. Due to more advance security setting and antivirus protection and detection, this method is easily identified and defended against. The second method of exploit deployment is the method mainly focused upon within this paper. This method required extensive research on the best way to compromise a security enabled protected network. Once a network has been compromised, then any and all devices connected to such network has the potential to be compromised as well. With a compromised network, computers and devices can be penetrated through deployed exploits. This paper will illustrate the research done to test ability to penetrate a network without user interaction, in order to retrieve personal information from a targeted host.
Accuracy of remote chest X-ray interpretation using Google Glass technology.
Spaedy, Emily; Christakopoulos, Georgios E; Tarar, Muhammad Nauman J; Christopoulos, Georgios; Rangan, Bavana V; Roesle, Michele; Ochoa, Cristhiaan D; Yarbrough, William; Banerjee, Subhash; Brilakis, Emmanouil S
2016-09-15
We sought to explore the accuracy of remote chest X-ray reading using hands-free, wearable technology (Google Glass, Google, Mountain View, California). We compared interpretation of twelve chest X-rays with 23 major cardiopulmonary findings by faculty and fellows from cardiology, radiology, and pulmonary-critical care via: (1) viewing the chest X-ray image on the Google Glass screen; (2) viewing a photograph of the chest X-ray taken using Google Glass and interpreted on a mobile device; (3) viewing the original chest X-ray on a desktop computer screen. One point was given for identification of each correct finding and a subjective rating of user experience was recorded. Fifteen physicians (5 faculty and 10 fellows) participated. The average chest X-ray reading score (maximum 23 points) as viewed through the Google Glass, Google Glass photograph on a mobile device, and the original X-ray viewed on a desktop computer was 14.1±2.2, 18.5±1.5 and 21.3±1.7, respectively (p<0.0001 between Google Glass and mobile device, p<0.0001 between Google Glass and desktop computer and p=0.0004 between mobile device and desktop computer). Of 15 physicians, 11 (73.3%) felt confident in detecting findings using the photograph taken by Google Glass as viewed on a mobile device. Remote chest X-ray interpretation using hands-free, wearable technology (Google Glass) is less accurate than interpretation using a desktop computer or a mobile device, suggesting that further technical improvements are needed before widespread application of this novel technology. Published by Elsevier Ireland Ltd.
Custom Sky-Image Mosaics from NASA's Information Power Grid
NASA Technical Reports Server (NTRS)
Jacob, Joseph; Collier, James; Craymer, Loring; Curkendall, David
2005-01-01
yourSkyG is the second generation of the software described in yourSky: Custom Sky-Image Mosaics via the Internet (NPO-30556), NASA Tech Briefs, Vol. 27, No. 6 (June 2003), page 45. Like its predecessor, yourSkyG supplies custom astronomical image mosaics of sky regions specified by requesters using client computers connected to the Internet. Whereas yourSky constructs mosaics on a local multiprocessor system, yourSkyG performs the computations on NASA s Information Power Grid (IPG), which is capable of performing much larger mosaicking tasks. (The IPG is high-performance computation and data grid that integrates geographically distributed 18 NASA Tech Briefs, September 2005 computers, databases, and instruments.) A user of yourSkyG can specify parameters describing a mosaic to be constructed. yourSkyG then constructs the mosaic on the IPG and makes it available for downloading by the user. The complexities of determining which input images are required to construct a mosaic, retrieving the required input images from remote sky-survey archives, uploading the images to the computers on the IPG, performing the computations remotely on the Grid, and downloading the resulting mosaic from the Grid are all transparent to the user
Efficacy of a telerehabilitation intervention programme using biofeedback among computer operators.
Golebowicz, Merav; Levanon, Yafa; Palti, Ram; Ratzon, Navah Z
2015-01-01
Computer operators spend long periods of time sitting in a static posture at computer workstations and therefore have an increased exposure to work-related musculoskeletal disorders (WRMSD). The present study is aimed at investigating the feasibility and effectiveness of a tele-biofeedback ergonomic intervention programme among computer operators suffering from WRMSD. Twelve subjects with WRMSD were assigned an ergonomic intervention accompanied by remote tele-biofeedback training, which was practised at their workstations. Evaluations of pain symptoms and locations, body posture and psychosocial characteristics were carried out before and after the intervention in the workplace. The hypothesis was partially verified as it showed improved body position at the workstation and decreased pain in some body parts. Tele-biofeedback, as part of an intervention, appears to be feasible and efficient for computer operators who suffer from WRMSD. This study encourages further research on tele-health within the scope of occupational therapy practice. Practitioner summary: Research concerning tele-health using biofeedback is scarce. The present study analyses the feasibility and partial effectiveness of a tele-biofeedback ergonomic intervention programme for computer operators suffering from WRMSD. The uniqueness and singularity of this study is the usage of remote communication between participants and practitioners through the Internet.
A Novel Centrality Measure for Network-wide Cyber Vulnerability Assessment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sathanur, Arun V.; Haglin, David J.
In this work we propose a novel formulation that models the attack and compromise on a cyber network as a combination of two parts - direct compromise of a host and the compromise occurring through the spread of the attack on the network from a compromised host. The model parameters for the nodes are a concise representation of the host profiles that can include the risky behaviors of the associated human users while the model parameters for the edges are based on the existence of vulnerabilities between each pair of connected hosts. The edge models relate to the summary representationsmore » of the corresponding attack-graphs. This results in a formulation based on Random Walk with Restart (RWR) and the resulting centrality metric can be solved for in an efficient manner through the use of sparse linear solvers. Thus the formulation goes beyond mere topological considerations in centrality computations by summarizing the host profiles and the attack graphs into the model parameters. The computational efficiency of the method also allows us to also quantify the uncertainty in the centrality measure through Monte Carlo analysis.« less
NaradaBrokering as Middleware Fabric for Grid-based Remote Visualization Services
NASA Astrophysics Data System (ADS)
Pallickara, S.; Erlebacher, G.; Yuen, D.; Fox, G.; Pierce, M.
2003-12-01
Remote Visualization Services (RVS) have tended to rely on approaches based on the client server paradigm. The simplicity in these approaches is offset by problems such as single-point-of-failures, scaling and availability. Furthermore, as the complexity, scale and scope of the services hosted on this paradigm increase, this approach becomes increasingly unsuitable. We propose a scheme based on top of a distributed brokering infrastructure, NaradaBrokering, which comprises a distributed network of broker nodes. These broker nodes are organized in a cluster-based architecture that can scale to very large sizes. The broker network is resilient to broker failures and efficiently routes interactions to entities that expressed an interest in them. In our approach to RVS, services advertise their capabilities to the broker network, which manages these service advertisements. Among the services considered within our system are those that perform graphic transformations, mediate access to specialized datasets and finally those that manage the execution of specified tasks. There could be multiple instances of each of these services and the system ensures that load for a given service is distributed efficiently over these service instances. Among the features provided in our approach are efficient discovery of services and asynchronous interactions between services and service requestors (which could themselves be other services). Entities need not be online during the execution of the service request. The system also ensures that entities can be notified about task executions, partial results and failures that might have taken place during service execution. The system also facilitates specification of task overrides, distribution of execution results to alternate devices (which were not used to originally request service execution) and to multiple users. These RVS services could of course be either OGSA (Open Grid Services Architecture) based Grid services or traditional Web services. The brokering infrastructure will manage the service advertisements and the invocation of these services. This scheme ensures that the fundamental Grid computing concept is met - provide computing capabilities of those that are willing to provide it to those that seek the same. {[1]} The NaradaBrokering Project: http://www.naradabrokering.org
Mathematic modeling of the Earth's surface and the process of remote sensing
NASA Technical Reports Server (NTRS)
Balter, B. M.
1979-01-01
It is shown that real data from remote sensing of the Earth from outer space are not best suited to the search for optimal procedures with which to process such data. To work out the procedures, it was proposed that data synthesized with the help of mathematical modeling be used. A criterion for simularity to reality was formulated. The basic principles for constructing methods for modeling the data from remote sensing are recommended. A concrete method is formulated for modeling a complete cycle of radiation transformations in remote sensing. A computer program is described which realizes the proposed method. Some results from calculations are presented which show that the method satisfies the requirements imposed on it.
Scientific Programming Using Java: A Remote Sensing Example
NASA Technical Reports Server (NTRS)
Prados, Don; Mohamed, Mohamed A.; Johnson, Michael; Cao, Changyong; Gasser, Jerry
1999-01-01
This paper presents results of a project to port remote sensing code from the C programming language to Java. The advantages and disadvantages of using Java versus C as a scientific programming language in remote sensing applications are discussed. Remote sensing applications deal with voluminous data that require effective memory management, such as buffering operations, when processed. Some of these applications also implement complex computational algorithms, such as Fast Fourier Transformation analysis, that are very performance intensive. Factors considered include performance, precision, complexity, rapidity of development, ease of code reuse, ease of maintenance, memory management, and platform independence. Performance of radiometric calibration code written in Java for the graphical user interface and of using C for the domain model are also presented.
Perturbation approach for nuclear magnetic resonance solid-state quantum computation
Berman, G. P.; Kamenev, D. I.; Tsifrinovich, V. I.
2003-01-01
A dynmore » amics of a nuclear-spin quantum computer with a large number ( L = 1000 ) of qubits is considered using a perturbation approach. Small parameters are introduced and used to compute the error in an implementation of an entanglement between remote qubits, using a sequence of radio-frequency pulses. The error is computed up to the different orders of the perturbation theory and tested using exact numerical solution.« less
Desert Research and Technology Studies (RATS) Local and Remote Test Sites
NASA Technical Reports Server (NTRS)
Janoiko, Barbara; Kosmo, Joseph; Eppler, Dean
2007-01-01
Desert RATS (Research and Technology Studies) is a combined group of inter-NASA center scientists and engineers, collaborating with representatives of industry and academia, for the purpose of conducting remote field exercises. These exercises provide the capability to validate experimental hardware and software, to evaluate and develop mission operational techniques, and to identify and establish technical requirements applicable for future planetary exploration. D-RATS completed its ninth year of field testing in September 2006. Dry run test activities prior to testing at designated remote field site locations are initially conducted at the Johnson Space Center (JSC) Remote Field Demonstration Test Site. This is a multi-acre external test site located at JSC and has detailed representative terrain features simulating both Lunar and Mars surface characteristics. The majority of the remote field tests have been subsequently conducted in various high desert areas adjacent to Flagstaff, Arizona. Both the local JSC and remote field test sites have terrain conditions that are representative of both the Moon and Mars, such as strewn rock and volcanic ash fields, meteorite crater ejecta blankets, rolling plains, hills, gullies, slopes, and outcrops. Flagstaff is the preferred remote test site location for many reasons. First, there are nine potential test sites with representative terrain features within a 75-mile radius. Second, Flagstaff is the location of the United States Geologic Survey (USGS)/Astrogeology Branch, which historically supported Apollo astronaut geologic training and currently supports and provides host accommodations to the D-RATS team. Finally, in considering the importance of logistics in regard to providing the necessary level of support capabilities, the Flagstaff area provides substantial logistics support and lodging accommodations to take care of team members during long hours of field operations.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Linenberg, A.; Lander, N.J.
1994-12-31
The need for remote monitoring of certain compounds in a sparsely populated area with limited user assistance led to the development and manufacture of a self contained, portable gas chromatography with the appropriate software. Part per billion levels of vinyl chloride, cis 1,2 dichloroethylene and trichloroethylene were detected in air using a trap for preconcentration of the compounds. The units were continuously calibrated with certified standards from Scott Specialty Gases, which in one case was 1 part per billion of the aforementioned compounds. The entire operation of the units, including monitoring instrument responses, changing operating parameters, data transfer, data reviewmore » and data reporting was done entirely on a remote basis from approximately 600 miles away using a remote computer with a modem and remote operating software. The entire system concept promises the availability of highly sensitive remote monitoring in sparsely populated areas for long periods of time.« less
[Personnel Management and Computer Systems].
ERIC Educational Resources Information Center
Reeves, Robert F.
The organization and use of computerized management information systems at the Oakland Schools intermediate school district in Michigan is utilized by 24 local school districts. The use of remote terminals provides access for the development of ongoing personnel programs. Emphasis is given to four major computer subsystems that directly involve…
Intelligent Systems: Terrestrial Observation and Prediction Using Remote Sensing Data
NASA Technical Reports Server (NTRS)
Coughlan, Joseph C.
2005-01-01
NASA has made science and technology investments to better utilize its large space-borne remote sensing data holdings of the Earth. With the launch of Terra, NASA created a data-rich environment where the challenge is to fully utilize the data collected from EOS however, despite unprecedented amounts of observed data, there is a need for increasing the frequency, resolution, and diversity of observations. Current terrestrial models that use remote sensing data were constructed in a relatively data and compute limited era and do not take full advantage of on-line learning methods and assimilation techniques that can exploit these data. NASA has invested in visualization, data mining and knowledge discovery methods which have facilitated data exploitation, but these methods are insufficient for improving Earth science models that have extensive background knowledge nor do these methods refine understanding of complex processes. Investing in interdisciplinary teams that include computational scientists can lead to new models and systems for online operation and analysis of data that can autonomously improve in prediction skill over time.
A remote sensing computer-assisted learning tool developed using the unified modeling language
NASA Astrophysics Data System (ADS)
Friedrich, J.; Karslioglu, M. O.
The goal of this work has been to create an easy-to-use and simple-to-make learning tool for remote sensing at an introductory level. Many students struggle to comprehend what seems to be a very basic knowledge of digital images, image processing and image arithmetic, for example. Because professional programs are generally too complex and overwhelming for beginners and often not tailored to the specific needs of a course regarding functionality, a computer-assisted learning (CAL) program was developed based on the unified modeling language (UML), the present standard for object-oriented (OO) system development. A major advantage of this approach is an easier transition from modeling to coding of such an application, if modern UML tools are being used. After introducing the constructed UML model, its implementation is briefly described followed by a series of learning exercises. They illustrate how the resulting CAL tool supports students taking an introductory course in remote sensing at the author's institution.
ERIC Educational Resources Information Center
Kelly, Clare
2015-01-01
This article discusses a small-scale study that explores how members of one family based in Australia and the United Kingdom use remote technology to develop and maintain family relationships across generations and distance. Of particular interest was the manner in which Skype computer software was mediated to develop intersubjectivity between a…
Scalable Trust of Next-Generation Management (STRONGMAN)
2004-10-01
remote logins might be policy controlled to allow only strongly encrypted IPSec tunnels to log in remotely, to access selected files, etc. The...and Angelos D. Keromytis. Drop-in Security for Distributed and Portable Computing Elements. Emerald Journal of Internet Research. Electronic...Security and Privacy, pp. 17-31, May 1999. [2] S. M. Bellovin. Distributed Firewalls. ; login : magazine, special issue on security, November 1999. [3] M
ERIC Educational Resources Information Center
Grober, S.; Vetter, M.; Eckert, B.; Jodl, H.-J.
2007-01-01
We suggest that different string pendulums are positioned at different locations on Earth and measure at each place the gravitational acceleration (accuracy [delta]g is approximately equal to 0.01 m s[superscript -2]). Each pendulum can be remotely controlled via the internet by a computer located somewhere on Earth. The theoretical part describes…
ADP of multispectral scanner data for land use mapping
NASA Technical Reports Server (NTRS)
Hoffer, R. M.
1971-01-01
The advantages and disadvantages of various remote sensing instrumentation and analysis techniques are reviewed. The use of multispectral scanner data and the automatic data processing techniques are considered. A computer-aided analysis system for remote sensor data is described with emphasis on the image display, statistics processor, wavelength band selection, classification processor, and results display. Advanced techniques in using spectral and temporal data are also considered.
ERIC Educational Resources Information Center
Caminero, Agustín C.; Ros, Salvador; Hernández, Roberto; Robles-Gómez, Antonio; Tobarra, Llanos; Tolbaños Granjo, Pedro J.
2016-01-01
The use of practical laboratories is a key in engineering education in order to provide our students with the resources needed to acquire practical skills. This is specially true in the case of distance education, where no physical interactions between lecturers and students take place, so virtual or remote laboratories must be used. UNED has…
Beacon data acquisition and display system
Skogmo, D.G.; Black, B.D.
1991-12-17
A system for transmitting aircraft beacon information received by a secondary surveillance radar through telephone lines to a remote display includes a digitizer connected to the radar for preparing a serial file of data records containing position and identification information of the beacons detected by each sweep of the radar. This information is transmitted through the telephone lines to a remote computer where it is displayed. 6 figures.
Beacon data acquisition and display system
Skogmo, David G.; Black, Billy D.
1991-01-01
A system for transmitting aircraft beacon information received by a secondary surveillance radar through telephone lines to a remote display includes a digitizer connected to the radar for preparing a serial file of data records containing position and identification information of the beacons detected by each sweep of the radar. This information is transmitted through the telephone lines to a remote computer where it is displayed.
Modeling of Explorative Procedures for Remote Object Identification
1991-09-01
haptic sensory system and the simulated foveal component of the visual system. Eventually it will allow multiple applications in remote sensing and...superposition of sensory channels. The use of a force reflecting telemanipulator and computer simulated visual foveal component are the tools which...representation of human search models is achieved by using the proprioceptive component of the haptic sensory system and the simulated foveal component of the
NASA Technical Reports Server (NTRS)
Talay, T. A.; Sykes, K. W.; Kuo, C. Y.
1979-01-01
On May 17, 1977, a remote sensing experiment was conducted on the James River, Virginia, whereby thermal spectrometer and near-infrared photography data of thermal discharges at Hopewell and the Surry nuclear power plant were obtained by an aircraft for one tidal cycle. These data were used in subsequent investigations into the near field discharge trajectories. For the Gravelly Run thermal plume at Hopewell, several empirical expressions for the plume centerline were evaluated by comparisons of the computed trajectories and those observed in the remote sensing images.
Remote Sensing: A valuable tool in the Forest Service decision making process. [in Utah
NASA Technical Reports Server (NTRS)
Stanton, F. L.
1975-01-01
Forest Service studies for integrating remotely sensed data into existing information systems highlight a need to: (1) re-examine present methods of collecting and organizing data, (2) develop an integrated information system for rapidly processing and interpreting data, (3) apply existing technological tools in new ways, and (4) provide accurate and timely information for making right management decisions. The Forest Service developed an integrated information system using remote sensors, microdensitometers, computer hardware and software, and interactive accessories. Their efforts substantially reduce the time it takes for collecting and processing data.
A study of Minnesota land and water resources using remote sensing, volume 13
NASA Technical Reports Server (NTRS)
1980-01-01
Progress in the use of LANDSAT data to classify wetlands in the Upper Mississippi River Valley and efforts to evaluate stress in corn and soybean crops are described. Satellite remote sensing data was used to measure particle concentrations in Lake Superior and several different kinds of remote sensing data were synergistically combined in order to identify near surface bedrock in Minnesota. Data analysis techniques which separate those activities requiring extensive computing form those involving a great deal of user interaction were developed to allow the latter to be done in the user's office or in the field.
NASA Technical Reports Server (NTRS)
Ingels, F. M.; Rives, T. B.
1987-01-01
An analytical analysis of the HOSC Generic Peripheral processing system was conducted. The results are summarized and they indicate that the maximum delay in performing screen change requests should be less than 2.5 sec., occurring for a slow VAX host to video screen I/O rate of 50 KBps. This delay is due to the average I/O rate from the video terminals to their host computer. Software structure of the main computers and the host computers will have greater impact on screen change or refresh response times. The HOSC data system model was updated by a newly coded PASCAL based simulation program which was installed on the HOSC VAX system. This model is described and documented. Suggestions are offered to fine tune the performance of the ETERNET interconnection network. Suggestions for using the Nutcracker by Excelan to trace itinerate packets which appear on the network from time to time were offered in discussions with the HOSC personnel. Several visits to the HOSC facility were to install and demonstrate the simulation model.
Computational Modeling and Real-Time Control of Patient-Specific Laser Treatment of Cancer
Fuentes, D.; Oden, J. T.; Diller, K. R.; Hazle, J. D.; Elliott, A.; Shetty, A.; Stafford, R. J.
2014-01-01
An adaptive feedback control system is presented which employs a computational model of bioheat transfer in living tissue to guide, in real-time, laser treatments of prostate cancer monitored by magnetic resonance thermal imaging (MRTI). The system is built on what can be referred to as cyberinfrastructure - a complex structure of high-speed network, large-scale parallel computing devices, laser optics, imaging, visualizations, inverse-analysis algorithms, mesh generation, and control systems that guide laser therapy to optimally control the ablation of cancerous tissue. The computational system has been successfully tested on in-vivo, canine prostate. Over the course of an 18 minute laser induced thermal therapy (LITT) performed at M.D. Anderson Cancer Center (MDACC) in Houston, Texas, the computational models were calibrated to intra-operative real time thermal imaging treatment data and the calibrated models controlled the bioheat transfer to within 5°C of the predetermined treatment plan. The computational arena is in Austin, Texas and managed at the Institute for Computational Engineering and Sciences (ICES). The system is designed to control the bioheat transfer remotely while simultaneously providing real-time remote visualization of the on-going treatment. Post operative histology of the canine prostate reveal that the damage region was within the targeted 1.2cm diameter treatment objective. PMID:19148754
Computational modeling and real-time control of patient-specific laser treatment of cancer.
Fuentes, D; Oden, J T; Diller, K R; Hazle, J D; Elliott, A; Shetty, A; Stafford, R J
2009-04-01
An adaptive feedback control system is presented which employs a computational model of bioheat transfer in living tissue to guide, in real-time, laser treatments of prostate cancer monitored by magnetic resonance thermal imaging. The system is built on what can be referred to as cyberinfrastructure-a complex structure of high-speed network, large-scale parallel computing devices, laser optics, imaging, visualizations, inverse-analysis algorithms, mesh generation, and control systems that guide laser therapy to optimally control the ablation of cancerous tissue. The computational system has been successfully tested on in vivo, canine prostate. Over the course of an 18 min laser-induced thermal therapy performed at M.D. Anderson Cancer Center (MDACC) in Houston, Texas, the computational models were calibrated to intra-operative real-time thermal imaging treatment data and the calibrated models controlled the bioheat transfer to within 5 degrees C of the predetermined treatment plan. The computational arena is in Austin, Texas and managed at the Institute for Computational Engineering and Sciences (ICES). The system is designed to control the bioheat transfer remotely while simultaneously providing real-time remote visualization of the on-going treatment. Post-operative histology of the canine prostate reveal that the damage region was within the targeted 1.2 cm diameter treatment objective.
Integration and validation of a data grid software
NASA Astrophysics Data System (ADS)
Carenton-Madiec, Nicolas; Berger, Katharina; Cofino, Antonio
2014-05-01
The Earth System Grid Federation (ESGF) Peer-to-Peer (P2P) is a software infrastructure for the management, dissemination, and analysis of model output and observational data. The ESGF grid is composed with several types of nodes which have different roles. About 40 data nodes host model outputs and datasets using thredds catalogs. About 25 compute nodes offer remote visualization and analysis tools. About 15 index nodes crawl data nodes catalogs and implement faceted and federated search in a web interface. About 15 Identity providers nodes manage accounts, authentication and authorization. Here we will present an actual size test federation spread across different institutes in different countries and a python test suite that were started in December 2013. The first objective of the test suite is to provide a simple tool that helps to test and validate a single data node and its closest index, compute and identity provider peer. The next objective will be to run this test suite on every data node of the federation and therefore test and validate every single node of the whole federation. The suite already implements nosetests, requests, myproxy-logon, subprocess, selenium and fabric python libraries in order to test both web front ends, back ends and security services. The goal of this project is to improve the quality of deliverable in a small developers team context. Developers are widely spread around the world working collaboratively and without hierarchy. This kind of working organization context en-lighted the need of a federated integration test and validation process.