Vlachogiannis, J G
2003-01-01
Taguchi's technique is a helpful tool to achieve experimental optimization of a large number of decision variables with a small number of off-line experiments. The technique appears to be an ideal tool for improving the performance of X-ray medical radiographic screens under a noise source. Currently there are very many guides available for improving the efficiency of X-ray medical radiographic screens. These guides can be refined using a second-stage parameter optimization. based on Taguchi's technique, selecting the optimum levels of controllable X-ray radiographic screen factors. A real example of the proposed technique is presented giving certain performance criteria. The present research proposes the reinforcement of X-ray radiography by Taguchi's technique as a novel hardware mechanism.
Optimization of segmented thermoelectric generator using Taguchi and ANOVA techniques.
Kishore, Ravi Anant; Sanghadasa, Mohan; Priya, Shashank
2017-12-01
Recent studies have demonstrated that segmented thermoelectric generators (TEGs) can operate over large thermal gradient and thus provide better performance (reported efficiency up to 11%) as compared to traditional TEGs, comprising of single thermoelectric (TE) material. However, segmented TEGs are still in early stages of development due to the inherent complexity in their design optimization and manufacturability. In this study, we demonstrate physics based numerical techniques along with Analysis of variance (ANOVA) and Taguchi optimization method for optimizing the performance of segmented TEGs. We have considered comprehensive set of design parameters, such as geometrical dimensions of p-n legs, height of segmentation, hot-side temperature, and load resistance, in order to optimize output power and efficiency of segmented TEGs. Using the state-of-the-art TE material properties and appropriate statistical tools, we provide near-optimum TEG configuration with only 25 experiments as compared to 3125 experiments needed by the conventional optimization methods. The effect of environmental factors on the optimization of segmented TEGs is also studied. Taguchi results are validated against the results obtained using traditional full factorial optimization technique and a TEG configuration for simultaneous optimization of power and efficiency is obtained.
Taguchi optimization of bismuth-telluride based thermoelectric cooler
NASA Astrophysics Data System (ADS)
Anant Kishore, Ravi; Kumar, Prashant; Sanghadasa, Mohan; Priya, Shashank
2017-07-01
In the last few decades, considerable effort has been made to enhance the figure-of-merit (ZT) of thermoelectric (TE) materials. However, the performance of commercial TE devices still remains low due to the fact that the module figure-of-merit not only depends on the material ZT, but also on the operating conditions and configuration of TE modules. This study takes into account comprehensive set of parameters to conduct the numerical performance analysis of the thermoelectric cooler (TEC) using a Taguchi optimization method. The Taguchi method is a statistical tool that predicts the optimal performance with a far less number of experimental runs than the conventional experimental techniques. Taguchi results are also compared with the optimized parameters obtained by a full factorial optimization method, which reveals that the Taguchi method provides optimum or near-optimum TEC configuration using only 25 experiments against 3125 experiments needed by the conventional optimization method. This study also shows that the environmental factors such as ambient temperature and cooling coefficient do not significantly affect the optimum geometry and optimum operating temperature of TECs. The optimum TEC configuration for simultaneous optimization of cooling capacity and coefficient of performance is also provided.
Modified Mahalanobis Taguchi System for Imbalance Data Classification
2017-01-01
The Mahalanobis Taguchi System (MTS) is considered one of the most promising binary classification algorithms to handle imbalance data. Unfortunately, MTS lacks a method for determining an efficient threshold for the binary classification. In this paper, a nonlinear optimization model is formulated based on minimizing the distance between MTS Receiver Operating Characteristics (ROC) curve and the theoretical optimal point named Modified Mahalanobis Taguchi System (MMTS). To validate the MMTS classification efficacy, it has been benchmarked with Support Vector Machines (SVMs), Naive Bayes (NB), Probabilistic Mahalanobis Taguchi Systems (PTM), Synthetic Minority Oversampling Technique (SMOTE), Adaptive Conformal Transformation (ACT), Kernel Boundary Alignment (KBA), Hidden Naive Bayes (HNB), and other improved Naive Bayes algorithms. MMTS outperforms the benchmarked algorithms especially when the imbalance ratio is greater than 400. A real life case study on manufacturing sector is used to demonstrate the applicability of the proposed model and to compare its performance with Mahalanobis Genetic Algorithm (MGA). PMID:28811820
Investigating the effects of PDC cutters geometry on ROP using the Taguchi technique
NASA Astrophysics Data System (ADS)
Jamaludin, A. A.; Mehat, N. M.; Kamaruddin, S.
2017-10-01
At times, the polycrystalline diamond compact (PDC) bit’s performance dropped and affects the rate of penetration (ROP). The objective of this project is to investigate the effect of PDC cutter geometry and optimize them. An intensive study in cutter geometry would further enhance the ROP performance. The relatively extended analysis was carried out and four significant geometry factors have been identified that directly improved the ROP. Cutter size, back rake angle, side rake angle and chamfer angle are the stated geometry factors. An appropriate optimization technique that effectively controls all influential geometry factors during cutters manufacturing is introduced and adopted in this project. By adopting L9 Taguchi OA, simulation experiment is conducted by using explicit dynamics finite element analysis. Through a structure Taguchi analysis, ANOVA confirms that the most significant geometry to improve ROP is cutter size (99.16% percentage contribution). The optimized cutter is expected to drill with high ROP that can reduce the rig time, which in its turn, may reduce the total drilling cost.
Simulation reduction using the Taguchi method
NASA Technical Reports Server (NTRS)
Mistree, Farrokh; Lautenschlager, Ume; Erikstad, Stein Owe; Allen, Janet K.
1993-01-01
A large amount of engineering effort is consumed in conducting experiments to obtain information needed for making design decisions. Efficiency in generating such information is the key to meeting market windows, keeping development and manufacturing costs low, and having high-quality products. The principal focus of this project is to develop and implement applications of Taguchi's quality engineering techniques. In particular, we show how these techniques are applied to reduce the number of experiments for trajectory simulation of the LifeSat space vehicle. Orthogonal arrays are used to study many parameters simultaneously with a minimum of time and resources. Taguchi's signal to noise ratio is being employed to measure quality. A compromise Decision Support Problem and Robust Design are applied to demonstrate how quality is designed into a product in the early stages of designing.
Wear behavior of electroless Ni-P-W coating under lubricated condition - a Taguchi based approach
NASA Astrophysics Data System (ADS)
Mukhopadhyay, Arkadeb; Duari, Santanu; Barman, Tapan Kumar; Sahoo, Prasanta
2016-09-01
The present study aims to investigate the tribological behavior of electroless Ni-P-W coating under engine oil lubricated condition to ascertain its suitability in automotive applications. Coating is deposited onto mild steel specimens by the electroless method. The experiments are carried out on a pin - on - disc type tribo tester under lubrication. Three tribotesting parameters namely the applied normal load, sliding speed and sliding duration are varied at their three levels and their effects on the wear depth of the deposits are studied. The experiments are carried out based on the combinations available in Taguchi's L27 orthogonal array (OA). Optimization of the tribo-testing parameters is carried out using Taguchi's S/N ratio method to minimize the wear depth. Analysis of variance carried out at a confidence level of 99% indicates that the sliding speed is the most significant parameter in controlling the wear behavior of the deposits. Coating characterization is done using scanning electron microscope, energy dispersive X-ray analysis and X-ray diffraction techniques. It is seen that the wear mechanism under lubricated condition is abrasive in nature.
Study of Dimple Effect on the Friction Characteristics of a Journal Bearing using Taguchi Method
NASA Astrophysics Data System (ADS)
Murthy, A. Amar; Raghunandana, Dr.
2018-02-01
The effect of producing dimples using chemically etched techniques or by machining process on the surface of a journal bearing bushing to reduce the friction using Taguchi method is investigated. The data used in the present analysis is based on the results obtained by the series of experiments conducted to study the dimples effect on the Stribeck curve. It is statistically proved that producing dimples on the bushing surface of a journal bearing has significant effect on the friction coefficient when used with light oils. Also it is seen that there is an interaction effect between speeds-load and load-dimples. Hence the interaction effect, which are usually neglected should be considered during actual experiments that significantly contributes in reducing the friction in mixed lubrication regime. The experiments, if were conducted after Taguchi method, then the number of experiments would have been reduced to half of the actual set of experiments that were essentially conducted.
Assessing the applicability of the Taguchi design method to an interrill erosion study
NASA Astrophysics Data System (ADS)
Zhang, F. B.; Wang, Z. L.; Yang, M. Y.
2015-02-01
Full-factorial experimental designs have been used in soil erosion studies, but are time, cost and labor intensive, and sometimes they are impossible to conduct due to the increasing number of factors and their levels to consider. The Taguchi design is a simple, economical and efficient statistical tool that only uses a portion of the total possible factorial combinations to obtain the results of a study. Soil erosion studies that use the Taguchi design are scarce and no comparisons with full-factorial designs have been made. In this paper, a series of simulated rainfall experiments using a full-factorial design of five slope lengths (0.4, 0.8, 1.2, 1.6, and 2 m), five slope gradients (18%, 27%, 36%, 48%, and 58%), and five rainfall intensities (48, 62.4, 102, 149, and 170 mm h-1) were conducted. Validation of the applicability of a Taguchi design to interrill erosion experiments was achieved by extracting data from the full dataset according to a theoretical Taguchi design. The statistical parameters for the mean quasi-steady state erosion and runoff rates of each test, the optimum conditions for producing maximum erosion and runoff, and the main effect and percentage contribution of each factor obtained from the full-factorial and Taguchi designs were compared. Both designs generated almost identical results. Using the experimental data from the Taguchi design, it was possible to accurately predict the erosion and runoff rates under the conditions that had been excluded from the Taguchi design. All of the results obtained from analyzing the experimental data for both designs indicated that the Taguchi design could be applied to interrill erosion studies and could replace full-factorial designs. This would save time, labor and costs by generally reducing the number of tests to be conducted. Further work should test the applicability of the Taguchi design to a wider range of conditions.
Multi-Response Optimization of Resin Finishing by Using a Taguchi-Based Grey Relational Analysis
Shafiq, Faizan; Sarwar, Zahid; Jilani, Muhammad Munib; Cai, Yingjie
2018-01-01
In this study, the influence and optimization of the factors of a non-formaldehyde resin finishing process on cotton fabric using a Taguchi-based grey relational analysis were experimentally investigated. An L27 orthogonal array was selected for five parameters and three levels by applying Taguchi’s design of experiments. The Taguchi technique was coupled with a grey relational analysis to obtain a grey relational grade for evaluating multiple responses, i.e., crease recovery angle (CRA), tearing strength (TE), and whiteness index (WI). The optimum parameters (values) for resin finishing were the resin concentration (80 g·L−1), the polyethylene softener (40 g·L−1), the catalyst (25 g·L−1), the curing temperature (140 °C), and the curing time (2 min). The goodness-of-fit of the data was validated by an analysis of variance (ANOVA). The optimized sample was characterized by Fourier-transform infrared (FTIR) spectroscopy, thermogravimetric analysis (TGA), and scanning electron microscope (SEM) to better understand the structural details of the resin finishing process. The results showed an improved thermal stability and confirmed the presence of well deposited of resin on the optimized fabric surface. PMID:29543724
A Taguchi study of the aeroelastic tailoring design process
NASA Technical Reports Server (NTRS)
Bohlmann, Jonathan D.; Scott, Robert C.
1991-01-01
A Taguchi study was performed to determine the important players in the aeroelastic tailoring design process and to find the best composition of the optimization's objective function. The Wing Aeroelastic Synthesis Procedure (TSO) was used to ascertain the effects that factors such as composite laminate constraints, roll effectiveness constraints, and built-in wing twist and camber have on the optimum, aeroelastically tailored wing skin design. The results show the Taguchi method to be a viable engineering tool for computational inquiries, and provide some valuable lessons about the practice of aeroelastic tailoring.
Application of Taguchi methods to infrared window design
NASA Astrophysics Data System (ADS)
Osmer, Kurt A.; Pruszynski, Charles J.
1990-10-01
Dr. Genichi Taguchi, a prominent quality consultant, reduced a branch of statistics known as "Design of Experiments" to a cookbook methodology that can be employed by any competent engineer. This technique has been extensively employed by Japanese manufacturers, and is widely credited with helping them attain their current level of success in low cost, high quality product design and fabrication. Although this technique was originally put forth as a tool to streamline the determination of improved production processes, it can also be applied to a wide range of engineering problems. As part of an internal research project, this method of experimental design has been adapted to window trade studies and materials research. Two of these analyses are presented herein, and have been chosen to illustrate the breadth of applications to which the Taguchi method can be utilized.
Incorporating Servqual-QFD with Taguchi Design for optimizing service quality design
NASA Astrophysics Data System (ADS)
Arbi Hadiyat, M.
2018-03-01
Deploying good service design in service companies has been updated issue in improving customer satisfaction, especially based on the level of service quality measured by Parasuraman’s SERVQUAL. Many researchers have been proposing methods in designing the service, and some of them are based on engineering viewpoint, especially by implementing the QFD method or even using robust Taguchi method. The QFD method would found the qualitative solution by generating the “how’s”, while Taguchi method gives more quantitative calculation in optimizing best solution. However, incorporating both QFD and Taguchi has been done in this paper and yields better design process. The purposes of this research is to evaluate the incorporated methods by implemented it to a case study, then analyze the result and see the robustness of those methods to customer perception of service quality. Started by measuring service attributes using SERVQUAL and find the improvement with QFD, the deployment of QFD solution then generated by defining Taguchi factors levels and calculating the Signal-to-noise ratio in its orthogonal array, and optimized Taguchi response then found. A case study was given for designing service in local bank. Afterward, the service design obtained from previous analysis was then evaluated and shows that it was still meet the customer satisfaction. Incorporating QFD and Taguchi has performed well and can be adopted and developed for another research for evaluating the robustness of result.
The parameters effect on the structural performance of damaged steel box beam using Taguchi method
NASA Astrophysics Data System (ADS)
El-taly, Boshra A.; Abd El Hameed, Mohamed F.
2018-03-01
In the current study, the influence of notch or opening parameters and the positions of the applied load on the structural performance of steel box beams up to failure was investigated using Finite Element analysis program, ANSYS. The Taguchi-based design of experiments technique was used to plan the current study. The plan included 12 box steel beams; three intact beams, and nine damaged beams (with opening) in the beams web. The numerical studies were conducted under varying the spacing between the two concentrated point loads (location of applied loads), the notch (opening) position, and the ratio between depth and width of the notch with a constant notch area. According to Taguchi analysis, factor X (location of the applied loads) was found the highest contributing parameters for the variation of the ultimate load, vertical deformation, shear stresses, and the compressive normal stresses.
A comparative study of electrochemical machining process parameters by using GA and Taguchi method
NASA Astrophysics Data System (ADS)
Soni, S. K.; Thomas, B.
2017-11-01
In electrochemical machining quality of machined surface strongly depend on the selection of optimal parameter settings. This work deals with the application of Taguchi method and genetic algorithm using MATLAB to maximize the metal removal rate and minimize the surface roughness and overcut. In this paper a comparative study is presented for drilling of LM6 AL/B4C composites by comparing the significant impact of numerous machining process parameters such as, electrolyte concentration (g/l),machining voltage (v),frequency (hz) on the response parameters (surface roughness, material removal rate and over cut). Taguchi L27 orthogonal array was chosen in Minitab 17 software, for the investigation of experimental results and also multiobjective optimization done by genetic algorithm is employed by using MATLAB. After obtaining optimized results from Taguchi method and genetic algorithm, a comparative results are presented.
Multidisciplinary design of a rocket-based combined cycle SSTO launch vehicle using Taguchi methods
NASA Technical Reports Server (NTRS)
Olds, John R.; Walberg, Gerald D.
1993-01-01
Results are presented from the optimization process of a winged-cone configuration SSTO launch vehicle that employs a rocket-based ejector/ramjet/scramjet/rocket operational mode variable-cycle engine. The Taguchi multidisciplinary parametric-design method was used to evaluate the effects of simultaneously changing a total of eight design variables, rather than changing them one at a time as in conventional tradeoff studies. A combination of design variables was in this way identified which yields very attractive vehicle dry and gross weights.
Taguchi method for partial differential equations with application in tumor growth.
Ilea, M; Turnea, M; Rotariu, M; Arotăriţei, D; Popescu, Marilena
2014-01-01
The growth of tumors is a highly complex process. To describe this process, mathematical models are needed. A variety of partial differential mathematical models for tumor growth have been developed and studied. Most of those models are based on the reaction-diffusion equations and mass conservation law. A variety of modeling strategies have been developed, each focusing on tumor growth. Systems of time-dependent partial differential equations occur in many branches of applied mathematics. The vast majority of mathematical models in tumor growth are formulated in terms of partial differential equations. We propose a mathematical model for the interactions between these three cancer cell populations. The Taguchi methods are widely used by quality engineering scientists to compare the effects of multiple variables, together with their interactions, with a simple and manageable experimental design. In Taguchi's design of experiments, variation is more interesting to study than the average. First, Taguchi methods are utilized to search for the significant factors and the optimal level combination of parameters. Except the three parameters levels, other factors levels other factors levels would not be considered. Second, cutting parameters namely, cutting speed, depth of cut, and feed rate are designed using the Taguchi method. Finally, the adequacy of the developed mathematical model is proved by ANOVA. According to the results of ANOVA, since the percentage contribution of the combined error is as small. Many mathematical models can be quantitatively characterized by partial differential equations. The use of MATLAB and Taguchi method in this article illustrates the important role of informatics in research in mathematical modeling. The study of tumor growth cells is an exciting and important topic in cancer research and will profit considerably from theoretical input. Interpret these results to be a permanent collaboration between math's and medical oncologists.
NASA Astrophysics Data System (ADS)
Ţîţu, M. A.; Pop, A. B.; Ţîţu, Ș
2017-06-01
This paper presents a study on the modelling and optimization of certain variables by using the Taguchi Method with a view to modelling and optimizing the process of pressing tappets into anchors, process conducted in an organization that promotes knowledge-based management. The paper promotes practical concepts of the Taguchi Method and describes the way in which the objective functions are obtained and used during the modelling and optimization of the process of pressing tappets into the anchors.
NASA Astrophysics Data System (ADS)
Pourbaghi-Masouleh, M.; Asgharzadeh, H.
2013-08-01
In this study, the Taguchi method of design of experiment (DOE) was used to optimize the hydroxyapatite (HA) coatings on various metallic substrates deposited by sol-gel dip-coating technique. The experimental design consisted of five factors including substrate material (A), surface preparation of substrate (B), dipping/withdrawal speed (C), number of layers (D), and calcination temperature (E) with three levels of each factor. An orthogonal array of L18 type with mixed levels of the control factors was utilized. The image processing of the micrographs of the coatings was conducted to determine the percentage of coated area ( PCA). Chemical and phase composition of HA coatings were studied by XRD, FT-IR, SEM, and EDS techniques. The analysis of variance (ANOVA) indicated that the PCA of HA coatings was significantly affected by the calcination temperature. The optimum conditions from signal-to-noise ( S/N) ratio analysis were A: pure Ti, B: polishing and etching for 24 h, C: 50 cm min-1, D: 1, and E: 300 °C. In the confirmation experiment using the optimum conditions, the HA coating with high PCA of 98.5 % was obtained.
Optimization of porthole die geometrical variables by Taguchi method
NASA Astrophysics Data System (ADS)
Gagliardi, F.; Ciancio, C.; Ambrogio, G.; Filice, L.
2017-10-01
Porthole die extrusion is commonly used to manufacture hollow profiles made of lightweight alloys for numerous industrial applications. The reliability of extruded parts is affected strongly by the quality of the longitudinal and transversal seam welds. According to that, the die geometry must be designed correctly and the process parameters must be selected properly to achieve the desired product quality. In this study, numerical 3D simulations have been created and run to investigate the role of various geometrical variables on punch load and maximum pressure inside the welding chamber. These are important outputs to take into account affecting, respectively, the necessary capacity of the extrusion press and the quality of the welding lines. The Taguchi technique has been used to reduce the number of the required numerical simulations necessary for considering the influence of twelve different geometric variables. Moreover, the Analysis of variance (ANOVA) has been implemented to individually analyze the effect of each input parameter on the two responses. Then, the methodology has been utilized to determine the optimal process configuration individually optimizing the two investigated process outputs. Finally, the responses of the optimized parameters have been verified through finite element simulations approximating the predicted value closely. This study shows the feasibility of the Taguchi technique for predicting performance, optimization and therefore for improving the design of a porthole extrusion process.
NASA Astrophysics Data System (ADS)
Maleque, M. A.; Bello, K. A.; Adebisi, A. A.; Akma, N.
2017-03-01
Tungsten inert gas (TIG) torch is one of the most recently used heat source for surface modification of engineering parts, giving similar results to the more expensive high power laser technique. In this study, ceramic-based embedded composite coating has been produced by precoated silicon carbide (SiC) powders on the AISI 4340 low alloy steel substrate using TIG welding torch process. A design of experiment based on Taguchi approach has been adopted to optimize the TIG cladding process parameters. The L9 orthogonal array and the signal-to-noise was used to study the effect of TIG welding parameters such as arc current, travelling speed, welding voltage and argon flow rate on tribological response behaviour (wear rate, surface roughness and wear track width). The objective of the study was to identify optimal design parameter that significantly minimizes each of the surface quality characteristics. The analysis of the experimental results revealed that the argon flow rate was found to be the most influential factor contributing to the minimum wear and surface roughness of the modified coating surface. On the other hand, the key factor in reducing wear scar is the welding voltage. Finally, a convenient and economical Taguchi approach used in this study was efficient to find out optimal factor settings for obtaining minimum wear rate, wear scar and surface roughness responses in TIG-coated surfaces.
Sadeghi, Maryam; Faghihi, Reza; Sina, Sedigheh
2017-06-15
Thermoluminescence dosimetry (TLD) is a powerful technique with wide applications in personal, environmental and clinical dosimetry. The optimum annealing, storage and reading protocols are very effective in accuracy of TLD response. The purpose of this study is to obtain an optimum protocol for GR-200; LiF: Mg, Cu, P, by optimizing the effective parameters, to increase the reliability of the TLD response using Taguchi method. Taguchi method has been used in this study for optimization of annealing, storage and reading protocols of the TLDs. A number of 108 GR-200 chips were divided into 27 groups, each containing four chips. The TLDs were exposed to three different doses, and stored, annealed and read out by different procedures as suggested by Taguchi Method. By comparing the signal-to-noise ratios the optimum dosimetry procedure was obtained. According to the results, the optimum values for annealing temperature (°C), Annealing Time (s), Annealing to Exposure time (d), Exposure to Readout time (d), Pre-heat Temperature (°C), Pre-heat Time (s), Heating Rate (°C/s), Maximum Temperature of Readout (°C), readout time (s) and Storage Temperature (°C) are 240, 90, 1, 2, 50, 0, 15, 240, 13 and -20, respectively. Using the optimum protocol, an efficient glow curve with low residual signals can be achieved. Using optimum protocol obtained by Taguchi method, the dosimetry can be effectively performed with great accuracy. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
NASA Astrophysics Data System (ADS)
Rastkerdar, E.; Shamanian, M.; Saatchi, A.
2013-04-01
In this study, the Taguchi method was used as a design of experiment (DOE) technique to optimize the pulsed current gas tungsten arc welding (GTAW) parameters for improved pitting corrosion resistance of AA5083-H18 aluminum alloy welds. A L9 (34) orthogonal array of the Taguchi design was used, which involves nine experiments for four parameters: peak current ( P), base current ( B), percent pulse-on time ( T), and pulse frequency ( F) with three levels was used. Pitting corrosion resistance in 3.5 wt.% NaCl solution was evaluated by anodic polarization tests at room temperature and calculating the width of the passive region (∆ E pit). Analysis of variance (ANOVA) was performed on the measured data and S/ N (signal to noise) ratios. The "bigger is better" was selected as the quality characteristic (QC). The optimum conditions were found as 170 A, 85 A, 40%, and 6 Hz for P, B, T, and F factors, respectively. The study showed that the percent pulse-on time has the highest influence on the pitting corrosion resistance (50.48%) followed by pulse frequency (28.62%), peak current (11.05%) and base current (9.86%). The range of optimum ∆ E pit at optimum conditions with a confidence level of 90% was predicted to be between 174.81 and 177.74 mVSCE. Under optimum conditions, the confirmation test was carried out, and the experimental value of ∆ E pit of 176 mVSCE was in agreement with the predicted value from the Taguchi model. In this regard, the model can be effectively used to predict the ∆ E pit of pulsed current gas tungsten arc welded joints.
Optimization of bone drilling parameters using Taguchi method based on finite element analysis
NASA Astrophysics Data System (ADS)
Rosidi, Ayip; Lenggo Ginta, Turnad; Rani, Ahmad Majdi Bin Abdul
2017-05-01
Thermal necrosis results fracture problems and implant failure if temperature exceeds 47 °C for one minute during bone drilling. To solve this problem, this work studied a new thermal model by using three drilling parameters: drill diameter, feed rate and spindle speed. Effects of those parameters to heat generation were studied. The drill diameters were 4 mm, 6 mm and 6 mm; the feed rates were 80 mm/min, 100 mm/min and 120 mm/min whereas the spindle speeds were 400 rpm, 500 rpm and 600 rpm then an optimization was done by Taguchi method to which combination parameter can be used to prevent thermal necrosis during bone drilling. The results showed that all the combination of parameters produce confidence results which were below 47 °C and finite element analysis combined with Taguchi method can be used for predicting temperature generation and optimizing bone drilling parameters prior to clinical bone drilling. All of the combination parameters can be used for surgeon to achieve sustainable orthopaedic surgery.
SVM-RFE based feature selection and Taguchi parameters optimization for multiclass SVM classifier.
Huang, Mei-Ling; Hung, Yung-Hsiang; Lee, W M; Li, R K; Jiang, Bo-Ru
2014-01-01
Recently, support vector machine (SVM) has excellent performance on classification and prediction and is widely used on disease diagnosis or medical assistance. However, SVM only functions well on two-group classification problems. This study combines feature selection and SVM recursive feature elimination (SVM-RFE) to investigate the classification accuracy of multiclass problems for Dermatology and Zoo databases. Dermatology dataset contains 33 feature variables, 1 class variable, and 366 testing instances; and the Zoo dataset contains 16 feature variables, 1 class variable, and 101 testing instances. The feature variables in the two datasets were sorted in descending order by explanatory power, and different feature sets were selected by SVM-RFE to explore classification accuracy. Meanwhile, Taguchi method was jointly combined with SVM classifier in order to optimize parameters C and γ to increase classification accuracy for multiclass classification. The experimental results show that the classification accuracy can be more than 95% after SVM-RFE feature selection and Taguchi parameter optimization for Dermatology and Zoo databases.
SVM-RFE Based Feature Selection and Taguchi Parameters Optimization for Multiclass SVM Classifier
Huang, Mei-Ling; Hung, Yung-Hsiang; Lee, W. M.; Li, R. K.; Jiang, Bo-Ru
2014-01-01
Recently, support vector machine (SVM) has excellent performance on classification and prediction and is widely used on disease diagnosis or medical assistance. However, SVM only functions well on two-group classification problems. This study combines feature selection and SVM recursive feature elimination (SVM-RFE) to investigate the classification accuracy of multiclass problems for Dermatology and Zoo databases. Dermatology dataset contains 33 feature variables, 1 class variable, and 366 testing instances; and the Zoo dataset contains 16 feature variables, 1 class variable, and 101 testing instances. The feature variables in the two datasets were sorted in descending order by explanatory power, and different feature sets were selected by SVM-RFE to explore classification accuracy. Meanwhile, Taguchi method was jointly combined with SVM classifier in order to optimize parameters C and γ to increase classification accuracy for multiclass classification. The experimental results show that the classification accuracy can be more than 95% after SVM-RFE feature selection and Taguchi parameter optimization for Dermatology and Zoo databases. PMID:25295306
Taguchi method of experimental design in materials education
NASA Technical Reports Server (NTRS)
Weiser, Martin W.
1993-01-01
Some of the advantages and disadvantages of the Taguchi Method of experimental design as applied to Materials Science will be discussed. This is a fractional factorial method that employs the minimum number of experimental trials for the information obtained. The analysis is also very simple to use and teach, which is quite advantageous in the classroom. In addition, the Taguchi loss function can be easily incorporated to emphasize that improvements in reproducibility are often at least as important as optimization of the response. The disadvantages of the Taguchi Method include the fact that factor interactions are normally not accounted for, there are zero degrees of freedom if all of the possible factors are used, and randomization is normally not used to prevent environmental biasing. In spite of these disadvantages it is felt that the Taguchi Method is extremely useful for both teaching experimental design and as a research tool, as will be shown with a number of brief examples.
Azadeh, Ali; Sheikhalishahi, Mohammad
2015-06-01
A unique framework for performance optimization of generation companies (GENCOs) based on health, safety, environment, and ergonomics (HSEE) indicators is presented. To rank this sector of industry, the combination of data envelopment analysis (DEA), principal component analysis (PCA), and Taguchi are used for all branches of GENCOs. These methods are applied in an integrated manner to measure the performance of GENCO. The preferred model between DEA, PCA, and Taguchi is selected based on sensitivity analysis and maximum correlation between rankings. To achieve the stated objectives, noise is introduced into input data. The results show that Taguchi outperforms other methods. Moreover, a comprehensive experiment is carried out to identify the most influential factor for ranking GENCOs. The approach developed in this study could be used for continuous assessment and improvement of GENCO's performance in supplying energy with respect to HSEE factors. The results of such studies would help managers to have better understanding of weak and strong points in terms of HSEE factors.
NASA Astrophysics Data System (ADS)
Soni, Sourabh Kumar; Thomas, Benedict
2018-04-01
The term "weldability" has been used to describe a wide variety of characteristics when a material is subjected to welding. In our analysis we perform experimental investigation to estimate the tensile strength of welded joint strength and then optimization of welding process parameters by using taguchi method and Artificial Neural Network (ANN) tool in MINITAB and MATLAB software respectively. The study reveals the influence on weldability of steel by varying composition of steel by mechanical characterization. At first we prepare the samples of different grades of steel (EN8, EN 19, EN 24). The samples were welded together by metal inert gas welding process and then tensile testing on Universal testing machine (UTM) was conducted for the same to evaluate the tensile strength of the welded steel specimens. Further comparative study was performed to find the effects of welding parameter on quality of weld strength by employing Taguchi method and Neural Network tool. Finally we concluded that taguchi method and Neural Network Tool is much efficient technique for optimization.
Permeability Evaluation Through Chitosan Membranes Using Taguchi Design
Sharma, Vipin; Marwaha, Rakesh Kumar; Dureja, Harish
2010-01-01
In the present study, chitosan membranes capable of imitating permeation characteristics of diclofenac diethylamine across animal skin were prepared using cast drying method. The effect of concentration of chitosan, concentration of cross-linking agent (NaTPP), crosslinking time was studied using Taguchi design. Taguchi design ranked concentration of chitosan as the most important factor influencing the permeation parameters of diclofenac diethylamine. The flux of the diclofenac diethylamine solution through optimized chitosan membrane (T9) was found to be comparable to that obtained across rat skin. The mathematical model developed using multilinear regression analysis can be used to formulate chitosan membranes that can mimic the desired permeation characteristics. The developed chitosan membranes can be utilized as a substitute to animal skin for in vitro permeation studies. PMID:21179329
Permeability evaluation through chitosan membranes using taguchi design.
Sharma, Vipin; Marwaha, Rakesh Kumar; Dureja, Harish
2010-01-01
In the present study, chitosan membranes capable of imitating permeation characteristics of diclofenac diethylamine across animal skin were prepared using cast drying method. The effect of concentration of chitosan, concentration of cross-linking agent (NaTPP), crosslinking time was studied using Taguchi design. Taguchi design ranked concentration of chitosan as the most important factor influencing the permeation parameters of diclofenac diethylamine. The flux of the diclofenac diethylamine solution through optimized chitosan membrane (T9) was found to be comparable to that obtained across rat skin. The mathematical model developed using multilinear regression analysis can be used to formulate chitosan membranes that can mimic the desired permeation characteristics. The developed chitosan membranes can be utilized as a substitute to animal skin for in vitro permeation studies.
NASA Astrophysics Data System (ADS)
Singh, Rupinder
2018-02-01
Hot chamber (HC) die casting process is one of the most widely used commercial processes for the casting of low temperature metals and alloys. This process gives near-net shape product with high dimensional accuracy. However in actual field environment the best settings of input parameters is often conflicting as the shape and size of the casting changes and one have to trade off among various output parameters like hardness, dimensional accuracy, casting defects, microstructure etc. So for online inspection of the cast components properties (without affecting the production line) the weight measurement has been established as one of the cost effective method (as the difference in weight of sound and unsound casting reflects the possible casting defects) in field environment. In the present work at first stage the effect of three input process parameters (namely: pressure at 2nd phase in HC die casting; metal pouring temperature and die opening time) has been studied for optimizing the cast component weight `W' as output parameter in form of macro model based upon Taguchi L9 OA. After this Buckingham's π approach has been applied on Taguchi based macro model for the development of micro model. This study highlights the Taguchi-Buckingham based combined approach as a case study (for conversion of macro model into micro model) by identification of optimum levels of input parameters (based on Taguchi approach) and development of mathematical model (based on Buckingham's π approach). Finally developed mathematical model can be used for predicting W in HC die casting process with more flexibility. The results of study highlights second degree polynomial equation for predicting cast component weight in HC die casting and suggest that pressure at 2nd stage is one of the most contributing factors for controlling the casting defect/weight of casting.
Rapid development of xylanase assay conditions using Taguchi methodology.
Prasad Uday, Uma Shankar; Bandyopadhyay, Tarun Kanti; Bhunia, Biswanath
2016-11-01
The present investigation is mainly concerned with the rapid development of extracellular xylanase assay conditions by using Taguchi methodology. The extracellular xylanase was produced from Aspergillus niger (KP874102.1), a new strain isolated from a soil sample of the Baramura forest, Tripura West, India. Four physical parameters including temperature, pH, buffer concentration and incubation time were considered as key factors for xylanase activity and were optimized using Taguchi robust design methodology for enhanced xylanase activity. The main effect, interaction effects and optimal levels of the process factors were determined using signal-to-noise (S/N) ratio. The Taguchi method recommends the use of S/N ratio to measure quality characteristics. Based on analysis of the S/N ratio, optimal levels of the process factors were determined. Analysis of variance (ANOVA) was performed to evaluate statistically significant process factors. ANOVA results showed that temperature contributed the maximum impact (62.58%) on xylanase activity, followed by pH (22.69%), buffer concentration (9.55%) and incubation time (5.16%). Predicted results showed that enhanced xylanase activity (81.47%) can be achieved with pH 2, temperature 50°C, buffer concentration 50 Mm and incubation time 10 min.
NASA Astrophysics Data System (ADS)
Sudibyo, Aji, B. B.; Sumardi, S.; Mufakir, F. R.; Junaidi, A.; Nurjaman, F.; Karna, Aziza, Aulia
2017-01-01
Gold amalgamation process was widely used to treat gold ore. This process produces the tailing or amalgamation solid waste, which still contains gold at 8-9 ppm. Froth flotation is one of the promising methods to beneficiate gold from this tailing. However, this process requires optimal conditions which depends on the type of raw material. In this study, Taguchi method was used to optimize the optimum conditions of the froth flotation process. The Taguchi optimization shows that the gold recovery was strongly influenced by the particle size which is the best particle size at 150 mesh followed by the Potassium amyl xanthate concentration, pH and pine oil concentration at 1133.98, 4535.92 and 68.04 gr/ton amalgamation tailing, respectively.
Azadeh, Ali; Sheikhalishahi, Mohammad
2014-01-01
Background A unique framework for performance optimization of generation companies (GENCOs) based on health, safety, environment, and ergonomics (HSEE) indicators is presented. Methods To rank this sector of industry, the combination of data envelopment analysis (DEA), principal component analysis (PCA), and Taguchi are used for all branches of GENCOs. These methods are applied in an integrated manner to measure the performance of GENCO. The preferred model between DEA, PCA, and Taguchi is selected based on sensitivity analysis and maximum correlation between rankings. To achieve the stated objectives, noise is introduced into input data. Results The results show that Taguchi outperforms other methods. Moreover, a comprehensive experiment is carried out to identify the most influential factor for ranking GENCOs. Conclusion The approach developed in this study could be used for continuous assessment and improvement of GENCO's performance in supplying energy with respect to HSEE factors. The results of such studies would help managers to have better understanding of weak and strong points in terms of HSEE factors. PMID:26106505
NASA Astrophysics Data System (ADS)
Kulkarni, R. D.; Agarwal, Vivek
2008-08-01
An ion chamber amplifier (ICA) is used as a safety device for neutronic power (flux) measurement in regulation and protection systems of nuclear reactors. Therefore, performance reliability of an ICA is an important issue. Appropriate quality engineering is essential to achieve a robust design and performance of the ICA circuit. It is observed that the low input bias current operational amplifiers used in the input stage of the ICA circuit are the most critical devices for proper functioning of the ICA. They are very sensitive to the gamma radiation present in their close vicinity. Therefore, the response of the ICA deteriorates with exposure to gamma radiation resulting in a decrease in the overall reliability, unless desired performance is ensured under all conditions. This paper presents a performance enhancement scheme for an ICA operated in the nuclear environment. The Taguchi method, which is a proven technique for reliability enhancement, has been used in this work. It is demonstrated that if a statistical, optimal design approach, like the Taguchi method is used, the cost of high quality and reliability may be brought down drastically. The complete methodology and statistical calculations involved are presented, as are the experimental and simulation results to arrive at a robust design of the ICA.
NASA Astrophysics Data System (ADS)
Venkata Subbaiah, K.; Raju, Ch.; Suresh, Ch.
2017-08-01
The present study aims to compare the conventional cutting inserts with wiper cutting inserts during the hard turning of AISI 4340 steel at different workpiece hardness. Type of insert, hardness, cutting speed, feed, and depth of cut are taken as process parameters. Taguchi’s L18 orthogonal array was used to conduct the experimental tests. Parametric analysis carried in order to know the influence of each process parameter on the three important Surface Roughness Characteristics (Ra, Rz, and Rt) and Material Removal Rate. Taguchi based Grey Relational Analysis (GRA) used to optimize the process parameters for individual response and multi-response outputs. Additionally, the analysis of variance (ANOVA) is also applied to identify the most significant factor.
Ozdemir, Utkan; Ozbay, Bilge; Ozbay, Ismail; Veli, Sevil
2014-09-01
In this work, Taguchi L32 experimental design was applied to optimize biosorption of Cu(2+) ions by an easily available biosorbent, Spaghnum moss. With this aim, batch biosorption tests were performed to achieve targeted experimental design with five factors (concentration, pH, biosorbent dosage, temperature and agitation time) at two different levels. Optimal experimental conditions were determined by calculated signal-to-noise ratios. "Higher is better" approach was followed to calculate signal-to-noise ratios as it was aimed to obtain high metal removal efficiencies. The impact ratios of factors were determined by the model. Within the study, Cu(2+) biosorption efficiencies were also predicted by using Taguchi method. Results of the model showed that experimental and predicted values were close to each other demonstrating the success of Taguchi approach. Furthermore, thermodynamic, isotherm and kinetic studies were performed to explain the biosorption mechanism. Calculated thermodynamic parameters were in good accordance with the results of Taguchi model. Copyright © 2014 Elsevier Inc. All rights reserved.
A Gradient Taguchi Method for Engineering Optimization
NASA Astrophysics Data System (ADS)
Hwang, Shun-Fa; Wu, Jen-Chih; He, Rong-Song
2017-10-01
To balance the robustness and the convergence speed of optimization, a novel hybrid algorithm consisting of Taguchi method and the steepest descent method is proposed in this work. Taguchi method using orthogonal arrays could quickly find the optimum combination of the levels of various factors, even when the number of level and/or factor is quite large. This algorithm is applied to the inverse determination of elastic constants of three composite plates by combining numerical method and vibration testing. For these problems, the proposed algorithm could find better elastic constants in less computation cost. Therefore, the proposed algorithm has nice robustness and fast convergence speed as compared to some hybrid genetic algorithms.
NASA Astrophysics Data System (ADS)
Vasu, M.; Shivananda, Nayaka H.
2018-04-01
EN47 steel samples are machined on a self-centered lathe using Chemical Vapor Deposition of coated TiCN/Al2O3/TiN and uncoated tungsten carbide tool inserts, with nose radius 0.8mm. Results are compared with each other and optimized using statistical tool. Input (cutting) parameters that are considered in this work are feed rate (f), cutting speed (Vc), and depth of cut (ap), the optimization criteria are based on the Taguchi (L9) orthogonal array. ANOVA method is adopted to evaluate the statistical significance and also percentage contribution for each model. Multiple response characteristics namely cutting force (Fz), tool tip temperature (T) and surface roughness (Ra) are evaluated. The results discovered that coated tool insert (TiCN/Al2O3/TiN) exhibits 1.27 and 1.29 times better than the uncoated tool insert for tool tip temperature and surface roughness respectively. A slight increase in cutting force was observed for coated tools.
Determining flexor-tendon repair techniques via soft computing
NASA Technical Reports Server (NTRS)
Johnson, M.; Firoozbakhsh, K.; Moniem, M.; Jamshidi, M.
2001-01-01
An SC-based multi-objective decision-making method for determining the optimal flexor-tendon repair technique from experimental and clinical survey data, and with variable circumstances, was presented. Results were compared with those from the Taguchi method. Using the Taguchi method results in the need to perform ad-hoc decisions when the outcomes for individual objectives are contradictory to a particular preference or circumstance, whereas the SC-based multi-objective technique provides a rigorous straightforward computational process in which changing preferences and importance of differing objectives are easily accommodated. Also, adding more objectives is straightforward and easily accomplished. The use of fuzzy-set representations of information categories provides insight into their performance throughout the range of their universe of discourse. The ability of the technique to provide a "best" medical decision given a particular physician, hospital, patient, situation, and other criteria was also demonstrated.
Determining flexor-tendon repair techniques via soft computing.
Johnson, M; Firoozbakhsh, K; Moniem, M; Jamshidi, M
2001-01-01
An SC-based multi-objective decision-making method for determining the optimal flexor-tendon repair technique from experimental and clinical survey data, and with variable circumstances, was presented. Results were compared with those from the Taguchi method. Using the Taguchi method results in the need to perform ad-hoc decisions when the outcomes for individual objectives are contradictory to a particular preference or circumstance, whereas the SC-based multi-objective technique provides a rigorous straightforward computational process in which changing preferences and importance of differing objectives are easily accommodated. Also, adding more objectives is straightforward and easily accomplished. The use of fuzzy-set representations of information categories provides insight into their performance throughout the range of their universe of discourse. The ability of the technique to provide a "best" medical decision given a particular physician, hospital, patient, situation, and other criteria was also demonstrated.
NASA Astrophysics Data System (ADS)
Novareza, O.; Sulistiyarini, D. H.; Wiradmoko, R.
2018-02-01
This paper presents the result of using Taguchi method in turning process of medium carbon steel of AISI 4140. The primary concern is to find the optimal surface roughness after turning process. The taguchi method is used to get a combination of factors and factor levels in order to get the optimum surface roughness level. Four important factors with three levels were used in experiment based on Taguchi method. A number of 27 experiments were carried out during the research and analysed using analysis of variance (ANOVA) method. The result of surface finish was determined in Ra type surface roughness. The depth of cut was found to be the most important factors for reducing the surface roughness of AISI 4140 steel. On the contrary, the other important factors i.e. spindle speed and rake side angle of the tool were proven to be less factors that affecting the surface finish. It is interesting to see the effect of coolant composition that gained the second important factors to reduce the roughness. It may need further research to explain this result.
NASA Astrophysics Data System (ADS)
Nirmala, D. B.; Raviraj, S.
2016-06-01
This paper presents the application of Taguchi approach to obtain optimal mix proportion for Self Compacting Concrete (SCC) containing spent foundry sand and M-sand. Spent foundry sand is used as a partial replacement for M-sand. The SCC mix has seven control factors namely, Coarse aggregate, M-sand with Spent Foundry sand, Cement, Fly ash, Water, Super plasticizer and Viscosity modifying agent. Modified Nan Su method is used to proportion the initial SCC mix. L18 (21×37) Orthogonal Arrays (OA) with the seven control factors having 3 levels is used in Taguchi approach which resulted in 18 SCC mix proportions. All mixtures are extensively tested both in fresh and hardened states to verify whether they meet the practical and technical requirements of SCC. The quality characteristics considering "Nominal the better" situation is applied to the test results to arrive at the optimal SCC mix proportion. Test results indicate that the optimal mix satisfies the requirements of fresh and hardened properties of SCC. The study reveals the feasibility of using spent foundry sand as a partial replacement of M-sand in SCC and also that Taguchi method is a reliable tool to arrive at optimal mix proportion of SCC.
Ebrahimi Zarandi, Mohammad Javad; Sohrabi, Mahmoud Reza; Khosravi, Morteza; Mansouriieh, Nafiseh; Davallo, Mehran; Khosravan, Azita
2016-01-01
This study synthesized magnetic nanoparticles (Fe(3)O(4)) immobilized on activated carbon (AC) and used them as an effective adsorbent for Cu(II) removal from aqueous solution. The effect of three parameters, including the concentration of Cu(II), dosage of Fe(3)O(4)/AC magnetic nanocomposite and pH on the removal of Cu(II) using Fe(3)O(4)/AC nanocomposite were studied. In order to examine and describe the optimum condition for each of the mentioned parameters, Taguchi's optimization method was used in a batch system and L9 orthogonal array was used for the experimental design. The removal percentage (R%) of Cu(II) and uptake capacity (q) were transformed into an accurate signal-to-noise ratio (S/N) for a 'larger-the-better' response. Taguchi results, which were analyzed based on choosing the best run by examining the S/N, were statistically tested using analysis of variance; the tests showed that all the parameters' main effects were significant within a 95% confidence level. The best conditions for removal of Cu(II) were determined at pH of 7, nanocomposite dosage of 0.1 gL(-1) and initial Cu(II) concentration of 20 mg L(-1) at constant temperature of 25 °C. Generally, the results showed that the simple Taguchi's method is suitable to optimize the Cu(II) removal experiments.
NASA Technical Reports Server (NTRS)
Stanley, Douglas O.; Unal, Resit; Joyner, C. R.
1992-01-01
The application of advanced technologies to future launch vehicle designs would allow the introduction of a rocket-powered, single-stage-to-orbit (SSTO) launch system early in the next century. For a selected SSTO concept, a dual mixture ratio, staged combustion cycle engine that employs a number of innovative technologies was selected as the baseline propulsion system. A series of parametric trade studies are presented to optimize both a dual mixture ratio engine and a single mixture ratio engine of similar design and technology level. The effect of varying lift-off thrust-to-weight ratio, engine mode transition Mach number, mixture ratios, area ratios, and chamber pressure values on overall vehicle weight is examined. The sensitivity of the advanced SSTO vehicle to variations in each of these parameters is presented, taking into account the interaction of each of the parameters with each other. This parametric optimization and sensitivity study employs a Taguchi design method. The Taguchi method is an efficient approach for determining near-optimum design parameters using orthogonal matrices from design of experiments (DOE) theory. Using orthogonal matrices significantly reduces the number of experimental configurations to be studied. The effectiveness and limitations of the Taguchi method for propulsion/vehicle optimization studies as compared to traditional single-variable parametric trade studies is also discussed.
NASA Astrophysics Data System (ADS)
Wisessirikul, W.; Loykulnant, S.; Montha, S.; Fhulua, T.; Prapainainar, P.
2016-06-01
This research studied the quantity of free formaldehyde released from 2- [(hydroxymethyl)amino]ethanol (HAE) in DI water and natural rubber latex mixture using high-performance liquid chromatography (HPLC) technique. The quantity of formaldehyde retained in the solution was cross-checked by using titration technique. The investigated factors were the concentration of preservative (HAE), pH, and temperature. Taguchi's method was used to design the experiments. The number of experiments was reduced to 16 experiments from all possible experiments by orthogonal arrays (3 factors and 4 levels in each factor). Minitab program was used as a tool for statistical calculation and for finding the suitable condition for the preservative system. HPLC studies showed that higher temperature and higher concentration of the preservative influence the amount of formaldehyde released. It was found that conditions at which formaldehyde was released in the lowest amount were 1.6%w/v HAE, 4 to 40 °C, and the original pH. Nevertheless, the pH value of NR latex should be more than 10 (the suitable pH value was found to be 13). This preservative can be used to replace current preservative systems and can maintain the quality of latex for long-term storage. Use of the proposed preservative system was also shown to have reduced impact on the toxicity of the environment.
Rao, Ravella Sreenivas; Kumar, C Ganesh; Prakasham, R Shetty; Hobbs, Phil J
2008-04-01
Success in experiments and/or technology mainly depends on a properly designed process or product. The traditional method of process optimization involves the study of one variable at a time, which requires a number of combinations of experiments that are time, cost and labor intensive. The Taguchi method of design of experiments is a simple statistical tool involving a system of tabulated designs (arrays) that allows a maximum number of main effects to be estimated in an unbiased (orthogonal) fashion with a minimum number of experimental runs. It has been applied to predict the significant contribution of the design variable(s) and the optimum combination of each variable by conducting experiments on a real-time basis. The modeling that is performed essentially relates signal-to-noise ratio to the control variables in a 'main effect only' approach. This approach enables both multiple response and dynamic problems to be studied by handling noise factors. Taguchi principles and concepts have made extensive contributions to industry by bringing focused awareness to robustness, noise and quality. This methodology has been widely applied in many industrial sectors; however, its application in biological sciences has been limited. In the present review, the application and comparison of the Taguchi methodology has been emphasized with specific case studies in the field of biotechnology, particularly in diverse areas like fermentation, food processing, molecular biology, wastewater treatment and bioremediation.
NASA Astrophysics Data System (ADS)
Azmi, Nur Iffah Mohamed; Arifin Mat Piah, Kamal; Yusoff, Wan Azhar Wan; Romlay, Fadhlur Rahman Mohd
2018-03-01
Controller that uses PID parameters requires a good tuning method in order to improve the control system performance. Tuning PID control method is divided into two namely the classical methods and the methods of artificial intelligence. Particle swarm optimization algorithm (PSO) is one of the artificial intelligence methods. Previously, researchers had integrated PSO algorithms in the PID parameter tuning process. This research aims to improve the PSO-PID tuning algorithms by integrating the tuning process with the Variable Weight Grey- Taguchi Design of Experiment (DOE) method. This is done by conducting the DOE on the two PSO optimizing parameters: the particle velocity limit and the weight distribution factor. Computer simulations and physical experiments were conducted by using the proposed PSO- PID with the Variable Weight Grey-Taguchi DOE and the classical Ziegler-Nichols methods. They are implemented on the hydraulic positioning system. Simulation results show that the proposed PSO-PID with the Variable Weight Grey-Taguchi DOE has reduced the rise time by 48.13% and settling time by 48.57% compared to the Ziegler-Nichols method. Furthermore, the physical experiment results also show that the proposed PSO-PID with the Variable Weight Grey-Taguchi DOE tuning method responds better than Ziegler-Nichols tuning. In conclusion, this research has improved the PSO-PID parameter by applying the PSO-PID algorithm together with the Variable Weight Grey-Taguchi DOE method as a tuning method in the hydraulic positioning system.
NASA Astrophysics Data System (ADS)
Kavimani, V.; Prakash, K. Soorya
2017-11-01
This paper deals with the fabrication of reduced graphene oxide (r-GO) reinforced Magnesium Metal Matrix Composite (MMC) through a novel solvent based powder metallurgy route. Investigations over basic and functional properties of developed MMC reveals that addition of r-GO improvises the microhardness upto 64 HV but however decrement in specific wear rate is also notified. Visualization of worn out surfaces through SEM images clearly explains for the occurrence of plastic deformation and the presence of wear debris because of ploughing out action. Taguchi coupled Artificial Neural Network (ANN) technique is adopted to arrive at optimal values of the input parameters such as load, reinforcement weight percentage, sliding distance and sliding velocity and thereby achieve minimal target output value viz. specific wear rate. Influence of any of the input parameter over specific wear rate studied through ANOVA reveals that load acting on pin has a major influence with 38.85% followed by r-GO wt. % of 25.82%. ANN model developed to predict specific wear rate value based on the variation of input parameter facilitates better predictability with R-value of 98.4% when compared with the outcomes of regression model.
Optimisation Of Cutting Parameters Of Composite Material Laser Cutting Process By Taguchi Method
NASA Astrophysics Data System (ADS)
Lokesh, S.; Niresh, J.; Neelakrishnan, S.; Rahul, S. P. Deepak
2018-03-01
The aim of this work is to develop a laser cutting process model that can predict the relationship between the process input parameters and resultant surface roughness, kerf width characteristics. The research conduct is based on the Design of Experiment (DOE) analysis. Response Surface Methodology (RSM) is used in this work. It is one of the most practical and most effective techniques to develop a process model. Even though RSM has been used for the optimization of the laser process, this research investigates laser cutting of materials like Composite wood (veneer)to be best circumstances of laser cutting using RSM process. The input parameters evaluated are focal length, power supply and cutting speed, the output responses being kerf width, surface roughness, temperature. To efficiently optimize and customize the kerf width and surface roughness characteristics, a machine laser cutting process model using Taguchi L9 orthogonal methodology was proposed.
Yadav, Kartikey K; Dasgupta, Kinshuk; Singh, Dhruva K; Varshney, Lalit; Singh, Harvinderpal
2015-03-06
Polyethersulfone-based beads encapsulating di-2-ethylhexyl phosphoric acid have been synthesized and evaluated for the recovery of rare earth values from the aqueous media. Percentage recovery and the sorption behavior of Dy(III) have been investigated under wide range of experimental parameters using these beads. Taguchi method utilizing L-18 orthogonal array has been adopted to identify the most influential process parameters responsible for higher degree of recovery with enhanced sorption of Dy(III) from chloride medium. Analysis of variance indicated that the feed concentration of Dy(III) is the most influential factor for equilibrium sorption capacity, whereas aqueous phase acidity influences the percentage recovery most. The presence of polyvinyl alcohol and multiwalled carbon nanotube modified the internal structure of the composite beads and resulted in uniform distribution of organic extractant inside polymeric matrix. The experiment performed under optimum process conditions as predicted by Taguchi method resulted in enhanced Dy(III) recovery and sorption capacity by polymeric beads with minimum standard deviation. Copyright © 2015 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Bharti, P. K.; Khan, M. I.; Singh, Harbinder
2010-10-01
Off-line quality control is considered to be an effective approach to improve product quality at a relatively low cost. The Taguchi method is one of the conventional approaches for this purpose. Through this approach, engineers can determine a feasible combination of design parameters such that the variability of a product's response can be reduced and the mean is close to the desired target. The traditional Taguchi method was focused on ensuring good performance at the parameter design stage with one quality characteristic, but most products and processes have multiple quality characteristics. The optimal parameter design minimizes the total quality loss for multiple quality characteristics. Several studies have presented approaches addressing multiple quality characteristics. Most of these papers were concerned with maximizing the parameter combination of signal to noise (SN) ratios. The results reveal the advantages of this approach are that the optimal parameter design is the same as the traditional Taguchi method for the single quality characteristic; the optimal design maximizes the amount of reduction of total quality loss for multiple quality characteristics. This paper presents a literature review on solving multi-response problems in the Taguchi method and its successful implementation in various industries.
Taguchi experimental design to determine the taste quality characteristic of candied carrot
NASA Astrophysics Data System (ADS)
Ekawati, Y.; Hapsari, A. A.
2018-03-01
Robust parameter design is used to design product that is robust to noise factors so the product’s performance fits the target and delivers a better quality. In the process of designing and developing the innovative product of candied carrot, robust parameter design is carried out using Taguchi Method. The method is used to determine an optimal quality design. The optimal quality design is based on the process and the composition of product ingredients that are in accordance with consumer needs and requirements. According to the identification of consumer needs from the previous research, quality dimensions that need to be assessed are the taste and texture of the product. The quality dimension assessed in this research is limited to the taste dimension. Organoleptic testing is used for this assessment, specifically hedonic testing that makes assessment based on consumer preferences. The data processing uses mean and signal to noise ratio calculation and optimal level setting to determine the optimal process/composition of product ingredients. The optimal value is analyzed using confirmation experiments to prove that proposed product match consumer needs and requirements. The result of this research is identification of factors that affect the product taste and the optimal quality of product according to Taguchi Method.
Taguchi Approach to Design Optimization for Quality and Cost: An Overview
NASA Technical Reports Server (NTRS)
Unal, Resit; Dean, Edwin B.
1990-01-01
Calibrations to existing cost of doing business in space indicate that to establish human presence on the Moon and Mars with the Space Exploration Initiative (SEI) will require resources, felt by many, to be more than the national budget can afford. In order for SEI to succeed, we must actually design and build space systems at lower cost this time, even with tremendous increases in quality and performance requirements, such as extremely high reliability. This implies that both government and industry must change the way they do business. Therefore, new philosophy and technology must be employed to design and produce reliable, high quality space systems at low cost. In recognizing the need to reduce cost and improve quality and productivity, Department of Defense (DoD) and National Aeronautics and Space Administration (NASA) have initiated Total Quality Management (TQM). TQM is a revolutionary management strategy in quality assurance and cost reduction. TQM requires complete management commitment, employee involvement, and use of statistical tools. The quality engineering methods of Dr. Taguchi, employing design of experiments (DOE), is one of the most important statistical tools of TQM for designing high quality systems at reduced cost. Taguchi methods provide an efficient and systematic way to optimize designs for performance, quality, and cost. Taguchi methods have been used successfully in Japan and the United States in designing reliable, high quality products at low cost in such areas as automobiles and consumer electronics. However, these methods are just beginning to see application in the aerospace industry. The purpose of this paper is to present an overview of the Taguchi methods for improving quality and reducing cost, describe the current state of applications and its role in identifying cost sensitive design parameters.
NASA Astrophysics Data System (ADS)
Khanna, Rajesh; Kumar, Anish; Garg, Mohinder Pal; Singh, Ajit; Sharma, Neeraj
2015-12-01
Electric discharge drill machine (EDDM) is a spark erosion process to produce micro-holes in conductive materials. This process is widely used in aerospace, medical, dental and automobile industries. As for the performance evaluation of the electric discharge drilling machine, it is very necessary to study the process parameters of machine tool. In this research paper, a brass rod 2 mm diameter was selected as a tool electrode. The experiments generate output responses such as tool wear rate (TWR). The best parameters such as pulse on-time, pulse off-time and water pressure were studied for best machining characteristics. This investigation presents the use of Taguchi approach for better TWR in drilling of Al-7075. A plan of experiments, based on L27 Taguchi design method, was selected for drilling of material. Analysis of variance (ANOVA) shows the percentage contribution of the control factor in the machining of Al-7075 in EDDM. The optimal combination levels and the significant drilling parameters on TWR were obtained. The optimization results showed that the combination of maximum pulse on-time and minimum pulse off-time gives maximum MRR.
Mohaptra, Sonali; Dash, Preeti Krishna; Behera, Sudhanshu Shekar; Thatoi, Hrudayanath
2016-01-01
In the bioconversion of lignocelluloses for bioethanol, pretreatment seems to be the most important step which improves the elimination of the lignin and hemicelluloses content, exposing cellulose to further hydrolysis. The present study discusses the application of dynamic statistical techniques like the Taguchi method and artificial neural network (ANN) in the optimization of pretreatment of lignocellulosic biomasses such as Hybrid Napier grass (HNG) (Pennisetum purpureum) and Denanath grass (DG) (Pennisetum pedicellatum), using alkali sodium hydroxide. This study analysed and determined a parameter combination with a low number of experiments by using the Taguchi method in which both the substrates can be efficiently pretreated. The optimized parameters obtained from the L16 orthogonal array are soaking time (18 and 26 h), temperature (60°C and 55°C), and alkali concentration (1%) for HNG and DG, respectively. High performance liquid chromatography analysis of the optimized pretreated grass varieties confirmed the presence of glucan (47.94% and 46.50%), xylan (9.35% and 7.95%), arabinan (2.15% and 2.2%), and galactan/mannan (1.44% and 1.52%) for HNG and DG, respectively. Physicochemical characterization studies of native and alkali-pretreated grasses were carried out by scanning electron microscopy and Fourier transformation Infrared spectroscopy which revealed some morphological differences between the native and optimized pretreated samples. Model validation by ANN showed a good agreement between experimental results and the predicted responses.
NASA Astrophysics Data System (ADS)
Sudibyo, Hermida, L.; Junaedi, A.; Putra, F. A.
2017-11-01
Nickel and cobalt metal able to process from low grade laterite using solvent extraction and electrowinning. One of electrowinning methods which has good performance to produce pure metal is electrometal-electrowinninge(EMEW). In this work, solventextraction was used to separate nickel and cobalt which useCyanex-Versatic Acid in toluene as an organic phase. An aqueous phase of extraction was processed using EMEW in order to deposit the nickel metal in Cathode electrode. The parameters which used in this work were batch temperature, operation time, voltage, and boric acid concentration. Those parameters were studied and optimized using the design of experiment of Taguchi. The Taguchi analysis result shows that the optimum result of EMEW was at 60°C of batch temperature, 2 Voltage, 6 hours operation and 0.5 M of boric acid.
The Taguchi Method Application to Improve the Quality of a Sustainable Process
NASA Astrophysics Data System (ADS)
Titu, A. M.; Sandu, A. V.; Pop, A. B.; Titu, S.; Ciungu, T. C.
2018-06-01
Taguchi’s method has always been a method used to improve the quality of the analyzed processes and products. This research shows an unusual situation, namely the modeling of some parameters, considered technical parameters, in a process that is wanted to be durable by improving the quality process and by ensuring quality using an experimental research method. Modern experimental techniques can be applied in any field and this study reflects the benefits of interacting between the agriculture sustainability principles and the Taguchi’s Method application. The experimental method used in this practical study consists of combining engineering techniques with experimental statistical modeling to achieve rapid improvement of quality costs, in fact seeking optimization at the level of existing processes and the main technical parameters. The paper is actually a purely technical research that promotes a technical experiment using the Taguchi method, considered to be an effective method since it allows for rapid achievement of 70 to 90% of the desired optimization of the technical parameters. The missing 10 to 30 percent can be obtained with one or two complementary experiments, limited to 2 to 4 technical parameters that are considered to be the most influential. Applying the Taguchi’s Method in the technique and not only, allowed the simultaneous study in the same experiment of the influence factors considered to be the most important in different combinations and, at the same time, determining each factor contribution.
NASA Astrophysics Data System (ADS)
Vo, Thanh Tu; Chen, Xiaopeng; Shen, Weixiang; Kapoor, Ajay
2015-01-01
In this paper, a new charging strategy of lithium-polymer batteries (LiPBs) has been proposed based on the integration of Taguchi method (TM) and state of charge estimation. The TM is applied to search an optimal charging current pattern. An adaptive switching gain sliding mode observer (ASGSMO) is adopted to estimate the SOC which controls and terminates the charging process. The experimental results demonstrate that the proposed charging strategy can successfully charge the same types of LiPBs with different capacities and cycle life. The proposed charging strategy also provides much shorter charging time, narrower temperature variation and slightly higher energy efficiency than the equivalent constant current constant voltage charging method.
Investigation of Structures of Microwave Microelectromechanical-System Switches by Taguchi Method
NASA Astrophysics Data System (ADS)
Lai, Yeong-Lin; Lin, Chien-Hung
2007-10-01
The optimal design of microwave microelectromechanical-system (MEMS) switches by the Taguchi method is presented. The structures of the switches are analyzed and optimized in terms of the effective stiffness constant, the maximum von Mises stress, and the natural frequency in order to improve the reliability and the performance of the MEMS switches. There are four factors, each of which has three levels in the Taguchi method for the MEMS switches. An L9(34) orthogonal array is used for the matrix experiments. The characteristics of the experiments are studied by the finite-element method and the analytical method. The responses of the signal-to-noise (S/N) ratios of the characteristics of the switches are investigated. The statistical analysis of variance (ANOVA) is used to interpret the experimental results and decide the significant factors. The final optimum setting, A1B3C1D2, predicts that the effective stiffness constant is 1.06 N/m, the maximum von Mises stress is 76.9 MPa, and the natural frequency is 29.331 kHz. The corresponding switching time is 34 μs, and the pull-down voltage is 9.8 V.
Ozturk, Ismet; Golec, Adem; Karaman, Safa; Sagdic, Osman; Kayacier, Ahmed
2010-10-01
In this study, the effects of the incorporation of some herbal teas at different concentrations into the ice cream mix on the population of Listeria monocytogenes were studied using Taguchi method. The ice cream mix samples flavored with herbal teas were prepared using green tea and sage at different concentrations. Afterward, fresh culture of L. monocytogenes was inoculated into the samples and the L. monocytogenes was counted at different storage periods. Taguchi method was used for experimental design and analysis. In addition, some physicochemical properties of samples were examined. Results suggested that there was some effect, although little, on the population of L. monocytogenes when herbal tea was incorporated into the ice cream mix. Additionally, the use of herbal tea caused a decrease in the pH values of the samples and significant changes in the color values.
NASA Astrophysics Data System (ADS)
Khavekar, Rajendra; Vasudevan, Hari, Dr.; Modi, Bhavik
2017-08-01
Two well-known Design of Experiments (DoE) methodologies, such as Taguchi Methods (TM) and Shainin Systems (SS) are compared and analyzed in this study through their implementation in a plastic injection molding unit. Experiments were performed at a perfume bottle cap manufacturing company (made by acrylic material) using TM and SS to find out the root cause of defects and to optimize the process parameters for minimum rejection. Experiments obtained the rejection rate to be 8.57% from 40% (appx.) during trial runs, which is quiet low, representing successful implementation of these DoE methods. The comparison showed that both methodologies gave same set of variables as critical for defect reduction, but with change in their significance order. Also, Taguchi methods require more number of experiments and consume more time compared to the Shainin System. Shainin system is less complicated and is easy to implement, whereas Taguchi methods is statistically more reliable for optimization of process parameters. Finally, experimentations implied that DoE methods are strong and reliable in implementation, as organizations attempt to improve the quality through optimization.
Improving the Glucose Meter Error Grid With the Taguchi Loss Function.
Krouwer, Jan S
2016-07-01
Glucose meters often have similar performance when compared by error grid analysis. This is one reason that other statistics such as mean absolute relative deviation (MARD) are used to further differentiate performance. The problem with MARD is that too much information is lost. But additional information is available within the A zone of an error grid by using the Taguchi loss function. Applying the Taguchi loss function gives each glucose meter difference from reference a value ranging from 0 (no error) to 1 (error reaches the A zone limit). Values are averaged over all data which provides an indication of risk of an incorrect medical decision. This allows one to differentiate glucose meter performance for the common case where meters have a high percentage of values in the A zone and no values beyond the B zone. Examples are provided using simulated data. © 2015 Diabetes Technology Society.
NASA Astrophysics Data System (ADS)
Gaaz, Tayser Sumer; Sulong, Abu Bakar; Kadhum, Abdul Amir H.; Nassir, Mohamed H.; Al-Amiery, Ahmed A.
The variation of the results of the mechanical properties of halloysite nanotubes (HNTs) reinforced thermoplastic polyurethane (TPU) at different HNTs loadings was implemented as a tool for analysis. The preparation of HNTs-TPU nanocomposites was performed under four controlled parameters of mixing temperature, mixing speed, mixing time, and HNTs loading at three levels each to satisfy Taguchi method orthogonal array L9 aiming to optimize these parameters for the best measurements of tensile strength, Young's modulus, and tensile strain (known as responses). The maximum variation of the experimental results for each response was determined and analysed based on the optimized results predicted by Taguchi method and ANOVA. It was found that the maximum absolute variations of the three mentioned responses are 69%, 352%, and 126%, respectively. The analysis has shown that the preparation of the optimized tensile strength requires 1 wt.% HNTs loading (excluding 2 wt.% and 3 wt.%), mixing temperature of 190 °C (excluding 200 °C and 210 °C), and mixing speed of 30 rpm (excluding 40 rpm and 50 rpm). In addition, the analysis has determined that the mixing time at 20 min has no effect on the preparation. The mentioned analysis was fortified by ANOVA, images of FESEM, and DSC results. Seemingly, the agglomeration and distribution of HNTs in the nanocomposite play an important role in the process. The outcome of the analysis could be considered as a very important step towards the reliability of Taguchi method.
NASA Astrophysics Data System (ADS)
Harlapur, M. D.; Mallapur, D. G.; Udupa, K. Rajendra
2018-04-01
In the present study, an experimental study of the volumetric wear behaviour of Aluminium (Al-25Mg2Si2Cu4Ni) alloy in as cast and 5Hr homogenized with T6 heat treatment is carried out at constant load. The Pin on disc apparatus was used to carry out the sliding wear test. Taguchi method based on L-16 orthogonal array was employed to evaluate the data on the wear behavior. Signal-to-noise ratio among the objective of smaller the better and mean of means results were used. General regression model is obtained by correlation. Lastly confirmation test was completed to compose a comparison between the experimental results foreseen from the mention correlation. The mathematical model reveals the load has maximum contribution on the wear rate compared to speed. Scanning Electron Microscope was used to analyze the worn-out wear surfaces. Wear results show that 5Hr homogenized Al-25Mg2Si2Cu4Ni alloy samples with T6 treated had better volumetric wear resistance as compared to as cast samples.
NASA Astrophysics Data System (ADS)
Abu, M. Y.; Nor, E. E. Mohd; Rahman, M. S. Abd
2018-04-01
Integration between quality and costing system is very crucial in order to achieve an accurate product cost and profit. Current practice by most of remanufacturers, there are still lacking on optimization during the remanufacturing process which contributed to incorrect variables consideration to the costing system. Meanwhile, traditional costing accounting being practice has distortion in the cost unit which lead to inaccurate cost of product. The aim of this work is to identify the critical and non-critical variables during remanufacturing process using Mahalanobis-Taguchi System and simultaneously estimate the cost using Activity Based Costing method. The orthogonal array was applied to indicate the contribution of variables in the factorial effect graph and the critical variables were considered with overhead costs that are actually demanding the activities. This work improved the quality inspection together with costing system to produce an accurate profitability information. As a result, the cost per unit of remanufactured crankshaft of MAN engine model with 5 critical crankpins is MYR609.50 while Detroit engine model with 4 critical crankpins is MYR1254.80. The significant of output demonstrated through promoting green by reducing re-melting process of damaged parts to ensure consistent benefit of return cores.
Yeh, Chia-Hsien; Zhao, Zi-Qi; Shen, Pi-Lan; Lin, Yu-Cheng
2014-01-01
This study presents an optical inspection system for detecting a commercial point-of-care testing product and a new detection model covering from qualitative to quantitative analysis. Human chorionic gonadotropin (hCG) strips (cut-off value of the hCG commercial product is 25 mIU/mL) were the detection target in our study. We used a complementary metal-oxide semiconductor (CMOS) sensor to detect the colors of the test line and control line in the specific strips and to reduce the observation errors by the naked eye. To achieve better linearity between the grayscale and the concentration, and to decrease the standard deviation (increase the signal to noise ratio, S/N), the Taguchi method was used to find the optimal parameters for the optical inspection system. The pregnancy test used the principles of the lateral flow immunoassay, and the colors of the test and control line were caused by the gold nanoparticles. Because of the sandwich immunoassay model, the color of the gold nanoparticles in the test line was darkened by increasing the hCG concentration. As the results reveal, the S/N increased from 43.48 dB to 53.38 dB, and the hCG concentration detection increased from 6.25 to 50 mIU/mL with a standard deviation of less than 10%. With the optimal parameters to decrease the detection limit and to increase the linearity determined by the Taguchi method, the optical inspection system can be applied to various commercial rapid tests for the detection of ketamine, troponin I, and fatty acid binding protein (FABP). PMID:25256108
Abrasive wear response of TIG-melted TiC composite coating: Taguchi approach
NASA Astrophysics Data System (ADS)
Maleque, M. A.; Bello, K. A.; Adebisi, A. A.; Dube, A.
2017-03-01
In this study, Taguchi design of experiment approach has been applied to assess wear behaviour of TiC composite coatings deposited on AISI 4340 steel substrates by novel powder preplacement and TIG torch melting processes. To study the abrasive wear behaviour of these coatings against alumina ball at 600° C, a Taguchi’s orthogonal array is used to acquire the wear test data for determining optimal parameters that lead to the minimization of wear rate. Composite coatings are developed based on Taguchi’s L-16 orthogonal array experiment with three process parameters (welding current, welding speed, welding voltage and shielding gas flow rate) at four levels. In this technique, mean response and signal-to-noise ratio are used to evaluate the influence of the TIG process parameters on the wear rate performance of the composite coated surfaces. The results reveal that welding voltage is the most significant control parameter for minimizing wear rate while the current presents the least contribution to the wear rate reduction. The study also shows the best optimal condition has been arrived at A3 (90 A), B4 (2.5 mm/s), C3 (30 V) and D3 (20 L/min), which gives minimum wear rate in TiC embedded coatings. Finally, a confirmatory experiment has been conducted to verify the optimized result and shows that the error between the predicted values and the experimental observation at the optimal condition lies within the limit of 4.7 %. Thus, the validity of the optimum condition for the coatings is established.
NASA Astrophysics Data System (ADS)
Chen, Junxun; Cheng, Longsheng; Yu, Hui; Hu, Shaolin
2018-01-01
NASA Astrophysics Data System (ADS)
Nkuissi Tchognia, Joël Hervé; Hartiti, Bouchaib; Ridah, Abderraouf; Ndjaka, Jean-Marie; Thevenin, Philippe
2016-07-01
Present research deals with the optimal deposition parameters configuration for the synthesis of Cu2ZnSnS4 (CZTS) thin films using the sol-gel method associated to spin coating on ordinary glass substrates without sulfurization. The Taguchi design with a L9 (34) orthogonal array, a signal-to-noise (S/N) ratio and an analysis of variance (ANOVA) are used to optimize the performance characteristic (optical band gap) of CZTS thin films. Four deposition parameters called factors namely the annealing temperature, the annealing time, the ratios Cu/(Zn + Sn) and Zn/Sn were chosen. To conduct the tests using the Taguchi method, three levels were chosen for each factor. The effects of the deposition parameters on structural and optical properties are studied. The determination of the most significant factors of the deposition process on optical properties of as-prepared films is also done. The results showed that the significant parameters are Zn/Sn ratio and the annealing temperature by applying the Taguchi method.
Besseris, George J
2018-03-01
Generalized regression neural networks (GRNN) may act as crowdsourcing cognitive agents to screen small, dense and complex datasets. The concurrent screening and optimization of several complex physical and sensory traits of bread is developed using a structured Taguchi-type micro-mining technique. A novel product outlook is offered to industrial operations to cover separate aspects of smart product design, engineering and marketing. Four controlling factors were selected to be modulated directly on a modern production line: 1) the dough weight, 2) the proofing time, 3) the baking time, and 4) the oven zone temperatures. Concentrated experimental recipes were programmed using the Taguchi-type L 9 (3 4 ) OA-sampler to detect potentially non-linear multi-response tendencies. The fused behavior of the master-ranked bread characteristics behavior was smart sampled with GRNN-crowdsourcing and robust analysis. It was found that the combination of the oven zone temperatures to play a highly influential role in all investigated scenarios. Moreover, the oven zone temperatures and the dough weight appeared to be instrumental when attempting to synchronously adjusting all four physical characteristics. The optimal oven-zone temperature setting for concurrent screening-and-optimization was found to be 270-240 °C. The optimized (median) responses for loaf weight, moisture, height, width, color, flavor, crumb structure, softness, and elasticity are: 782 g, 34.8 %, 9.36 cm, 10.41 cm, 6.6, 7.2, 7.6, 7.3, and 7.0, respectively.
Pervez, Hifsa; Mozumder, Mohammad S; Mourad, Abdel-Hamid I
2016-08-22
The current study presents an investigation on the optimization of injection molding parameters of HDPE/TiO₂ nanocomposites using grey relational analysis with the Taguchi method. Four control factors, including filler concentration (i.e., TiO₂), barrel temperature, residence time and holding time, were chosen at three different levels of each. Mechanical properties, such as yield strength, Young's modulus and elongation, were selected as the performance targets. Nine experimental runs were carried out based on the Taguchi L₉ orthogonal array, and the data were processed according to the grey relational steps. The optimal process parameters were found based on the average responses of the grey relational grades, and the ideal operating conditions were found to be a filler concentration of 5 wt % TiO₂, a barrel temperature of 225 °C, a residence time of 30 min and a holding time of 20 s. Moreover, analysis of variance (ANOVA) has also been applied to identify the most significant factor, and the percentage of TiO₂ nanoparticles was found to have the most significant effect on the properties of the HDPE/TiO₂ nanocomposites fabricated through the injection molding process.
Ashengroph, Morahem; Nahvi, Iraj; Amini, Jahanshir
2013-01-01
For all industrial processes, modelling, optimisation and control are the keys to enhance productivity and ensure product quality. In the current study, the optimization of process parameters for improving the conversion of isoeugenol to vanillin by Psychrobacter sp. CSW4 was investigated by means of Taguchi approach and Box-Behnken statistical design under resting cell conditions. Taguchi design was employed for screening the significant variables in the bioconversion medium. Sequentially, Box-Behnken design experiments under Response Surface Methodology (RSM) was used for further optimization. Four factors (isoeugenol, NaCl, biomass and tween 80 initial concentrations), which have significant effects on vanillin yield, were selected from ten variables by Taguchi experimental design. With the regression coefficient analysis in the Box-Behnken design, a relationship between vanillin production and four significant variables was obtained, and the optimum levels of the four variables were as follows: initial isoeugenol concentration 6.5 g/L, initial tween 80 concentration 0.89 g/L, initial NaCl concentration 113.2 g/L and initial biomass concentration 6.27 g/L. Under these optimized conditions, the maximum predicted concentration of vanillin was 2.25 g/L. These optimized values of the factors were validated in a triplicate shaking flask study and an average of 2.19 g/L for vanillin, which corresponded to a molar yield 36.3%, after a 24 h bioconversion was obtained. The present work is the first one reporting the application of Taguchi design and Response surface methodology for optimizing bioconversion of isoeugenol into vanillin under resting cell conditions.
NASA Astrophysics Data System (ADS)
Rahmani, Mashaallah; Kaykhaii, Massoud; Sasani, Mojtaba
2018-01-01
This study aimed to investigate the efficiency of 3A zeolite as a novel adsorbent for removal of Rhodamine B and Malachite green dyes from water samples. To increase the removal efficiency, effecting parameters on adsorption process were investigated and optimized by adopting Taguchi design of experiments approach. The percentage contribution of each parameter on the removal of Rhodamine B and Malachite green dyes determined using ANOVA and showed that the most effective parameters in removal of RhB and MG by 3A zeolite are initial concentration of dye and pH, respectively. Under optimized condition, the amount predicted by Taguchi design method and the value obtained experimentally, showed good closeness (more than 94.86%). Good adsorption efficiency obtained for proposed methods indicates that, the 3A zeolite is capable to remove the significant amounts of Rhodamine B and Malachite green from environmental water samples.
NASA Astrophysics Data System (ADS)
Krishnudu, D. Mohana; Sreeramulu, D.; Reddy, P. Venkateshwar
2018-04-01
In the current study mechanical properties of particles filled hybrid composites have been studied. The mechanical properties of the hybrid composite mainly depend on the proportions of the coir weight, Luffa weight and filler weight. RSM along with Taguchi method have been applied to find the optimized parameters of the hybrid composites. From the current study it was observed that the tensile strength of the composite mainly depends on the coir percent than the other two particles.
Pervez, Hifsa; Mozumder, Mohammad S.; Mourad, Abdel-Hamid I.
2016-01-01
The current study presents an investigation on the optimization of injection molding parameters of HDPE/TiO2 nanocomposites using grey relational analysis with the Taguchi method. Four control factors, including filler concentration (i.e., TiO2), barrel temperature, residence time and holding time, were chosen at three different levels of each. Mechanical properties, such as yield strength, Young’s modulus and elongation, were selected as the performance targets. Nine experimental runs were carried out based on the Taguchi L9 orthogonal array, and the data were processed according to the grey relational steps. The optimal process parameters were found based on the average responses of the grey relational grades, and the ideal operating conditions were found to be a filler concentration of 5 wt % TiO2, a barrel temperature of 225 °C, a residence time of 30 min and a holding time of 20 s. Moreover, analysis of variance (ANOVA) has also been applied to identify the most significant factor, and the percentage of TiO2 nanoparticles was found to have the most significant effect on the properties of the HDPE/TiO2 nanocomposites fabricated through the injection molding process. PMID:28773830
Chung, C K; Shih, T R; Chen, T C; Wu, B H
2008-10-01
A planar micromixer with rhombic microchannels and a converging-diverging element has been systematically investigated by the Taguchi method, CFD-ACE simulations and experiments. To reduce the footprint and extend the operation range of Reynolds number, Taguchi method was used to numerically study the performance of the micromixer in a L(9) orthogonal array. Mixing efficiency is prominently influenced by geometrical parameters and Reynolds number (Re). The four factors in a L(9) orthogonal array are number of rhombi, turning angle, width of the rhombic channel and width of the throat. The degree of sensitivity by Taguchi method can be ranked as: Number of rhombi > Width of the rhombic channel > Width of the throat > Turning angle of the rhombic channel. Increasing the number of rhombi, reducing the width of the rhombic channel and throat and lowering the turning angle resulted in better fluid mixing efficiency. The optimal design of the micromixer in simulations indicates over 90% mixing efficiency at both Re > or = 80 and Re < or = 0.1. Experimental results in the optimal simulations are consistent with the simulated one. This planar rhombic micromixer has simplified the complex fabrication process of the multi-layer or three-dimensional micromixers and improved the performance of a previous rhombic micromixer at a reduced footprint and lower Re.
NASA Astrophysics Data System (ADS)
Singh, Akhilesh Kumar; Debnath, Tapas; Dey, Vidyut; Rai, Ram Naresh
2017-10-01
P-91 is modified 9Cr-1Mo steel. Fabricated structures and components of P-91 has a lot of application in power and chemical industry owing to its excellent properties like high temperature stress corrosion resistance, less susceptibility to thermal fatigue at high operating temperatures. The weld quality and surface finish of fabricated structure of P91 is very good when welded by Tungsten Inert Gas welding (TIG). However, the process has its limitation regarding weld penetration. The success of a welding process lies in fabricating with such a combination of parameters that gives maximum weld penetration and minimum weld width. To carry out an investigation on the effect of the autogenous TIG welding parameters on weld penetration and weld width, bead-on-plate welds were carried on P91 plates of thickness 6 mm in accordance to a Taguchi L9 design. Welding current, welding speed and gas flow rate were the three control variables in the investigation. After autogenous (TIG) welding, the dimension of the weld width, weld penetration and weld area were successfully measured by an image analysis technique developed for the study. The maximum error for the measured dimensions of the weld width, penetration and area with the developed image analysis technique was only 2 % compared to the measurements of Leica-Q-Win-V3 software installed in optical microscope. The measurements with the developed software, unlike the measurements under a microscope, required least human intervention. An Analysis of Variance (ANOVA) confirms the significance of the selected parameters. Thereafter, Taguchi's method was successfully used to trade-off between maximum penetration and minimum weld width while keeping the weld area at a minimum.
NASA Astrophysics Data System (ADS)
Luangpaiboon, P.; Suwankham, Y.; Homrossukon, S.
2010-10-01
This research presents a development of a design of experiment technique for quality improvement in automotive manufacturing industrial. The quality of interest is the colour shade, one of the key feature and exterior appearance for the vehicles. With low percentage of first time quality, the manufacturer has spent a lot of cost for repaired works as well as the longer production time. To permanently dissolve such problem, the precisely spraying condition should be optimized. Therefore, this work will apply the full factorial design, the multiple regression, the constrained response surface optimization methods or CRSOM, and Taguchi's method to investigate the significant factors and to determine the optimum factor level in order to improve the quality of paint shop. Firstly, 2κ full factorial was employed to study the effect of five factors including the paint flow rate at robot setting, the paint levelling agent, the paint pigment, the additive slow solvent, and non volatile solid at spraying of atomizing spraying machine. The response values of colour shade at 15 and 45 degrees were measured using spectrophotometer. Then the regression models of colour shade at both degrees were developed from the significant factors affecting each response. Consequently, both regression models were placed into the form of linear programming to maximize the colour shade subjected to 3 main factors including the pigment, the additive solvent and the flow rate. Finally, Taguchi's method was applied to determine the proper level of key variable factors to achieve the mean value target of colour shade. The factor of non volatile solid was found to be one more additional factor at this stage. Consequently, the proper level of all factors from both experiment design methods were used to set a confirmation experiment. It was found that the colour shades, both visual at 15 and 45 angel of measurement degrees of spectrophotometer, were nearly closed to the target and the defective at
Rahmani, Mashaallah; Kaykhaii, Massoud; Sasani, Mojtaba
2018-01-05
This study aimed to investigate the efficiency of 3A zeolite as a novel adsorbent for removal of Rhodamine B and Malachite green dyes from water samples. To increase the removal efficiency, effecting parameters on adsorption process were investigated and optimized by adopting Taguchi design of experiments approach. The percentage contribution of each parameter on the removal of Rhodamine B and Malachite green dyes determined using ANOVA and showed that the most effective parameters in removal of RhB and MG by 3A zeolite are initial concentration of dye and pH, respectively. Under optimized condition, the amount predicted by Taguchi design method and the value obtained experimentally, showed good closeness (more than 94.86%). Good adsorption efficiency obtained for proposed methods indicates that, the 3A zeolite is capable to remove the significant amounts of Rhodamine B and Malachite green from environmental water samples. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Ravi, D.; Parammasivam, K. M.
2016-09-01
Numerical investigations were conducted on a turbine cascade, with end-wall cooling by a single row of cylindrical holes, inclined at 30°. The mainstream fluid was hot air and the coolant was CO2 gas. Based on the Reynolds number, the flow was turbulent at the inlet. The film hole row position, its pitch and blowing ratio was varied with five different values. Taguchi approach was used in designing a L25 orthogonal array (OA) for these parameters. The end-wall averaged film cooling effectiveness (bar η) was chosen as the quality characteristic. CFD analyses were carried out using Ansys Fluent on computational domains designed with inputs from OA. Experiments were conducted for one chosen OA configuration and the computational results were found to correlate well with experimental measurements. The responses from the CFD analyses were fed to the statistical tool to develop a correlation for bar η using regression analysis.
Pandey, Rupesh Kumar; Panda, Sudhansu Sekhar
2014-11-01
Drilling of bone is a common procedure in orthopedic surgery to produce hole for screw insertion to fixate the fracture devices and implants. The increase in temperature during such a procedure increases the chances of thermal invasion of bone which can cause thermal osteonecrosis resulting in the increase of healing time or reduction in the stability and strength of the fixation. Therefore, drilling of bone with minimum temperature is a major challenge for orthopedic fracture treatment. This investigation discusses the use of fuzzy logic and Taguchi methodology for predicting and minimizing the temperature produced during bone drilling. The drilling experiments have been conducted on bovine bone using Taguchi's L25 experimental design. A fuzzy model is developed for predicting the temperature during orthopedic drilling as a function of the drilling process parameters (point angle, helix angle, feed rate and cutting speed). Optimum bone drilling process parameters for minimizing the temperature are determined using Taguchi method. The effect of individual cutting parameters on the temperature produced is evaluated using analysis of variance. The fuzzy model using triangular and trapezoidal membership predicts the temperature within a maximum error of ±7%. Taguchi analysis of the obtained results determined the optimal drilling conditions for minimizing the temperature as A3B5C1.The developed system will simplify the tedious task of modeling and determination of the optimal process parameters to minimize the bone drilling temperature. It will reduce the risk of thermal osteonecrosis and can be very effective for the online condition monitoring of the process. © IMechE 2014.
Optimization of radial-type superconducting magnetic bearing using the Taguchi method
NASA Astrophysics Data System (ADS)
Ai, Liwang; Zhang, Guomin; Li, Wanjie; Liu, Guole; Liu, Qi
2018-07-01
It is important and complicated to model and optimize the levitation behavior of superconducting magnetic bearing (SMB). That is due to the nonlinear constitutive relationships of superconductor and ferromagnetic materials, the relative movement between the superconducting stator and PM rotor, and the multi-parameter (e.g., air-gap, critical current density, and remanent flux density, etc.) affecting the levitation behavior. In this paper, we present a theoretical calculation and optimization method of the levitation behavior for radial-type SMB. A simplified model of levitation force calculation is established using 2D finite element method with H-formulation. In the model, the boundary condition of superconducting stator is imposed by harmonic series expressions to describe the traveling magnetic field generated by the moving PM rotor. Also, experimental measurements of the levitation force are performed and validate the model method. A statistical method called Taguchi method is adopted to carry out an optimization of load capacity for SMB. Then the factor effects of six optimization parameters on the target characteristics are discussed and the optimum parameters combination is determined finally. The results show that the levitation behavior of SMB is greatly improved and the Taguchi method is suitable for optimizing the SMB.
Study of optimal laser parameters for cutting QFN packages by Taguchi's matrix method
NASA Astrophysics Data System (ADS)
Li, Chen-Hao; Tsai, Ming-Jong; Yang, Ciann-Dong
2007-06-01
This paper reports the study of optimal laser parameters for cutting QFN (Quad Flat No-lead) packages by using a diode pumped solid-state laser system (DPSSL). The QFN cutting path includes two different materials, which are the encapsulated epoxy and a copper lead frame substrate. The Taguchi's experimental method with orthogonal array of L 9(3 4) is employed to obtain optimal combinatorial parameters. A quantified mechanism was proposed for examining the laser cutting quality of a QFN package. The influences of the various factors such as laser current, laser frequency, and cutting speed on the laser cutting quality is also examined. From the experimental results, the factors on the cutting quality in the order of decreasing significance are found to be (a) laser frequency, (b) cutting speed, and (c) laser driving current. The optimal parameters were obtained at the laser frequency of 2 kHz, the cutting speed of 2 mm/s, and the driving current of 29 A. Besides identifying this sequence of dominance, matrix experiment also determines the best level for each control factor. The verification experiment confirms that the application of laser cutting technology to QFN is very successfully by using the optimal laser parameters predicted from matrix experiments.
Panda, Jibitesh Kumar; Sastry, Gadepalli Ravi Kiran; Rai, Ram Naresh
2018-05-25
The energy situation and the concerns about global warming nowadays have ignited research interest in non-conventional and alternative fuel resources to decrease the emission and the continuous dependency on fossil fuels, particularly for various sectors like power generation, transportation, and agriculture. In the present work, the research is focused on evaluating the performance, emission characteristics, and combustion of biodiesel such as palm kernel methyl ester with the addition of diesel additive "triacetin" in it. A timed manifold injection (TMI) system was taken up to examine the influence of durations of several blends induced on the emission and performance characteristics as compared to normal diesel operation. This experimental study shows better performance and releases less emission as compared with mineral diesel and in turn, indicates that high performance and low emission is promising in PKME-triacetin fuel operation. This analysis also attempts to describe the application of the fuzzy logic-based Taguchi analysis to optimize the emission and performance parameters.
NASA Astrophysics Data System (ADS)
Lee, Yu-Cheng; Yen, Tieh-Min; Tsai, Chih-Hung
This study provides an integrated model of Supplier Quality Performance Assesment (SQPA) activity for the semiconductor industry through introducing the ISO 9001 management framework, Importance-Performance Analysis (IPA) Supplier Quality Performance Assesment and Taguchi`s Signal-to-Noise Ratio (S/N) techniques. This integrated model provides a SQPA methodology to create value for all members under mutual cooperation and trust in the supply chain. This method helps organizations build a complete SQPA framework, linking organizational objectives and SQPA activities to optimize rating techniques to promote supplier quality improvement. The techniques used in SQPA activities are easily understood. A case involving a design house is illustrated to show our model.
Castorena-Cortés, G; Roldán-Carrillo, T; Zapata-Peñasco, I; Reyes-Avila, J; Quej-Aké, L; Marín-Cruz, J; Olguín-Lora, P
2009-12-01
Microcosm assays and Taguchi experimental design was used to assess the biodegradation of an oil sludge produced by a gas processing unit. The study showed that the biodegradation of the sludge sample is feasible despite the high level of pollutants and complexity involved in the sludge. The physicochemical and microbiological characterization of the sludge revealed a high concentration of hydrocarbons (334,766+/-7001 mg kg(-1) dry matter, d.m.) containing a variety of compounds between 6 and 73 carbon atoms in their structure, whereas the concentration of Fe was 60,000 mg kg(-1) d.m. and 26,800 mg kg(-1) d.m. of sulfide. A Taguchi L(9) experimental design comprising 4 variables and 3 levels moisture, nitrogen source, surfactant concentration and oxidant agent was performed, proving that moisture and nitrogen source are the major variables that affect CO(2) production and total petroleum hydrocarbons (TPH) degradation. The best experimental treatment yielded a TPH removal of 56,092 mg kg(-1) d.m. The treatment was carried out under the following conditions: 70% moisture, no oxidant agent, 0.5% of surfactant and NH(4)Cl as nitrogen source.
NASA Astrophysics Data System (ADS)
Chandrakanth, Balaji; Venkatesan, G; Prakash Kumar, L. S. S; Jalihal, Purnima; Iniyan, S
2018-03-01
The present work discusses the design and selection of a shell and tube condenser used in Low Temperature Thermal Desalination (LTTD). To optimize the key geometrical and process parameters of the condenser with multiple parameters and levels, a design of an experiment approach using Taguchi method was chosen. An orthogonal array (OA) of 25 designs was selected for this study. The condenser was designed, analysed using HTRI software and the heat transfer area with respective tube side pressure drop were computed using the same, as these two objective functions determine the capital and running cost of the condenser. There was a complex trade off between the heat transfer area and pressure drop in the analysis, however second law analysis was worked out for determining the optimal heat transfer area vs pressure drop for condensing the required heat load.
NASA Astrophysics Data System (ADS)
Rodriguez, Tony F.; Cushman, David A.
2003-06-01
With the growing commercialization of watermarking techniques in various application scenarios it has become increasingly important to quantify the performance of watermarking products. The quantification of relative merits of various products is not only essential in enabling further adoption of the technology by society as a whole, but will also drive the industry to develop testing plans/methodologies to ensure quality and minimize cost (to both vendors & customers.) While the research community understands the theoretical need for a publicly available benchmarking system to quantify performance, there has been less discussion on the practical application of these systems. By providing a standard set of acceptance criteria, benchmarking systems can dramatically increase the quality of a particular watermarking solution, validating the product performances if they are used efficiently and frequently during the design process. In this paper we describe how to leverage specific design of experiments techniques to increase the quality of a watermarking scheme, to be used with the benchmark tools being developed by the Ad-Hoc Watermark Verification Group. A Taguchi Loss Function is proposed for an application and orthogonal arrays used to isolate optimal levels for a multi-factor experimental situation. Finally, the results are generalized to a population of cover works and validated through an exhaustive test.
Quality in the Operational Air Force: A Case of Misplaced Emphasis
1994-05-01
other quality advocates of the era. These men included Joseph Juran, Armand Feigenbaum, Kaoru Ishikawa , and Genichi Taguchi. Juran contributed disciplined...planning theories, while Feigenbaum felt that producing quality could actually reduce production costs. In addition, Ishikawa and Taguchi lent...statistically based problem solving techniques, but the more modem approaches of Ishikawa , Taguchi and others. The operative concept of TQM is ’continuous
Taguchi Method Applied in Optimization of Shipley SJR 5740 Positive Resist Deposition
NASA Technical Reports Server (NTRS)
Hui, A.; Blosiu, J. O.; Wiberg, D. V.
1998-01-01
Taguchi Methods of Robust Design presents a way to optimize output process performance through an organized set of experiments by using orthogonal arrays. Analysis of variance and signal-to-noise ratio is used to evaluate the contribution of each of the process controllable parameters in the realization of the process optimization. In the photoresist deposition process, there are numerous controllable parameters that can affect the surface quality and thickness of the final photoresist layer.
NASA Astrophysics Data System (ADS)
Ghosh, Nabendu; Kumar, Pradip; Nandi, Goutam
2016-10-01
Welding input process parameters play a very significant role in determining the quality of the welded joint. Only by properly controlling every element of the process can product quality be controlled. For better quality of MIG welding of Ferritic stainless steel AISI 409, precise control of process parameters, parametric optimization of the process parameters, prediction and control of the desired responses (quality indices) etc., continued and elaborate experiments, analysis and modeling are needed. A data of knowledge - base may thus be generated which may be utilized by the practicing engineers and technicians to produce good quality weld more precisely, reliably and predictively. In the present work, X-ray radiographic test has been conducted in order to detect surface and sub-surface defects of weld specimens made of Ferritic stainless steel. The quality of the weld has been evaluated in terms of yield strength, ultimate tensile strength and percentage of elongation of the welded specimens. The observed data have been interpreted, discussed and analyzed by considering ultimate tensile strength ,yield strength and percentage elongation combined with use of Grey-Taguchi methodology.
Thammarakcharoen, Faungchat; Suvannapruk, Waraporn; Suwanprateeb, Jintamai
2014-10-01
In this study, a statistical design of experimental methodology based on Taguchi orthogonal design has been used to study the effect of various processing parameters on the amount of calcium phosphate coating produced by such technique. Seven control factors with three levels each including sodium hydroxide concentration, pretreatment temperature, pretreatment time, cleaning method, coating time, coating temperature and surface area to solution volume ratio were studied. X-ray diffraction revealed that all the coatings consisted of the mixture of octacalcium phosphate (OCP) and hydroxyapatite (HA) and the presence of each phase depended on the process conditions used. Various content and size (-1-100 μm) of isolated spheroid particles with nanosized plate-like morphology deposited on the titanium surface or a continuous layer of plate-like nanocrystals having the plate thickness in the range of -100-300 nm and the plate width in the range of 3-8 μm were formed depending on the process conditions employed. The optimum condition of using sodium hydroxide concentration of 1 M, pretreatment temperature of 70 degrees C, pretreatment time of 24 h, cleaning by ultrasonic, coating time of 6 h, coating temperature of 50 degrees C and surface area to solution volume ratio of 32.74 for producing the greatest amount of the coating formed on the titanium surface was predicted and validated. In addition, coating temperature was found to be the dominant factor with the greatest contribution to the coating formation while coating time and cleaning method were significant factors. Other factors had negligible effects on the coating performance.
Thermochemical hydrolysis of macroalgae Ulva for biorefinery: Taguchi robust design method
NASA Astrophysics Data System (ADS)
Jiang, Rui; Linzon, Yoav; Vitkin, Edward; Yakhini, Zohar; Chudnovsky, Alexandra; Golberg, Alexander
2016-06-01
Understanding the impact of all process parameters on the efficiency of biomass hydrolysis and on the final yield of products is critical to biorefinery design. Using Taguchi orthogonal arrays experimental design and Partial Least Square Regression, we investigated the impact of change and the comparative significance of thermochemical process temperature, treatment time, %Acid and %Solid load on carbohydrates release from green macroalgae from Ulva genus, a promising biorefinery feedstock. The average density of hydrolysate was determined using a new microelectromechanical optical resonator mass sensor. In addition, using Flux Balance Analysis techniques, we compared the potential fermentation yields of these hydrolysate products using metabolic models of Escherichia coli, Saccharomyces cerevisiae wild type, Saccharomyces cerevisiae RN1016 with xylose isomerase and Clostridium acetobutylicum. We found that %Acid plays the most significant role and treatment time the least significant role in affecting the monosaccharaides released from Ulva biomass. We also found that within the tested range of parameters, hydrolysis with 121 °C, 30 min 2% Acid, 15% Solids could lead to the highest yields of conversion: 54.134-57.500 gr ethanol kg-1 Ulva dry weight by S. cerevisiae RN1016 with xylose isomerase. Our results support optimized marine algae utilization process design and will enable smart energy harvesting by thermochemical hydrolysis.
Design of Maternity Pillow by Using Kansei and Taguchi Methods
NASA Astrophysics Data System (ADS)
Ilma Rahmillah, Fety; Nanda kartika, Rachmah
2017-06-01
One of the customers’ considerations for purchasing a product is it can satisfy their feeling and emotion. It because of such product can enhance sleep quality of pregnant women. However, most of the existing product such as maternity pillows are still designed based on companies’ perspective. This study aims to capture the desire of pregnant women toward maternity pillow desired product by using kansei words and analyze the optimal design with Taguchi method. Eight collected kansei words were durable, aesthetic, comfort, portable, simple, multifunction, attractive motive, and easy to maintain. While L16 orthogonal array is used because there are three variables with two levels and four variables with four levels. It can be concluded that the best maternity pillow that can satisfy the customers can be designed by combining D1-E2-F2-G2-C1-B2-A2 means the model is U shape, flowery motive, medium color, Bag model B, cotton pillow cover, filled with silicon, and use double zipper. However, it is also possible to create combination of D1-E2-F2-G2-C1-B1-A1 by using consideration of cost which means that the zipper is switched to single as well as filled with dacron. In addition, the total percentage of contribution by using ANOVA reaches 95%.
NASA Astrophysics Data System (ADS)
Menon, P. S.; Kandiah, K.; Mandeep, J. S.; Shaari, S.; Apte, P. R.
Long-wavelength VCSELs (LW-VCSEL) operating in the 1.55 μm wavelength regime offer the advantages of low dispersion and optical loss in fiber optic transmission systems which are crucial in increasing data transmission speed and reducing implementation cost of fiber-to-the-home (FTTH) access networks. LW-VCSELs are attractive light sources because they offer unique features such as low power consumption, narrow beam divergence and ease of fabrication for two-dimensional arrays. This paper compares the near field and far field effects of the numerically investigated LW-VCSEL for various design parameters of the device. The optical intensity profile far from the device surface, in the Fraunhofer region, is important for the optical coupling of the laser with other optical components. The near field pattern is obtained from the structure output whereas the far-field pattern is essentially a two-dimensional fast Fourier Transform (FFT) of the near-field pattern. Design parameters such as the number of wells in the multi-quantum-well (MQW) region, the thickness of the MQW and the effect of using Taguchi's orthogonal array method to optimize the device design parameters on the near/far field patterns are evaluated in this paper. We have successfully increased the peak lasing power from an initial 4.84 mW to 12.38 mW at a bias voltage of 2 V and optical wavelength of 1.55 μm using Taguchi's orthogonal array. As a result of the Taguchi optimization and fine tuning, the device threshold current is found to increase along with a slight decrease in the modulation speed due to increased device widths.
Chenthamarakshan, Aiswarya; Parambayil, Nayana; Miziriya, Nafeesathul; Soumya, P S; Lakshmi, M S Kiran; Ramgopal, Anala; Dileep, Anuja; Nambisan, Padma
2017-02-13
Fungal laccase has profound applications in different fields of biotechnology due to its broad specificity and high redox potential. Any successful application of the enzyme requires large scale production. As laccase production is highly dependent on medium components and cultural conditions, optimization of the same is essential for efficient product production. Production of laccase by fungal strain Marasmiellus palmivorus LA1 under solid state fermentation was optimized by the Taguchi design of experiments (DOE) methodology. An orthogonal array (L8) was designed using Qualitek-4 software to study the interactions and relative influence of the seven selected factors by one factor at a time approach. The optimum condition formulated was temperature (28 °C), pH (5), galactose (0.8%w/v), cupric sulphate (3 mM), inoculum concentration (number of mycelial agar pieces) (6Nos.) and substrate length (0.05 m). Overall yield increase of 17.6 fold was obtained after optimization. Statistical optimization leads to the elimination of an insignificant medium component ammonium dihydrogen phosphate from the process and contributes to a 1.06 fold increase in enzyme production. A final production of 667.4 ± 13 IU/mL laccase activity paves way for the application of this strain for industrial applications. Study optimized lignin degrading laccases from Marasmiellus palmivorus LA1. This laccases can thus be used for further applications in different scales of production after analyzing the properties of the enzyme. Study also confirmed the use of taguchi method for optimizations of product production.
NASA Astrophysics Data System (ADS)
Kowalczyk, M.
2017-08-01
This paper describes the research results of surface quality research after the NiTi shape memory alloy (Nitinol) precise turning by the tools with edges made of polycrystalline diamonds (PCD). Nitinol, a nearly equiatomic nickel-titanium shape memory alloy, has wide applications in the arms industry, military, medicine and aerospace industry, and industrial robots. Due to their specific properties NiTi alloys are known to be difficult-to-machine materials particularly by using conventional techniques. The research trials were conducted for three independent parameters (vc, f, ap) affecting the surface roughness were analyzed. The choice of parameter configurations were performed by factorial design methods using orthogonal plan type L9, with three control factors, changing on three levels, developed by G. Taguchi. S/N ratio and ANOVA analyses were performed to identify the best of cutting parameters influencing surface roughness.
Bai, Mingsian R; Tung, Chih-Wei; Lee, Chih-Chung
2005-05-01
An optimal design technique of loudspeaker arrays for cross-talk cancellation with application in three-dimensional audio is presented. An array focusing scheme is presented on the basis of the inverse propagation that relates the transducers to a set of chosen control points. Tikhonov regularization is employed in designing the inverse cancellation filters. An extensive analysis is conducted to explore the cancellation performance and robustness issues. To best compromise the performance and robustness of the cross-talk cancellation system, optimal configurations are obtained with the aid of the Taguchi method and the genetic algorithm (GA). The proposed systems are further justified by physical as well as subjective experiments. The results reveal that large number of loudspeakers, closely spaced configuration, and optimal control point design all contribute to the robustness of cross-talk cancellation systems (CCS) against head misalignment.
NASA Astrophysics Data System (ADS)
Yousefieh, M.; Shamanian, M.; Saatchi, A.
2012-09-01
Taguchi design method with L9 orthogonal array was implemented to optimize the pulsed current gas tungsten arc welding parameters for the hardness and the toughness of super duplex stainless steel (SDSS, UNS S32760) welds. In this regard, the hardness and the toughness were considered as performance characteristics. Pulse current, background current, % on time, and pulse frequency were chosen as main parameters. Each parameter was varied at three different levels. As a result of pooled analysis of variance, the pulse current is found to be the most significant factor for both the hardness and the toughness of SDSS welds by percentage contribution of 71.81 for hardness and 78.18 for toughness. The % on time (21.99%) and the background current (17.81%) had also the next most significant effect on the hardness and the toughness, respectively. The optimum conditions within the selected parameter values for hardness were found as the first level of pulse current (100 A), third level of background current (70 A), first level of % on time (40%), and first level of pulse frequency (1 Hz), while they were found as the second level of pulse current (120 A), second level of background current (60 A), second level of % on time (60%), and third level of pulse frequency (5 Hz) for toughness. The Taguchi method was found to be a promising tool to obtain the optimum conditions for such studies. Finally, in order to verify experimental results, confirmation tests were carried out at optimum working conditions. Under these conditions, there were good agreements between the predicted and the experimental results for the both hardness and toughness.
Thermochemical hydrolysis of macroalgae Ulva for biorefinery: Taguchi robust design method
Jiang, Rui; Linzon, Yoav; Vitkin, Edward; Yakhini, Zohar; Chudnovsky, Alexandra; Golberg, Alexander
2016-01-01
Understanding the impact of all process parameters on the efficiency of biomass hydrolysis and on the final yield of products is critical to biorefinery design. Using Taguchi orthogonal arrays experimental design and Partial Least Square Regression, we investigated the impact of change and the comparative significance of thermochemical process temperature, treatment time, %Acid and %Solid load on carbohydrates release from green macroalgae from Ulva genus, a promising biorefinery feedstock. The average density of hydrolysate was determined using a new microelectromechanical optical resonator mass sensor. In addition, using Flux Balance Analysis techniques, we compared the potential fermentation yields of these hydrolysate products using metabolic models of Escherichia coli, Saccharomyces cerevisiae wild type, Saccharomyces cerevisiae RN1016 with xylose isomerase and Clostridium acetobutylicum. We found that %Acid plays the most significant role and treatment time the least significant role in affecting the monosaccharaides released from Ulva biomass. We also found that within the tested range of parameters, hydrolysis with 121 °C, 30 min 2% Acid, 15% Solids could lead to the highest yields of conversion: 54.134–57.500 gr ethanol kg−1 Ulva dry weight by S. cerevisiae RN1016 with xylose isomerase. Our results support optimized marine algae utilization process design and will enable smart energy harvesting by thermochemical hydrolysis. PMID:27291594
NASA Technical Reports Server (NTRS)
Deal, Don E.
1991-01-01
The chief goals of the summer project have been twofold - first, for my host group and myself to learn as much of the working details of Taguchi analysis as possible in the time allotted, and, secondly, to apply the methodology to a design problem with the intention of establishing a preliminary set of near-optimal (in the sense of producing a desired response) design parameter values from among a large number of candidate factor combinations. The selected problem is concerned with determining design factor settings for an automated approach program which is to have the capability of guiding the Shuttle into the docking port of the Space Station under controlled conditions so as to meet and/or optimize certain target criteria. The candidate design parameters under study were glide path (i.e., approach) angle, path intercept and approach gains, and minimum impulse bit mode (a parameter which defines how Shuttle jets shall be fired). Several performance criteria were of concern: terminal relative velocity at the instant the two spacecraft are mated; docking offset; number of Shuttle jet firings in certain specified directions (of interest due to possible plume impingement on the Station's solar arrays), and total RCS (a measure of the energy expended in performing the approach/docking maneuver). In the material discussed here, we have focused on single performance criteria - total RCS. An analysis of the possibility of employing a multiobjective function composed of a weighted sum of the various individual criteria has been undertaken, but is, at this writing, incomplete. Results from the Taguchi statistical analysis indicate that only three of the original four posited factors are significant in affecting RCS response. A comparison of model simulation output (via Monte Carlo) with predictions based on estimated factor effects inferred through the Taguchi experiment array data suggested acceptable or close agreement between the two except at the predicted optimum
Ashengroph, Morahem; Ababaf, Sajad
2014-12-01
Microbial caffeine removal is a green solution for treatment of caffeinated products and agro-industrial effluents. We directed this investigation to optimizing a bio-decaffeination process with growing cultures of Pseudomonas pseudoalcaligenes through Taguchi methodology which is a structured statistical approach that can be lowered variations in a process through Design of Experiments (DOE). Five parameters, i.e. initial fructose, tryptone, Zn(+2) ion and caffeine concentrations and also incubation time selected and an L16 orthogonal array was applied to design experiments with four 4-level factors and one 3-level factor (4(4) × 1(3)). Data analysis was performed using the statistical analysis of variance (ANOVA) method. Furthermore, the optimal conditions were determined by combining the optimal levels of the significant factors and verified by a confirming experiment. Measurement of residual caffeine concentration in the reaction mixture was performed using high-performance liquid chromatography (HPLC). Use of Taguchi methodology for optimization of design parameters resulted in about 86.14% reduction of caffeine in 48 h incubation when 5g/l fructose, 3 mM Zn(+2) ion and 4.5 g/l of caffeine are present in the designed media. Under the optimized conditions, the yield of degradation of caffeine (4.5 g/l) by the native strain of Pseudomonas pseudoalcaligenes TPS8 has been increased from 15.8% to 86.14% which is 5.4 fold higher than the normal yield. According to the experimental results, Taguchi methodology provides a powerful methodology for identifying the favorable parameters on caffeine removal using strain TPS8 which suggests that the approach also has potential application with similar strains to improve the yield of caffeine removal from caffeine containing solutions.
Chen, How-Ji; Chang, Sheng-Nan; Tang, Chao-Wei
2017-01-01
This study aimed to apply the Taguchi optimization technique to determine the process conditions for producing synthetic lightweight aggregate (LWA) by incorporating tile grinding sludge powder with reservoir sediments. An orthogonal array L16(45) was adopted, which consisted of five controllable four-level factors (i.e., sludge content, preheat temperature, preheat time, sintering temperature, and sintering time). Moreover, the analysis of variance method was used to explore the effects of the experimental factors on the particle density, water absorption, bloating ratio, and loss on ignition of the produced LWA. Overall, the produced aggregates had particle densities ranging from 0.43 to 2.1 g/cm3 and water absorption ranging from 0.6% to 13.4%. These values are comparable to the requirements for ordinary and high-performance LWAs. The results indicated that it is considerably feasible to produce high-performance LWA by incorporating tile grinding sludge with reservoir sediments. PMID:29125576
Chen, How-Ji; Chang, Sheng-Nan; Tang, Chao-Wei
2017-11-10
This study aimed to apply the Taguchi optimization technique to determine the process conditions for producing synthetic lightweight aggregate (LWA) by incorporating tile grinding sludge powder with reservoir sediments. An orthogonal array L 16 (4⁵) was adopted, which consisted of five controllable four-level factors (i.e., sludge content, preheat temperature, preheat time, sintering temperature, and sintering time). Moreover, the analysis of variance method was used to explore the effects of the experimental factors on the particle density, water absorption, bloating ratio, and loss on ignition of the produced LWA. Overall, the produced aggregates had particle densities ranging from 0.43 to 2.1 g/cm³ and water absorption ranging from 0.6% to 13.4%. These values are comparable to the requirements for ordinary and high-performance LWAs. The results indicated that it is considerably feasible to produce high-performance LWA by incorporating tile grinding sludge with reservoir sediments.
Kumar, Mukesh; Singh, Amrinder; Beniwal, Vikas; Salar, Raj Kumar
2016-12-01
Tannase (tannin acyl hydrolase E.C 3.1.1.20) is an inducible, largely extracellular enzyme that causes the hydrolysis of ester and depside bonds present in various substrates. Large scale industrial application of this enzyme is very limited owing to its high production costs. In the present study, cost effective production of tannase by Klebsiella pneumoniae KP715242 was studied under submerged fermentation using different tannin rich agro-residues like Indian gooseberry leaves (Phyllanthus emblica), Black plum leaves (Syzygium cumini), Eucalyptus leaves (Eucalyptus glogus) and Babul leaves (Acacia nilotica). Among all agro-residues, Indian gooseberry leaves were found to be the best substrate for tannase production under submerged fermentation. Sequential optimization approach using Taguchi orthogonal array screening and response surface methodology was adopted to optimize the fermentation variables in order to enhance the enzyme production. Eleven medium components were screened primarily by Taguchi orthogonal array design to identify the most contributing factors towards the enzyme production. The four most significant contributing variables affecting tannase production were found to be pH (23.62 %), tannin extract (20.70 %), temperature (20.33 %) and incubation time (14.99 %). These factors were further optimized with central composite design using response surface methodology. Maximum tannase production was observed at 5.52 pH, 39.72 °C temperature, 91.82 h of incubation time and 2.17 % tannin content. The enzyme activity was enhanced by 1.26 fold under these optimized conditions. The present study emphasizes the use of agro-residues as a potential substrate with an aim to lower down the input costs for tannase production so that the enzyme could be used proficiently for commercial purposes.
Oily wastewater treatment by ultrafiltration using Taguchi experimental design.
Salahi, A; Mohammadi, T
2011-01-01
In this research, results of an experimental investigation on separation of oil from a real oily wastewater using an ultrafiltration (UF) polymeric membrane are presented. In order to enhance the performance of UF in API separator effluent treatment and to get more permeation flux (PF), effects of operating factors on the yield of PF were studied. Five factors at four levels were investigated: trans-membrane pressure (TMP), temperature (T), cross flow velocity (CFV), pH and salt concentration (SC). Taguchi method (L(16) orthogonal array (OA)) was used. Analysis of variance (ANOVA) was applied to calculate sum of square, variance, error variance and contribution percentage of each factor on response. The optimal levels thus determined for the four influential factors were: TMP, 3 bar; T, 40˚C; CFV, 1.0 m/s; SC, 25 g/L and pH, 8. The results showed that CFV and SC are the most and the least effective factors on PF, respectively. Increasing CFV, TMP, T and pH caused the better performance of UF membrane process due to enhancement of driving force and fouling residence. Also, effects of oil concentration (OC) in the wastewater on PF and total organic carbon (TOC) rejection were investigated. Finally, the highest TOC rejection was found to be 85%.
Taguchi approach for co-gasification optimization of torrefied biomass and coal.
Chen, Wei-Hsin; Chen, Chih-Jung; Hung, Chen-I
2013-09-01
This study employs the Taguchi method to approach the optimum co-gasification operation of torrefied biomass (eucalyptus) and coal in an entrained flow gasifier. The cold gas efficiency is adopted as the performance index of co-gasification. The influences of six parameters, namely, the biomass blending ratio, oxygen-to-fuel mass ratio (O/F ratio), biomass torrefaction temperature, gasification pressure, steam-to-fuel mass ratio (S/F ratio), and inlet temperature of the carrier gas, on the performance of co-gasification are considered. The analysis of the signal-to-noise ratio suggests that the O/F ratio is the most important factor in determining the performance and the appropriate O/F ratio is 0.7. The performance is also significantly affected by biomass along with torrefaction, where a torrefaction temperature of 300°C is sufficient to upgrade eucalyptus. According to the recommended operating conditions, the values of cold gas efficiency and carbon conversion at the optimum co-gasification are 80.99% and 94.51%, respectively. Copyright © 2013 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Kar, Siddhartha; Chakraborty, Sujoy; Dey, Vidyut; Ghosh, Subrata Kumar
2017-10-01
This paper investigates the application of Taguchi method with fuzzy logic for multi objective optimization of roughness parameters in electro discharge coating process of Al-6351 alloy with powder metallurgical compacted SiC/Cu tool. A Taguchi L16 orthogonal array was employed to investigate the roughness parameters by varying tool parameters like composition and compaction load and electro discharge machining parameters like pulse-on time and peak current. Crucial roughness parameters like Centre line average roughness, Average maximum height of the profile and Mean spacing of local peaks of the profile were measured on the coated specimen. The signal to noise ratios were fuzzified to optimize the roughness parameters through a single comprehensive output measure (COM). Best COM obtained with lower values of compaction load, pulse-on time and current and 30:70 (SiC:Cu) composition of tool. Analysis of variance is carried out and a significant COM model is observed with peak current yielding highest contribution followed by pulse-on time, compaction load and composition. The deposited layer is characterised by X-Ray Diffraction analysis which confirmed the presence of tool materials on the work piece surface.
A Taguchi approach on optimal process control parameters for HDPE pipe extrusion process
NASA Astrophysics Data System (ADS)
Sharma, G. V. S. S.; Rao, R. Umamaheswara; Rao, P. Srinivasa
2017-06-01
High-density polyethylene (HDPE) pipes find versatile applicability for transportation of water, sewage and slurry from one place to another. Hence, these pipes undergo tremendous pressure by the fluid carried. The present work entails the optimization of the withstanding pressure of the HDPE pipes using Taguchi technique. The traditional heuristic methodology stresses on a trial and error approach and relies heavily upon the accumulated experience of the process engineers for determining the optimal process control parameters. This results in setting up of less-than-optimal values. Hence, there arouse a necessity to determine optimal process control parameters for the pipe extrusion process, which can ensure robust pipe quality and process reliability. In the proposed optimization strategy, the design of experiments (DoE) are conducted wherein different control parameter combinations are analyzed by considering multiple setting levels of each control parameter. The concept of signal-to-noise ratio ( S/ N ratio) is applied and ultimately optimum values of process control parameters are obtained as: pushing zone temperature of 166 °C, Dimmer speed at 08 rpm, and Die head temperature to be 192 °C. Confirmation experimental run is also conducted to verify the analysis and research result and values proved to be in synchronization with the main experimental findings and the withstanding pressure showed a significant improvement from 0.60 to 1.004 Mpa.
Karev, George B
2006-02-01
When assessing the relationship between direction and orientation in drawings by young children and adults, Taguchi and Noma used a fish-drawing task. However, the fish is not convenient enough as an object for such a task so it is highly preferable to use, instead of a single object, a set of several objects to assess directionality quantitatively. These authors' conclusions do not acknowledge alternative explanations.
Villarroel, Mario; Castro, Ruth; Junod, Julio
2003-06-01
The goal of this present study was the development of an optimized formula of damask marmalade low in calories applying Taguchi methodology to improve the quality of this product. The selection of this methodology lies on the fact that in real life conditions the result of an experiment frequently depends on the influence of several variables, therefore, one expedite way to solve this problem is utilizing factorial desings. The influence of acid, thickener, sweetener and aroma additives, as well as time of cooking, and possible interactions among some of them, were studied trying to get the best combination of these factors to optimize the sensorial quality of an experimental formulation of dietetic damask marmalade. An orthogonal array L8 (2(7)) was applied in this experience, as well as level average analysis was carried out according Taguchi methodology to determine the suitable working levels of the design factors previously choiced, to achieve a desirable product quality. A sensory trained panel was utilized to analyze the marmalade samples using a composite scoring test with a descriptive acuantitative scale ranging from 1 = Bad, 5 = Good. It was demonstrated that the design factors sugar/aspartame, pectin and damask aroma had a significant effect (p < 0.05) on the sensory quality of the marmalade with 82% of contribution on the response. The optimal combination result to be: citric acid 0.2%; pectin 1%; 30 g sugar/16 mg aspartame/100 g, damask aroma 0.5 ml/100 g, time of cooking 5 minutes. Regarding chemical composition, the most important results turned out to be the decrease in carbohydrate content compaired with traditional marmalade with a reduction of 56% in coloric value and also the amount of dietary fiber greater than similar commercial products. Assays of storage stability were carried out on marmalade samples submitted to different temperatures held in plastic bags of different density. Non percetible sensorial, microbiological and chemical changes
Hemmat, Abbas; Kafashan, Jalal; Huang, Hongying
2017-01-01
To study the optimum process conditions for pretreatments and anaerobic codigestion of oil refinery wastewater (ORWW) with chicken manure, L9 (34) Taguchi's orthogonal array was applied. The biogas production (BGP), biomethane content (BMP), and chemical oxygen demand solubilization (CODS) in stabilization rate were evaluated as the process outputs. The optimum conditions were obtained by using Design Expert software (Version 7.0.0). The results indicated that the optimum conditions could be achieved with 44% ORWW, 36°C temperature, 30 min sonication, and 6% TS in the digester. The optimum BGP, BMP, and CODS removal rates by using the optimum conditions were 294.76 mL/gVS, 151.95 mL/gVS, and 70.22%, respectively, as concluded by the experimental results. In addition, the artificial neural network (ANN) technique was implemented to develop an ANN model for predicting BGP yield and BMP content. The Levenberg-Marquardt algorithm was utilized to train ANN, and the architecture of 9-19-2 for the ANN model was obtained. PMID:29441352
Review of Fluorescence-Based Velocimetry Techniques to Study High-Speed Compressible Flows
NASA Technical Reports Server (NTRS)
Bathel, Brett F.; Johansen, Criag; Inman, Jennifer A.; Jones, Stephen B.; Danehy, Paul M.
2013-01-01
This paper reviews five laser-induced fluorescence-based velocimetry techniques that have been used to study high-speed compressible flows at NASA Langley Research Center. The techniques discussed in this paper include nitric oxide (NO) molecular tagging velocimetry (MTV), nitrogen dioxide photodissociation (NO2-to-NO) MTV, and NO and atomic oxygen (O-atom) Doppler-shift-based velocimetry. Measurements of both single-component and two-component velocity have been performed using these techniques. This paper details the specific application and experiment for which each technique has been used, the facility in which the experiment was performed, the experimental setup, sample results, and a discussion of the lessons learned from each experiment.
Experimental Validation for Hot Stamping Process by Using Taguchi Method
NASA Astrophysics Data System (ADS)
Fawzi Zamri, Mohd; Lim, Syh Kai; Razlan Yusoff, Ahmad
2016-02-01
Due to the demand for reduction in gas emissions, energy saving and producing safer vehicles has driven the development of Ultra High Strength Steel (UHSS) material. To strengthen UHSS material such as boron steel, it needed to undergo a process of hot stamping for heating at certain temperature and time. In this paper, Taguchi method is applied to determine the appropriate parameter of thickness, heating temperature and heating time to achieve optimum strength of boron steel. The experiment is conducted by using flat square shape of hot stamping tool with tensile dog bone as a blank product. Then, the value of tensile strength and hardness is measured as response. The results showed that the lower thickness, higher heating temperature and heating time give the higher strength and hardness for the final product. In conclusion, boron steel blank are able to achieve up to 1200 MPa tensile strength and 650 HV of hardness.
Vertically aligned N-doped CNTs growth using Taguchi experimental design
NASA Astrophysics Data System (ADS)
Silva, Ricardo M.; Fernandes, António J. S.; Ferro, Marta C.; Pinna, Nicola; Silva, Rui F.
2015-07-01
The Taguchi method with a parameter design L9 orthogonal array was implemented for optimizing the nitrogen incorporation in the structure of vertically aligned N-doped CNTs grown by thermal chemical deposition (TCVD). The maximization of the ID/IG ratio of the Raman spectra was selected as the target value. As a result, the optimal deposition configuration was NH3 = 90 sccm, growth temperature = 825 °C and catalyst pretreatment time of 2 min, the first parameter having the main effect on nitrogen incorporation. A confirmation experiment with these values was performed, ratifying the predicted ID/IG ratio of 1.42. Scanning electron microscopy (SEM) characterization revealed a uniform completely vertically aligned array of multiwalled CNTs which individually exhibit a bamboo-like structure, consisting of periodically curved graphitic layers, as depicted by high resolution transmission electron microscopy (HRTEM). The X-ray photoelectron spectroscopy (XPS) results indicated a 2.00 at.% of N incorporation in the CNTs in pyridine-like and graphite-like, as the predominant species.
Preparation of nanocellulose from Imperata brasiliensis grass using Taguchi method.
Benini, Kelly Cristina Coelho de Carvalho; Voorwald, Herman Jacobus Cornelis; Cioffi, Maria Odila Hilário; Rezende, Mirabel Cerqueira; Arantes, Valdeir
2018-07-15
Cellulose nanoparticles (CNs) were prepared by acid hydrolysis of the cellulose pulp extracted from the Brazilian satintail (Imperata Brasiliensis) plant using a conventional and a total chlorine free method. Initially, a statistical design of experiment was carried out using Taguchi orthogonal array to study the hydrolysis parameters, and the main properties (crystallinity, thermal stability, morphology, and sizes) of the nanocellulose. X-ray diffraction (XRD), fourier-transform infrared spectroscopy (FTIR), field-emission scanning electron microscopy (FE-SEM), dynamic light scattering (DLS), zeta potential and thermogravimetric analysis (TGA) were carried out to characterize the physical-chemical properties of the CNs obtained. Cellulose nanoparticles with diameter ranging from 10 to 60 nm and length between 150 and 250 nm were successfully obtained at sulfuric acid concentration of 64% (m/m), temperature 35 °C, reaction time 75 min, and a 1:20 (g/mL) pulp-to-solution ratio. Under this condition, the Imperata Brasiliensis CNs showed good stability in suspension, crystallinity index of 65%, and a cellulose degradation temperature of about 117 °C. Considering that these properties are similar to those of nanocelluloses from other lignocellulosics feedstocks, Imperata grass seems also to be a suitable source for nanocellulose production. Copyright © 2018 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Nath, Nayani Kishore
2017-08-01
The throat back up liners is used to protect the nozzle structural members from the severe thermal environment in solid rocket nozzles. The throat back up liners is made with E-glass phenolic prepregs by tape winding process. The objective of this work is to demonstrate the optimization of process parameters of tape winding process to achieve better insulative resistance using Taguchi's robust design methodology. In this method four control factors machine speed, roller pressure, tape tension, tape temperature that were investigated for the tape winding process. The presented work was to study the cogency and acceptability of Taguchi's methodology in manufacturing of throat back up liners. The quality characteristic identified was Back wall temperature. Experiments carried out using L 9 ' (34) orthogonal array with three levels of four different control factors. The test results were analyzed using smaller the better criteria for Signal to Noise ratio in order to optimize the process. The experimental results were analyzed conformed and successfully used to achieve the minimum back wall temperature of the throat back up liners. The enhancement in performance of the throat back up liners was observed by carrying out the oxy-acetylene tests. The influence of back wall temperature on the performance of throat back up liners was verified by ground firing test.
NASA Astrophysics Data System (ADS)
Sudibyo, Hermida, L.; Suwardi
2017-11-01
Tapioca waste water is very difficult to treat; hence many tapioca factories could not treat it well. One of method which able to overcome this problem is electrodeposition. This process has high performance when it conducted using batch recycle process and use aluminum bipolar electrode. However, the optimum operation conditions are having a significant effect in the tapioca wastewater treatment using bath recycle process. In this research, The Taguchi method was successfully applied to know the optimum condition and the interaction between parameters in electrocoagulation process. The results show that current density, conductivity, electrode distance, and pH have a significant effect on the turbidity removal of cassava starch waste water.
Flank wear analysing of high speed end milling for hardened steel D2 using Taguchi Method
NASA Astrophysics Data System (ADS)
Hazza Faizi Al-Hazza, Muataz; Ibrahim, Nur Asmawiyah bt; Adesta, Erry T. Y.; Khan, Ahsan Ali; Abdullah Sidek, Atiah Bt.
2017-03-01
One of the main challenges for any manufacturer is how to decrease the machining cost without affecting the final quality of the product. One of the new advanced machining processes in industry is the high speed hard end milling process that merges three advanced machining processes: high speed milling, hard milling and dry milling. However, one of the most important challenges in this process is to control the flank wear rate. Therefore a analyzing the flank wear rate during machining should be investigated in order to determine the best cutting levels that will not affect the final quality of the product. In this research Taguchi method has been used to investigate the effect of cutting speed, feed rate and depth of cut and determine the best level s to minimize the flank wear rate up to total length of 0.3mm based on the ISO standard to maintain the finishing requirements.
Errorless-based techniques can improve route finding in early Alzheimer's disease: a case study.
Provencher, Véronique; Bier, Nathalie; Audet, Thérèse; Gagnon, Lise
2008-01-01
Topographical disorientation is a common and early manifestation of dementia of Alzheimer type, which threatens independence in activities of daily living. Errorless-based techniques appear to be effective in helping patients with amnesia to learn routes, but little is known about their effectiveness in early dementia of Alzheimer type. A 77-year-old woman with dementia of Alzheimer type had difficulty in finding her way around her seniors residence, which reduced her social activities. This study used an ABA design (A is the baseline and B is the intervention) with multiple baselines across routes for going to the rosary (target), laundry, and game rooms (controls). The errorless-based technique intervention was applied to 2 of the 3 routes. Analyses showed significant improvement only for the routes learned with errorless-based techniques. Following the study, the participant increased her topographical knowledge of her surroundings. Route learning interventions based on errorless-based techniques appear to be a promising approach for improving the independence in early dementia of Alzheimer type.
Nano-Al Based Energetics: Rapid Heating Studies and a New Preparation Technique
NASA Astrophysics Data System (ADS)
Sullivan, Kyle; Kuntz, Josh; Gash, Alex; Zachariah, Michael
2011-06-01
Nano-Al based thermites have become an attractive alternative to traditional energetic formulations due to their increased energy density and high reactivity. Understanding the intrinsic reaction mechanism has been a difficult task, largely due to the lack of experimental techniques capable of rapidly and uniform heating a sample (~104- 108 K/s). The current work presents several studies on nano-Al based thermites, using rapid heating techniques. A new mechanism termed a Reactive Sintering Mechanism is proposed for nano-Al based thermites. In addition, new experimental techniques for nanocomposite thermite deposition onto thin Pt electrodes will be discussed. This combined technique will offer more precise control of the deposition, and will serve to further our understanding of the intrinsic reaction mechanism of rapidly heated energetic systems. An improved mechanistic understanding will lead to the development of optimized formulations and architectures. This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Ko, Cheolwoong; Yang, Mikyung; Byun, Taemin; Lee, Sang-Wook
2018-05-01
This study proposed a way to design femur fracture fixation plates made of shape memory alloy based on computed tomography (CT) images of Korean cadaveric femurs. To this end, 3 major design factors of femur fracture fixation plates (circumference angle, thickness, and inner diameter) were selected based on the contact pressure when a femur fracture fixation plate was applied to a cylinder model using the Taguchi method. Then, the effects of the design factors were analyzed. It was shown that the design factors were statistically significant at a level of p = 0.05 concerning the inner diameter and the thickness. The factors affecting the contact pressure were inner diameter, thickness, and circumference angle, in that order. Particularly, in the condition of Case 9 (inner diameter 27 mm, thickness 2.4 mm, and circumference angle 270°), the max. average contact pressure was 21.721 MPa, while the min. average contact pressure was 3.118 MPa in Case 10 (inner diameter 29 mm, thickness 2.0 mm, and circumference angle 210°). When the femur fracture fixation plate was applied to the cylinder model, the displacement due to external sliding and pulling forces was analyzed. As a result, the displacement in the sliding condition was at max. 3.75 times greater than that in the pulling condition, which indicated that the cohesion strength between the femur fracture fixation plate and the cylinder model was likely to be greater in the pulling condition. When a human femur model was applied, the max. average contact pressure was 10.76 MPa, which was lower than the yield strength of a human femur (108 MPa). In addition, the analysis of the rib behaviors of the femur fracture fixation plate in relation to the recovery effect of the shape memory alloy showed that the rib behaviors varied depending on the arbitrarily curved shapes of the femur sections. Copyright © 2018 John Wiley & Sons, Ltd.
Mnif, Inès; Sahnoun, Rihab; Ellouze-Chaabouni, Semia; Ghribi, Dhouha
2014-01-01
Low solubility of certain hydrophobic soil contaminants limits remediation process. Surface-active compounds can improve the solubility and removal of hydrophobic compounds from contaminated soils and, consequently, their biodegradation. Hence, this paper aims to study desorption efficiency of oil from soil of SPB1 lipopeptide biosurfactant. The effect of different physicochemical parameters on desorption potency was assessed. Taguchi experimental design method was applied in order to enhance the desorption capacity and establish the best washing parameters. Mobilization potency was compared to those of chemical surfactants under the newly defined conditions. Better desorption capacity was obtained using 0.1% biosurfacatnt solution and the mobilization potency shows great tolerance to acidic and alkaline pH values and salinity. Results show an optimum value of oil removal from diesel-contaminated soil of about 87%. The optimum washing conditions for surfactant solution volume, biosurfactant concentration, agitation speed, temperature, and time were found to be 12 ml/g of soil, 0.1% biosurfactant, 200 rpm, 30 °C, and 24 h, respectively. The obtained results were compared to those of SDS and Tween 80 at the optimal conditions described above, and the study reveals an effectiveness of SPB1 biosurfactant comparable to the reported chemical emulsifiers. (1) The obtained findings suggest (a) the competence of Bacillus subtilis biosurfactant in promoting diesel desorption from soil towards chemical surfactants and (b) the applicability of this method in decontaminating crude oil-contaminated soil and, therefore, improving bioavailability of hydrophobic compounds. (2) The obtained findings also suggest the adequacy of Taguchi design in promoting process efficiency. Our findings suggest that preoptimized desorption process using microbial-derived emulsifier can contribute significantly to enhancement of hydrophobic pollutants' bioavailability. This study can be
NASA Astrophysics Data System (ADS)
Wisnuadi, Alief Regyan; Damayanti, Retno Wulan; Pujiyanto, Eko
2018-02-01
Bearing is one of the most widely used parts in automotive industry. One of the leading bearing manufacturing companies in the world is SKF Indonesia. This company must produce bearing with international standard. SKF Indonesia must do continuous improvement in order to face competition. During this time, SKF Indonesia is only performing quality control at its Quality Assurance department. In other words, quality improvement at SKF Indonesia has not been done thoroughly. The purpose of this research is to improve quality of outer ring product at SKF Indonesia by conducting an internal grinding process experiment about setting speed ratio, fine position, and spark out grinding time. The specific purpose of this experiment is to optimize some quality responses such as roughness, roundness, and cycle time. All of the response in this experiment were smaller the better. Taguchi method and PCR-TOPSIS are used for the optimization process. The result of this research shows that by using Taguchi method and PCR-TOPSIS, the optimum condition occurs on speed ratio 36, fine position 18 µm/s and spark out 0.5 s. The optimum conditions result were roughness 0.398 µm, roundness 1.78 µm and cycle time 8.1 s. This results have been better than the previous results and meet the standards. The roughness of 0.523 µm decrease to 0.398 µm and the average cycle time of 8.5 s decrease to 8.1 s.
NASA Astrophysics Data System (ADS)
Van Hoten, Hendri; Gunawarman; Mulyadi, Ismet Hari; Kurniawan Mainil, Afdhal; Putra, Bismantoloa dan
2018-02-01
This research is about manufacture nanopowder Bioceramics from local materials used Ball Milling for biomedical applications. Source materials for the manufacture of medicines are plants, animal tissues, microbial structures and engineering biomaterial. The form of raw material medicines is a powder before mixed. In the case of medicines, research is to find sources of biomedical materials that will be in the nanoscale powders can be used as raw material for medicine. One of the biomedical materials that can be used as raw material for medicine is of the type of bioceramics is chicken eggshells. This research will develop methods for manufacture nanopowder material from chicken eggshells with Ball Milling using the Taguchi method and ANOVA. Eggshell milled using a variation of Milling rate on 150, 200 and 250 rpm, the time variation of 1, 2 and 3 hours and variations the grinding balls to eggshell powder weight ratio (BPR) 1: 6, 1: 8, 1: 10. Before milled with Ball Milling crushed eggshells in advance and calcinate to a temperature of 900°C. After the milled material characterization of the fine powder of eggshell using SEM to see its size. The result of this research is optimum parameter of Taguchi Design analysis that is 250 rpm milling rate, 3 hours milling time and BPR is 1: 6 with the average eggshell powder size is 1.305 μm. Milling speed, milling time and ball to powder weight of ratio have contribution successively equal to 60.82%, 30.76% and 6.64% by error equal to 1.78%.
Su, Chao-Ton; Wang, Pa-Chun; Chen, Yan-Cheng; Chen, Li-Fei
2012-08-01
Pressure ulcer is a serious problem during patient care processes. The high risk factors in the development of pressure ulcer remain unclear during long surgery. Moreover, past preventive policies are hard to implement in a busy operation room. The objective of this study is to use data mining techniques to construct the prediction model for pressure ulcers. Four data mining techniques, namely, Mahalanobis Taguchi System (MTS), Support Vector Machines (SVMs), decision tree (DT), and logistic regression (LR), are used to select the important attributes from the data to predict the incidence of pressure ulcers. Measurements of sensitivity, specificity, F(1), and g-means were used to compare the performance of four classifiers on the pressure ulcer data set. The results show that data mining techniques obtain good results in predicting the incidence of pressure ulcer. We can conclude that data mining techniques can help identify the important factors and provide a feasible model to predict pressure ulcer development.
Surface Roughness Optimization Using Taguchi Method of High Speed End Milling For Hardened Steel D2
NASA Astrophysics Data System (ADS)
Hazza Faizi Al-Hazza, Muataz; Ibrahim, Nur Asmawiyah bt; Adesta, Erry T. Y.; Khan, Ahsan Ali; Abdullah Sidek, Atiah Bt.
2017-03-01
The main challenge for any manufacturer is to achieve higher quality of their final products with maintains minimum machining time. In this research final surface roughness analysed and optimized with maximum 0.3 mm flank wear length. The experiment was investigated the effect of cutting speed, feed rate and depth of cut on the final surface roughness using D2 as a work piece hardened to 52-56 HRC, and coated carbide as cutting tool with higher cutting speed 120-240 mm/min. The experiment has been conducted using L9 design of Taguchi collection. The results have been analysed using JMP software.
Nitric acid treated multi-walled carbon nanotubes optimized by Taguchi method
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shamsuddin, Shahidah Arina; Hashim, Uda; Halim, Nur Hamidah Abdul
Electron transfer rate (ETR) of CNTs can be enhanced by increasing the amounts of COOH groups to their wall and opened tips. With the aim to achieve the highest production amount of COOH, Taguchi robust design has been used for the first time to optimize the surface modification of MWCNTs by nitric acid oxidation. Three main oxidation parameters which are concentration of acid, treatment temperature and treatment time have been selected as the control factors that will be optimized. The amounts of COOH produced are measured by using FTIR spectroscopy through the absorbance intensity. From the analysis, we found thatmore » acid concentration and treatment time had the most important influence on the production of COOH. Meanwhile, the treatment temperature will only give intermediate effect. The optimum amount of COOH can be achieved with the treatment by 8.0 M concentration of nitric acid at 120 °C for 2 hour.« less
Performance of Ti-multilayer coated tool during machining of MDN431 alloyed steel
NASA Astrophysics Data System (ADS)
Badiger, Pradeep V.; Desai, Vijay; Ramesh, M. R.
2018-04-01
Turbine forgings and other components are required to be high resistance to corrosion and oxidation because which they are highly alloyed with Ni and Cr. Midhani manufactures one of such material MDN431. It's a hard-to-machine steel with high hardness and strength. PVD coated insert provide an answer to problem with its state of art technique on the WC tool. Machinability studies is carried out on MDN431 steel using uncoated and Ti-multilayer coated WC tool insert using Taguchi optimisation technique. During the present investigation, speed (398-625rpm), feed (0.093-0.175mm/rev), and depth of cut (0.2-0.4mm) varied according to Taguchi L9 orthogonal array, subsequently cutting forces and surface roughness (Ra) were measured. Optimizations of the obtained results are done using Taguchi technique for cutting forces and surface roughness. Using Taguchi technique linear fit model regression analysis carried out for the combination of each input variable. Experimented results are compared and found the developed model is adequate which supported by proof trials. Speed, feed and depth of cut are linearly dependent on the cutting force and surface roughness for uncoated insert whereas Speed and depth of cut feed is inversely dependent in coated insert for both cutting force and surface roughness. Machined surface for coated and uncoated inserts during machining of MDN431 is studied using optical profilometer.
NASA Astrophysics Data System (ADS)
Mondal, Subrata; Bandyopadhyay, Asish.; Pal, Pradip Kumar
2010-10-01
This paper presents the prediction and evaluation of laser clad profile formed by means of CO2 laser applying Taguchi method and the artificial neural network (ANN). Laser cladding is one of the surface modifying technologies in which the desired surface characteristics of any component can be achieved such as good corrosion resistance, wear resistance and hardness etc. Laser is used as a heat source to melt the anti-corrosive powder of Inconel-625 (Super Alloy) to give a coating on 20 MnCr5 substrate. The parametric study of this technique is also attempted here. The data obtained from experiments have been used to develop the linear regression equation and then to develop the neural network model. Moreover, the data obtained from regression equations have also been used as supporting data to train the neural network. The artificial neural network (ANN) is used to establish the relationship between the input/output parameters of the process. The established ANN model is then indirectly integrated with the optimization technique. It has been seen that the developed neural network model shows a good degree of approximation with experimental data. In order to obtain the combination of process parameters such as laser power, scan speed and powder feed rate for which the output parameters become optimum, the experimental data have been used to develop the response surfaces.
Inquiry-Based Stress Reduction Meditation Technique for Teacher Burnout: A Qualitative Study
ERIC Educational Resources Information Center
Schnaider-Levi, Lia; Mitnik, Inbal; Zafrani, Keren; Goldman, Zehavit; Lev-Ari, Shahar
2017-01-01
An inquiry-based intervention has been found to have a positive effect on burnout and mental well-being parameters among teachers. The aim of the current study was to qualitatively evaluate the effect of the inquiry-based stress reduction (IBSR) meditation technique on the participants. Semi-structured interviews were conducted before and after…
Parameter optimization of flux-aided backing-submerged arc welding by using Taguchi method
NASA Astrophysics Data System (ADS)
Pu, Juan; Yu, Shengfu; Li, Yuanyuan
2017-07-01
Flux-aided backing-submerged arc welding has been conducted on D36 steel with thickness of 20 mm. The effects of processing parameters such as welding current, voltage, welding speed and groove angle on welding quality were investigated by Taguchi method. The optimal welding parameters were predicted and the individual importance of each parameter on welding quality was evaluated by examining the signal-to-noise ratio and analysis of variance (ANOVA) results. The importance order of the welding parameters for the welding quality of weld bead was: welding current > welding speed > groove angle > welding voltage. The welding quality of weld bead increased gradually with increasing welding current and welding speed and decreasing groove angle. The optimum values of the welding current, welding speed, groove angle and welding voltage were found to be 1050 A, 27 cm/min, 40∘ and 34 V, respectively.
Sánchez-Sánchez, Xavier; Elias-Zuñiga, Alex; Hernández-Avila, Marcelo
2018-06-01
Ultrasonic injection moulding was confirmed as an efficient processing technique for manufacturing ultra-high molecular weight polyethylene (UHMWPE)/graphite composites. Graphite contents of 1 wt%, 5 wt%, and 7 wt% were mechanically pre-mixed with UHMWPE powder, and each mixture was pressed at 135 °C. A precise quantity of the pre-composites mixtures cut into irregularly shaped small pieces were subjected to ultrasonic injection moulding to fabricate small tensile specimens. The Taguchi method was applied to achieve the optimal level of ultrasonic moulding parameters and to maximize the tensile strength of the composites; the results showed that mould temperature was the most significant parameter, followed by the graphite content and the plunger profile. The observed improvement in tensile strength in the specimen with 1 wt% graphite was of 8.8% and all composites showed an increase in the tensile modulus. Even though the presence of graphite produced a decrease in the crystallinity of all the samples, their thermal stability was considerably higher than that of pure UHMWPE. X-ray diffraction and scanning electron microscopy confirmed the exfoliation and dispersion of the graphite as a function of the ultrasonic processing. Fourier transform infrared spectra showed that the addition of graphite did not influence the molecular structure of the polymer matrix. Further, the ultrasonic energy led oxidative degradation and chain scission in the polymer. Copyright © 2018 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Abu, M. Y.; Norizan, N. S.; Rahman, M. S. Abd
2018-04-01
Remanufacturing is a sustainability strategic planning which transforming the end of life product to as new performance with their warranty is same or better than the original product. In order to quantify the advantages of this strategy, all the processes must implement the optimization to reach the ultimate goal and reduce the waste generated. The aim of this work is to evaluate the criticality of parameters on the end of life crankshaft based on Taguchi’s orthogonal array. Then, estimate the cost using traditional cost accounting by considering the critical parameters. By implementing the optimization, the remanufacturer obviously produced lower cost and waste during production with higher potential to gain the profit. Mahalanobis-Taguchi System was proven as a powerful method of optimization that revealed the criticality of parameters. When subjected the method to the MAN engine model, there was 5 out of 6 crankpins were critical which need for grinding process while no changes happened to the Caterpillar engine model. Meanwhile, the cost per unit for MAN engine model was changed from MYR1401.29 to RM1251.29 while for Caterpillar engine model have no changes due to the no changes on criticality of parameters consideration. Therefore, by integrating the optimization and costing through remanufacturing process, a better decision can be achieved after observing the potential profit will be gained. The significant of output demonstrated through promoting sustainability by reducing re-melting process of damaged parts to ensure consistent benefit of return cores.
Biosorption of malachite green from aqueous solutions by Pleurotus ostreatus using Taguchi method
2014-01-01
Dyes released into the environment have been posing a serious threat to natural ecosystems and aquatic life due to presence of heat, light, chemical and other exposures stable. In this study, the Pleurotus ostreatus (a macro-fungus) was used as a new biosorbent to study the biosorption of hazardous malachite green (MG) from aqueous solutions. The effective disposal of P. ostreatus is a meaningful work for environmental protection and maximum utilization of agricultural residues. The operational parameters such as biosorbent dose, pH, and ionic strength were investigated in a series of batch studies at 25°C. Freundlich isotherm model was described well for the biosorption equilibrium data. The biosorption process followed the pseudo-second-order kinetic model. Taguchi method was used to simplify the experimental number for determining the significance of factors and the optimum levels of experimental factors for MG biosorption. Biosorbent dose and initial MG concentration had significant influences on the percent removal and biosorption capacity. The highest percent removal reached 89.58% and the largest biosorption capacity reached 32.33 mg/g. The Fourier transform infrared spectroscopy (FTIR) showed that the functional groups such as, carboxyl, hydroxyl, amino and phosphonate groups on the biosorbent surface could be the potential adsorption sites for MG biosorption. P. ostreatus can be considered as an alternative biosorbent for the removal of dyes from aqueous solutions. PMID:24620852
Biosorption of malachite green from aqueous solutions by Pleurotus ostreatus using Taguchi method.
Chen, Zhengsuo; Deng, Hongbo; Chen, Can; Yang, Ying; Xu, Heng
2014-03-12
Dyes released into the environment have been posing a serious threat to natural ecosystems and aquatic life due to presence of heat, light, chemical and other exposures stable. In this study, the Pleurotus ostreatus (a macro-fungus) was used as a new biosorbent to study the biosorption of hazardous malachite green (MG) from aqueous solutions. The effective disposal of P. ostreatus is a meaningful work for environmental protection and maximum utilization of agricultural residues.The operational parameters such as biosorbent dose, pH, and ionic strength were investigated in a series of batch studies at 25°C. Freundlich isotherm model was described well for the biosorption equilibrium data. The biosorption process followed the pseudo-second-order kinetic model. Taguchi method was used to simplify the experimental number for determining the significance of factors and the optimum levels of experimental factors for MG biosorption. Biosorbent dose and initial MG concentration had significant influences on the percent removal and biosorption capacity. The highest percent removal reached 89.58% and the largest biosorption capacity reached 32.33 mg/g. The Fourier transform infrared spectroscopy (FTIR) showed that the functional groups such as, carboxyl, hydroxyl, amino and phosphonate groups on the biosorbent surface could be the potential adsorption sites for MG biosorption. P. ostreatus can be considered as an alternative biosorbent for the removal of dyes from aqueous solutions.
A study of trends and techniques for space base electronics
NASA Technical Reports Server (NTRS)
Trotter, J. D.; Wade, T. E.; Gassaway, J. D.; Mahmood, Q.
1978-01-01
A sputtering system was developed to deposit aluminum and aluminum alloys by the dc sputtering technique. This system is designed for a high level of cleanliness and for monitoring the deposition parameters during film preparation. This system is now ready for studying the deposition and annealing parameters upon double-level metal preparation. A technique recently applied for semiconductor analysis, the finite element method, was studied for use in the computer modeling of two dimensional MOS transistor structures. It was concluded that the method has not been sufficiently well developed for confident use at this time. An algorithm was developed for confident use at this time. An algorithm was developed for implementing a computer study which is based upon the finite difference method. The program which was developed was modified and used to calculate redistribution data for boron and phosphorous which had been predeposited by ion implantation with range and straggle conditions. Data were generated for 111 oriented SOS films with redistribution in N2, dry O2 and steam ambients.
NASA Astrophysics Data System (ADS)
Kadam, Nikhil Rajendra; Karthikeyan, Ganesarethinam
2016-10-01
The purpose of the experiments in this paper is to use the Taguchi methods to investigate the wear of WC/C coated nitrided AISI 4140 alloy steel. A study of lamellar WC/C coating which were deposited by a physical vapor deposition on nitrided AISI 4140 alloy steel. The investigation includes wear evaluation using Pin-on-disk configuration. When WC/C coated AISI 4140 alloy steel slides against EN 8 steel, it was found that carbon-rich coatings show much lower wear of the countersurface than nitrogen-rich coatings. The results were correlated with the properties determined from tribological and mechanical characterization, therefore by probably selecting the proper processing parameters the deposition of WC/C coating results in decreasing the wear rate of the substrate which shows a potential for tribological application.
NASA Technical Reports Server (NTRS)
Olds, John Robert; Walberg, Gerald D.
1993-01-01
Multidisciplinary design optimization (MDO) is an emerging discipline within aerospace engineering. Its goal is to bring structure and efficiency to the complex design process associated with advanced aerospace launch vehicles. Aerospace vehicles generally require input from a variety of traditional aerospace disciplines - aerodynamics, structures, performance, etc. As such, traditional optimization methods cannot always be applied. Several multidisciplinary techniques and methods were proposed as potentially applicable to this class of design problem. Among the candidate options are calculus-based (or gradient-based) optimization schemes and parametric schemes based on design of experiments theory. A brief overview of several applicable multidisciplinary design optimization methods is included. Methods from the calculus-based class and the parametric class are reviewed, but the research application reported focuses on methods from the parametric class. A vehicle of current interest was chosen as a test application for this research. The rocket-based combined-cycle (RBCC) single-stage-to-orbit (SSTO) launch vehicle combines elements of rocket and airbreathing propulsion in an attempt to produce an attractive option for launching medium sized payloads into low earth orbit. The RBCC SSTO presents a particularly difficult problem for traditional one-variable-at-a-time optimization methods because of the lack of an adequate experience base and the highly coupled nature of the design variables. MDO, however, with it's structured approach to design, is well suited to this problem. The result of the application of Taguchi methods, central composite designs, and response surface methods to the design optimization of the RBCC SSTO are presented. Attention is given to the aspect of Taguchi methods that attempts to locate a 'robust' design - that is, a design that is least sensitive to uncontrollable influences on the design. Near-optimum minimum dry weight solutions are
NASA Astrophysics Data System (ADS)
Sudhakara, Dara; Prasanthi, Guvvala
2017-04-01
Wire Cut EDM is an unconventional machining process used to build components of complex shape. The current work mainly deals with optimization of surface roughness while machining P/M CW TOOL STEEL by Wire cut EDM using Taguchi method. The process parameters of the Wire Cut EDM is ON, OFF, IP, SV, WT, and WP. L27 OA is used for to design of the experiments for conducting experimentation. In order to find out the effecting parameters on the surface roughness, ANOVA analysis is engaged. The optimum levels for getting minimum surface roughness is ON = 108 µs, OFF = 63 µs, IP = 11 A, SV = 68 V and WT = 8 g.
Genç, Nevim; Doğan, Esra Can; Narcı, Ali Oğuzhan; Bican, Emine
2017-05-01
In this study, a multi-response optimization method using Taguchi's robust design approach is proposed for imidacloprid removal by reverse osmosis. Tests were conducted with different membrane type (BW30, LFC-3, CPA-3), transmembrane pressure (TMP = 20, 25, 30 bar), volume reduction factor (VRF = 2, 3, 4), and pH (3, 7, 11). Quality and quantity of permeate are optimized with the multi-response characteristics of the total dissolved solid (TDS), conductivity, imidacloprid, and total organic carbon (TOC) rejection ratios and flux of permeate. The optimized conditions were determined as membrane type of BW30, TMP 30 bar, VRF 3, and pH 11. Under these conditions, TDS, conductivity, imidacloprid, and TOC rejections and permeate flux were 97.50 97.41, 97.80, 98.00% and 30.60 L/m2·h, respectively. Membrane type was obtained as the most effective factor; its contribution is 64%. The difference between the predicted and observed value of multi-response signal/noise (MRSN) is within the confidence interval.
Simulation verification techniques study
NASA Technical Reports Server (NTRS)
Schoonmaker, P. B.; Wenglinski, T. H.
1975-01-01
Results are summarized of the simulation verification techniques study which consisted of two tasks: to develop techniques for simulator hardware checkout and to develop techniques for simulation performance verification (validation). The hardware verification task involved definition of simulation hardware (hardware units and integrated simulator configurations), survey of current hardware self-test techniques, and definition of hardware and software techniques for checkout of simulator subsystems. The performance verification task included definition of simulation performance parameters (and critical performance parameters), definition of methods for establishing standards of performance (sources of reference data or validation), and definition of methods for validating performance. Both major tasks included definition of verification software and assessment of verification data base impact. An annotated bibliography of all documents generated during this study is provided.
Dutta, Subhasish; Basak, Bikram; Bhunia, Biswanath; Sinha, Ankan; Dey, Apurba
2017-05-01
The present research was conducted to define the approaches for enhanced production of rapamycin (Rap) by Streptomyces hygroscopicus microbial type culture collection (MTCC) 4003. Both physical mutagenesis by ultraviolet ray (UV) and chemical mutagenesis by N-methyl-N-nitro-N-nitrosoguanidine (NTG) have been applied successfully for the improvement of Rap production. Enhancing Rap yield by novel sequential UV mutagenesis technique followed by fermentation gives a significant difference in getting economically scalable amount of this industrially important macrolide compound. Mutant obtained through NTG mutagenesis (NTG-30-27) was found to be superior to others as it initially produced 67% higher Rap than wild type. Statistical optimization of nutritional and physiochemical parameters was carried out to find out most influential factors responsible for enhanced Rap yield by NTG-30-27 which was performed using Taguchi orthogonal array approach. Around 72% enhanced production was achieved with nutritional factors at their assigned level at 23 °C, 120 rpm and pH 7.6. Results were analysed in triplicate basis where validation and purification was carried out using high performance liquid chromatography. Stability study and potency of extracted Rap was supported by turbidimetric assay taking Candida albicans MTCC 227 as test organism.
NASA Astrophysics Data System (ADS)
Sarkar, Bitan Kumar; Kumar, Nikhil; Dey, Rajib; Das, Gopes Chandra
2018-06-01
In the present study, a unique method is adopted to achieve higher reducibility of titaniferous magnetite lump ore (TMO). In this method, TMO is initially heated followed by water quenching. The quenching process generates cracks due to thermal shock in the dense TMO lumps, which, in turn, increases the extent of reduction (EOR) using the lean grade coal as a reductant. The optimum combination of parameters found by using Taguchi's L27 orthogonal array (OA) (five factors, three levels) is - 8 + 4 mm of particle size (PS1), 1423 K of quenching temperature (Qtemp2), 15 minutes of quenching time (Qtime3), 3 times the number of quenching {(No. of Q)3}, and 120 minutes of reduction time (Rtime3) at fixed reduction temperature of 1473 K. At optimized levels of the parameters, 92.39 pct reduction is achieved. Isothermal reduction kinetics of the quenched TMO lumps at the optimized condition reveals mixed controlled mechanisms [initially contracting geometry (CG3) followed by diffusion (D3)]. Activation energies calculated are 69.895 KJ/mole for CG3 and 39.084 KJ/mole for D3.
NASA Astrophysics Data System (ADS)
Norcahyo, Rachmadi; Soepangkat, Bobby O. P.
2017-06-01
A research was conducted for the optimization of the end milling process of ASSAB XW-42 tool steel with multiple performance characteristics based on the orthogonal array with Taguchi-grey relational analysis method. Liquid nitrogen was applied as a coolant. The experimental studies were conducted under varying the liquid nitrogen cooling flow rates (FL), and the end milling process variables, i.e., cutting speed (Vc), feeding speed (Vf), and axial depth of cut (Aa). The optimized multiple performance characteristics were surface roughness (SR), flank wear (VB), and material removal rate (MRR). An orthogonal array, signal-to-noise (S/N) ratio, grey relational analysis, grey relational grade, and analysis of variance were employed to study the multiple performance characteristics. Experimental results showed that flow rate gave the highest contribution for reducing the total variation of the multiple responses, followed by cutting speed, feeding speed, and axial depth of cut. The minimum surface roughness, flank wear, and maximum material removal rate could be obtained by using the values of flow rate, cutting speed, feeding speed, and axial depth of cut of 0.5 l/minute, 109.9 m/minute, 440 mm/minute, and 0.9 mm, respectively.
NASA Astrophysics Data System (ADS)
Mehat, N. M.; Kamaruddin, S.
2017-10-01
An increase in demand for industrial gears has instigated the escalating uses of plastic-matrix composites, particularly carbon or glass fibre reinforced plastics as gear material to enhance the properties and limitation in plastic gears. However, the production of large quantity of these synthetic fibres reinforced composites has posed serious threat to ecosystem. Therefore, this work is conducted to study the applicability and practical ability of using bamboo fillers particularly in plastic gear manufacturing as opposed to synthetic fibres via the Taguchi optimization method. The results showed that no failure mechanism such as gear tooth root cracking and severe tooth wear were observed in gear tested made of 5-30 wt% of bamboo fillers in comparing with the unfilled PP gear. These results indicated that bamboo can be practically and economically used as an alternative filler in plastic material reinforcement as well as in minimizing the cost of raw material in general.
Interactive design optimization of magnetorheological-brake actuators using the Taguchi method
NASA Astrophysics Data System (ADS)
Erol, Ozan; Gurocak, Hakan
2011-10-01
This research explored an optimization method that would automate the process of designing a magnetorheological (MR)-brake but still keep the designer in the loop. MR-brakes apply resistive torque by increasing the viscosity of an MR fluid inside the brake. This electronically controllable brake can provide a very large torque-to-volume ratio, which is very desirable for an actuator. However, the design process is quite complex and time consuming due to many parameters. In this paper, we adapted the popular Taguchi method, widely used in manufacturing, to the problem of designing a complex MR-brake. Unlike other existing methods, this approach can automatically identify the dominant parameters of the design, which reduces the search space and the time it takes to find the best possible design. While automating the search for a solution, it also lets the designer see the dominant parameters and make choices to investigate only their interactions with the design output. The new method was applied for re-designing MR-brakes. It reduced the design time from a week or two down to a few minutes. Also, usability experiments indicated significantly better brake designs by novice users.
Metamodels for Computer-Based Engineering Design: Survey and Recommendations
NASA Technical Reports Server (NTRS)
Simpson, Timothy W.; Peplinski, Jesse; Koch, Patrick N.; Allen, Janet K.
1997-01-01
The use of statistical techniques to build approximations of expensive computer analysis codes pervades much of todays engineering design. These statistical approximations, or metamodels, are used to replace the actual expensive computer analyses, facilitating multidisciplinary, multiobjective optimization and concept exploration. In this paper we review several of these techniques including design of experiments, response surface methodology, Taguchi methods, neural networks, inductive learning, and kriging. We survey their existing application in engineering design and then address the dangers of applying traditional statistical techniques to approximate deterministic computer analysis codes. We conclude with recommendations for the appropriate use of statistical approximation techniques in given situations and how common pitfalls can be avoided.
Studies on the Parametric Effects of Plasma Arc Welding of 2205 Duplex Stainless Steel
NASA Astrophysics Data System (ADS)
Selva Bharathi, R.; Siva Shanmugam, N.; Murali Kannan, R.; Arungalai Vendan, S.
2018-03-01
This research study attempts to create an optimized parametric window by employing Taguchi algorithm for Plasma Arc Welding (PAW) of 2 mm thick 2205 duplex stainless steel. The parameters considered for experimentation and optimization are the welding current, welding speed and pilot arc length respectively. The experimentation involves the parameters variation and subsequently recording the depth of penetration and bead width. Welding current of 60-70 A, welding speed of 250-300 mm/min and pilot arc length of 1-2 mm are the range between which the parameters are varied. Design of experiments is used for the experimental trials. Back propagation neural network, Genetic algorithm and Taguchi techniques are used for predicting the bead width, depth of penetration and validated with experimentally achieved results which were in good agreement. Additionally, micro-structural characterizations are carried out to examine the weld quality. The extrapolation of these optimized parametric values yield enhanced weld strength with cost and time reduction.
Using the critical incident technique in community-based participatory research: a case study.
Belkora, Jeffrey; Stupar, Lauren; O'Donnell, Sara
2011-01-01
Successful community-based participatory research involves the community partner in every step of the research process. The primary study for this paper took place in rural, Northern California. Collaborative partners included an academic researcher and two community based resource centers that provide supportive services to people diagnosed with cancer. This paper describes our use of the Critical Incident Technique (CIT) to conduct Community-based Participatory Research. We ask: Did the CIT facilitate or impede the active engagement of the community in all steps of the study process? We identified factors about the Critical Incident Technique that were either barriers or facilitators to involving the community partner in every step of the research process. Facilitators included the CIT's ability to accommodate involvement from a large spectrum of the community, its flexible design, and its personal approach. Barriers to community engagement included training required to conduct interviews, depth of interview probes, and time required. Overall, our academic-community partners felt that our use of the CIT facilitated community involvement in our Community-Based Participatory Research Project, where we used it to formally document the forces promoting and inhibiting successful achievement of community aims.
Techniques for Cyber Attack Attribution
2003-10-01
Asaka, Midori, Shunji Okazawa, Atsushi Taguchi, and Shigeki Goto. June 1999. “A Method of Tracing Intruders by Use of Mobile Agents”, INET’99. http...Tsuchiya, Takefumi Onabuta, Shunji Okazawa, and Shigeki Goto. November 1999. “Local Attack Detection and Intrusion Route Tracing”, IEICE Transaction on
SSSFD manipulator engineering using statistical experiment design techniques
NASA Technical Reports Server (NTRS)
Barnes, John
1991-01-01
The Satellite Servicer System Flight Demonstration (SSSFD) program is a series of Shuttle flights designed to verify major on-orbit satellite servicing capabilities, such as rendezvous and docking of free flyers, Orbital Replacement Unit (ORU) exchange, and fluid transfer. A major part of this system is the manipulator system that will perform the ORU exchange. The manipulator must possess adequate toolplate dexterity to maneuver a variety of EVA-type tools into position to interface with ORU fasteners, connectors, latches, and handles on the satellite, and to move workpieces and ORUs through 6 degree of freedom (dof) space from the Target Vehicle (TV) to the Support Module (SM) and back. Two cost efficient tools were combined to perform a study of robot manipulator design parameters. These tools are graphical computer simulations and Taguchi Design of Experiment methods. Using a graphics platform, an off-the-shelf robot simulation software package, and an experiment designed with Taguchi's approach, the sensitivities of various manipulator kinematic design parameters to performance characteristics are determined with minimal cost.
Non-Conventional Techniques for the Study of Phase Transitions in NiTi-Based Alloys
NASA Astrophysics Data System (ADS)
Nespoli, Adelaide; Villa, Elena; Passaretti, Francesca; Albertini, Franca; Cabassi, Riccardo; Pasquale, Massimo; Sasso, Carlo Paolo; Coïsson, Marco
2014-07-01
Differential scanning calorimetry and electrical resistance measurements are the two most common techniques for the study of the phase transition path and temperatures of shape memory alloys (SMA) in stress-free condition. Besides, it is well known that internal friction measurements are also useful for this purpose. There are indeed some further techniques which are seldom used for the basic characterization of SMA transition: dilatometric analysis, magnetic measurements, and Seebeck coefficient study. In this work, we discuss the attitude of these techniques for the study of NiTi-based phase transition. Measurements were conducted on several fully annealed Ni50- x Ti50Cu x samples ranging from 3 to 10 at.% in Cu content, fully annealed at 850 °C for 1 h in vacuum and quenched in water at room temperature. Results show that all these techniques are sensitive to phase transition, and they provide significant information about the existence of intermediate phases.
Aerodynamic measurement techniques. [laser based diagnostic techniques
NASA Technical Reports Server (NTRS)
Hunter, W. W., Jr.
1976-01-01
Laser characteristics of intensity, monochromatic, spatial coherence, and temporal coherence were developed to advance laser based diagnostic techniques for aerodynamic related research. Two broad categories of visualization and optical measurements were considered, and three techniques received significant attention. These are holography, laser velocimetry, and Raman scattering. Examples of the quantitative laser velocimeter and Raman scattering measurements of velocity, temperature, and density indicated the potential of these nonintrusive techniques.
NASA Astrophysics Data System (ADS)
Meseguer Valdenebro, Jose Luis
Electric arc welding processes represent one of the most used techniques on manufacturing processes of mechanical components in modern industry. The electric arc welding processes have been adapted to current needs, becoming a flexible and versatile way to manufacture. Numerical results in the welding process are validated experimentally. The main numerical methods most commonly used today are three: finite difference method, finite element method and finite volume method. The most widely used numerical method for the modeling of welded joints is the finite element method because it is well adapted to the geometric and boundary conditions in addition to the fact that there is a variety of commercial programs which use the finite element method as a calculation basis. The content of this thesis shows an experimental study of a welded joint conducted by means of the MIG welding process of aluminum alloy 6063-T5. The numerical process is validated experimentally by applying the method of finite element through the calculation program ANSYS. The experimental results in this paper are the cooling curves, the critical cooling time t4/3, the weld bead geometry, the microhardness obtained in the welded joint, and the metal heat affected zone base, process dilution, critical areas intersected between the cooling curves and the curve TTP. The numerical results obtained in this thesis are: the thermal cycle curves, which represent both the heating to maximum temperature and subsequent cooling. The critical cooling time t4/3 and thermal efficiency of the process are calculated and the bead geometry obtained experimentally is represented. The heat affected zone is obtained by differentiating the zones that are found at different temperatures, the critical areas intersected between the cooling curves and the TTP curve. In order to conclude this doctoral thesis, an optimization has been conducted by means of the Taguchi method for welding parameters in order to obtain an
NASA Astrophysics Data System (ADS)
Tutar, Mumin; Aydin, Hakan; Bayram, Ali
2017-08-01
Formability and energy absorption capability of a steel sheet are highly desirable properties in manufacturing components for automotive applications. TWinning Induced Plastisity (TWIP) steels are, new generation high Mn alloyed steels, attractive for the automotive industry due to its outstanding elongation (%40-45) and tensile strength (~1000MPa). So, TWIP steels provide excellent formability and energy absorption capability. Another required property from the steel sheets is suitability for manufacturing methods such as welding. The use of the steel sheets in the automotive applications inevitably involves welding. Considering that there are 3000-5000 welded spots on a vehicle, it can be interpreted that one of the most important manufacturing method is Resistance Spot Welding (RSW) for the automotive industry. In this study; firstly, TWIP steel sheet were cold rolled to 15% reduction in thickness. Then, the cold rolled TWIP steel sheets were welded with RSW method. The welding parameters (welding current, welding time and electrode force) were optimized for maximizing the peak tensile shear load and minimizing the indentation of the joints using a Taguchi L9 orthogonal array. The effect of welding parameters was also evaluated by examining the signal-to-noise ratio and analysis of variance (ANOVA) results.
Retention of denture bases fabricated by three different processing techniques – An in vivo study
Chalapathi Kumar, V. H.; Surapaneni, Hemchand; Ravikiran, V.; Chandra, B. Sarat; Balusu, Srilatha; Reddy, V. Naveen
2016-01-01
Aim: Distortion due to Polymerization shrinkage compromises the retention. To evaluate the amount of retention of denture bases fabricated by conventional, anchorized, and injection molding polymerization techniques. Materials and Methods: Ten completely edentulous patients were selected, impressions were made, and master cast obtained was duplicated to fabricate denture bases by three polymerization techniques. Loop was attached to the finished denture bases to estimate the force required to dislodge them by retention apparatus. Readings were subjected to nonparametric Friedman two-way analysis of variance followed by Bonferroni correction methods and Wilcoxon matched-pairs signed-ranks test. Results: Denture bases fabricated by injection molding (3740 g), anchorized techniques (2913 g) recorded greater retention values than conventional technique (2468 g). Significant difference was seen between these techniques. Conclusions: Denture bases obtained by injection molding polymerization technique exhibited maximum retention, followed by anchorized technique, and least retention was seen in conventional molding technique. PMID:27382542
Design and operation of a bio-inspired micropump based on blood-sucking mechanism of mosquitoes
NASA Astrophysics Data System (ADS)
Leu, Tzong-Shyng; Kao, Ruei-Hung
2018-05-01
The study is to develop a novel bionic micropump, mimicking blood-suck mechanism of mosquitos with a similar efficiency of 36%. The micropump is produced by using micro-electro-mechanical system (MEMS) technology, PDMS (polydimethylsiloxane) to fabricate the microchannel, and an actuator membrane made by Fe-PDMS. It employs an Nd-FeB permanent magnet and PZT to actuate the Fe-PDMS membrane for generating flow rate. A lumped model theory and the Taguchi method are used for numerical simulation of pulsating flow in the micropump. Also focused is to change the size of mosquito mouth for identifying the best waveform for the transient flow processes. Based on computational results of channel size and the Taguchi method, an optimization actuation waveform is identified. The maximum pumping flow rate is 23.5 μL/min and the efficiency is 86%. The power density of micropump is about 8 times of that produced by mosquito’s suction. In addition to using theoretical design of the channel size, also combine with Taguchi method and asymmetric actuation to find the optimization actuation waveform, the experimental result shows the maximum pumping flowrate is 23.5 μL/min and efficiency is 86%, moreover, the power density of micropump is 8 times higher than mosquito’s.
Pant, Apourv; Rai, J P N
2018-04-15
Two phase bioreactor was constructed, designed and developed to evaluate the chlorpyrifos remediation. Six biotic and abiotic factors (substrate-loading rate, slurry phase pH, slurry phase dissolved oxygen (DO), soil water ratio, temperature and soil micro flora load) were evaluated by design of experimental (DOE) methodology employing Taguchi's orthogonal array (OA). The selected six factors were considered at two levels L-8 array (2^7, 15 experiments) in the experimental design. The optimum operating conditions obtained from the methodology showed enhanced chlorpyrifos degradation from 283.86µg/g to 955.364µg/g by overall 70.34% of enhancement. In the present study, with the help of few well defined experimental parameters a mathematical model was constructed to understand the complex bioremediation process and optimize the approximate parameters upto great accuracy. Copyright © 2017 Elsevier Inc. All rights reserved.
Kim, Dongcheol; Rhee, Sehun
2002-01-01
CO(2) welding is a complex process. Weld quality is dependent on arc stability and minimizing the effects of disturbances or changes in the operating condition commonly occurring during the welding process. In order to minimize these effects, a controller can be used. In this study, a fuzzy controller was used in order to stabilize the arc during CO(2) welding. The input variable of the controller was the Mita index. This index estimates quantitatively the arc stability that is influenced by many welding process parameters. Because the welding process is complex, a mathematical model of the Mita index was difficult to derive. Therefore, the parameter settings of the fuzzy controller were determined by performing actual control experiments without using a mathematical model of the controlled process. The solution, the Taguchi method was used to determine the optimal control parameter settings of the fuzzy controller to make the control performance robust and insensitive to the changes in the operating conditions.
NASA Astrophysics Data System (ADS)
Lin, Lianghua; Liu, Zhiyi; Ying, Puyou; Liu, Meng
2015-12-01
Multi-step heat treatment effectively enhances the stress corrosion cracking (SCC) resistance but usually degrades the mechanical properties of Al-Zn-Mg-Cu alloys. With the aim to enhance SCC resistance as well as strength of Al-Zn-Mg-Cu alloys, we have optimized the process parameters during two-step aging of Al-6.1Zn-2.8Mg-1.9Cu alloy by Taguchi's L9 orthogonal array. In this work, analysis of variance (ANOVA) was performed to find out the significant heat treatment parameters. The slow strain rate testing combined with scanning electron microscope and transmission electron microscope was employed to study the SCC behaviors of Al-Zn-Mg-Cu alloy. Results showed that the contour map produced by ANOVA offered a reliable reference for selection of optimum heat treatment parameters. By using this method, a desired combination of mechanical performances and SCC resistance was obtained.
NASA Astrophysics Data System (ADS)
Grewal, Harpreet Singh; Bhandari, Sanjeev; Singh, Harpreet
2012-09-01
WC-Co-Cr coatings were deposited on some hydroturbine 13Cr4Ni and 16Cr5Ni steels by the detonation-gun spray process. An in-depth characterization of the as-sprayed coating was done using X-ray diffraction (XRD) and scanning electron microscopy (SEM)/energy-dispersive X-ray spectroscopy (EDS) techniques. Microhardness and porosity measurements were also made. The coating was found to have a typical splat-like morphology with some indications of unmelted carbide particles. The XRD results showed the presence of WC as the primary phase along with W2C and Co6W6C as secondary phases. Furthermore, the slurry erosion behavior of the coatings was investigated to ascertain the usefulness of the coatings to reduce the slurry erosion of the steels. The effect of four operating factors viz. the velocity, impact angle, concentration, and particle size on the slurry erosion of coated and bare steels has been studied using a high-speed jet-type test rig. The sand used as an erodent was collected from a power plant to replicate the actual turbine conditions. It has been observed that the given cermet coating can enhance the erosion resistance of the steel. Velocity was found to be the most significant factor affecting the erosion behavior of the coating, whereas it was the erodent particle size in the case of uncoated steel. As evidenced from the SEM images, the platelet mechanism of erosion seemed to be the prominent one, causing the removal of material from the surface of the steel, whereas for the coating, the formation and interlinking of cracks resulted in the removal of material.
Contests versus Norms: Implications of Contest-Based and Norm-Based Intervention Techniques
Bergquist, Magnus; Nilsson, Andreas; Hansla, André
2017-01-01
Interventions using either contests or norms can promote environmental behavioral change. Yet research on the implications of contest-based and norm-based interventions is lacking. Based on Goal-framing theory, we suggest that a contest-based intervention frames a gain goal promoting intensive but instrumental behavioral engagement. In contrast, the norm-based intervention was expected to frame a normative goal activating normative obligations for targeted and non-targeted behavior and motivation to engage in pro-environmental behaviors in the future. In two studies participants (n = 347) were randomly assigned to either a contest- or a norm-based intervention technique. Participants in the contest showed more intensive engagement in both studies. Participants in the norm-based intervention tended to report higher intentions for future energy conservation (Study 1) and higher personal norms for non-targeted pro-environmental behaviors (Study 2). These findings suggest that contest-based intervention technique frames a gain goal, while norm-based intervention frames a normative goal. PMID:29218026
Contests versus Norms: Implications of Contest-Based and Norm-Based Intervention Techniques.
Bergquist, Magnus; Nilsson, Andreas; Hansla, André
2017-01-01
Interventions using either contests or norms can promote environmental behavioral change. Yet research on the implications of contest-based and norm-based interventions is lacking. Based on Goal-framing theory, we suggest that a contest-based intervention frames a gain goal promoting intensive but instrumental behavioral engagement. In contrast, the norm-based intervention was expected to frame a normative goal activating normative obligations for targeted and non-targeted behavior and motivation to engage in pro-environmental behaviors in the future. In two studies participants ( n = 347) were randomly assigned to either a contest- or a norm-based intervention technique. Participants in the contest showed more intensive engagement in both studies. Participants in the norm-based intervention tended to report higher intentions for future energy conservation (Study 1) and higher personal norms for non-targeted pro-environmental behaviors (Study 2). These findings suggest that contest-based intervention technique frames a gain goal, while norm-based intervention frames a normative goal.
Erosion performance studies on sansevieria cylindrica reinforced vinylester composite
NASA Astrophysics Data System (ADS)
Johnson, R. Deepak Joel; Arumugaprabu, V.; Uthayakumar, M.; Vigneshwaran, S.; Manikandan, V.; Bennet, C.
2018-03-01
The intent of the research is to study the erosion behaviour of NaOH treated and untreated sansevieria cylindrica reinforced vinyl ester composites (SCVEC). The SCVEC was fabricated by varying fiber length as 30 mm and 40 mm and the fiber concentration as 30 wt%, 40 wt% and 50 wt% respectively for both NaOH treated and untreated sansevieria cylindrica fibres. The fabricated SCVEC was subjected to erosion studies using abrasive air jet erosion test rig. Full factorial design of experiment for conducting the erosion studies was made using Taguchi technique. The erosion test process variables like impingement angle 30°, 60° and 90°, impact velocity 28, 41 and 72 m s‑1, erodent feed rate or discharge 2.5, 3.3 and 4 g min‑1 and exposure time 5, 10 and 15 min were used to study the erosion rate of the SCVEC specimen. From the Taguchi analysis the optimized erosion process parameter and fabrication process parameters were found to be as fiber length 30 mm, NaOH treated fiber, fiber content 40 wt.%, impingement angle 90°, impact velocity 41 m s‑1, erodent discharge 4 g min‑1 and exposure time 15 min. Further, the erosion mechanism on the surface of the eroded SCVEC specimen was studied using Scanning electron microscope (SEM).
Rai, Suchita; Wasewar, Kailas L; Lataye, Dilip H; Mishra, Rajshekhar S; Puttewar, Suresh P; Chaddha, Mukesh J; Mahindiran, P; Mukhopadhyay, Jyoti
2012-09-01
'Red mud' or 'bauxite residue', a waste generated from alumina refinery is highly alkaline in nature with a pH of 10.5-12.5. Red mud poses serious environmental problems such as alkali seepage in ground water and alkaline dust generation. One of the options to make red mud less hazardous and environmentally benign is its neutralization with acid or an acidic waste. Hence, in the present study, neutralization of alkaline red mud was carried out using a highly acidic waste (pickling waste liquor). Pickling waste liquor is a mixture of strong acids used for descaling or cleaning the surfaces in steel making industry. The aim of the study was to look into the feasibility of neutralization process of the two wastes using Taguchi's design of experimental methodology. This would make both the wastes less hazardous and safe for disposal. The effect of slurry solids, volume of pickling liquor, stirring time and temperature on the neutralization process were investigated. The analysis of variance (ANOVA) shows that the volume of the pickling liquor is the most significant parameter followed by quantity of red mud with 69.18% and 18.48% contribution each respectively. Under the optimized parameters, pH value of 7 can be achieved by mixing the two wastes. About 25-30% of the total soda from the red mud is being neutralized and alkalinity is getting reduced by 80-85%. Mineralogy and morphology of the neutralized red mud have also been studied. The data presented will be useful in view of environmental concern of red mud disposal.
Periasamy, Rathinasamy; Palvannan, Thayumanavan
2010-12-01
Production of laccase using a submerged culture of Pleurotus orstreatus IMI 395545 was optimized by the Taguchi orthogonal array (OA) design of experiments (DOE) methodology. This approach facilitates the study of the interactions of a large number of variables spanned by factors and their settings, with a small number of experiments, leading to considerable savings in time and cost for process optimization. This methodology optimizes the number of impact factors and enables to calculate their interaction in the production of industrial enzymes. Eight factors, viz. glucose, yeast extract, malt extract, inoculum, mineral solution, inducer (1 mM CuSO₄) and amino acid (l-asparagine) at three levels and pH at two levels, with an OA layout of L18 (2¹ × 3⁷) were selected for the proposed experimental design. The laccase yield obtained from the 18 sets of fermentation experiments performed with the selected factors and levels was further processed with Qualitek-4 software. The optimized conditions shared an enhanced laccase expression of 86.8% (from 485.0 to 906.3 U). The combination of factors was further validated for laccase production and reactive blue 221 decolorization. The results revealed an enhanced laccase yield of 32.6% and dye decolorization up to 84.6%. This methodology allows the complete evaluation of main and interaction factors. © 2010 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim
Aghili, Zahra Sadat; Zarkesh-Esfahani, Sayyed Hamid
2018-02-01
Growth hormone deficiency results in growth retardation in children and the GH deficiency syndrome in adults and they need to receive recombinant-GH in order to rectify the GH deficiency symptoms. Mammalian cells have become the favorite system for production of recombinant proteins for clinical application compared to prokaryotic systems because of their capability for appropriate protein folding, assembly, post-translational modification and proper signal. However, production level in mammalian cells is generally low compared to prokaryotic hosts. Taguchi has established orthogonal arrays to describe a large number of experimental situations mainly to reduce experimental errors and to enhance the efficiency and reproducibility of laboratory experiments.In the present study, rhGH was produced in CHO cells and production of rhGH was assessed using Dot blotting, western blotting and Elisa assay. For optimization of rhGH production in CHO cells using Taguchi method An M16 orthogonal experimental design was used to investigate four different culture components. The biological activity of rhGH was assessed using LHRE-TK-Luciferase reporter gene system in HEK-293 and compared to the biological activity of prokaryotic rhGH.A maximal productivity of rhGH was reached in the conditions of 1%DMSO, 1%glycerol, 25 µM ZnSO 4 and 0 mM NaBu. Our findings indicate that control of culture conditions such as the addition of chemical components helps to develop an efficient large-scale and industrial process for the production of rhGH in CHO cells. Results of bioassay indicated that rhGH produced by CHO cells is able to induce GH-mediated intracellular cell signaling and showed higher bioactivity when compared to prokaryotic GH at the same concentrations. © Georg Thieme Verlag KG Stuttgart · New York.
Wu, Peiwen; Yu, Yang; McGhee, Claire E.; Tan, Li Huey
2014-01-01
In this review, we summarize recent progresses in the application of synchrotron-based spectroscopic techniques for nucleic acid research that takes advantage of high-flux and high-brilliance electromagnetic radiation from synchrotron sources. The first section of the review focuses on the characterization of the structure and folding processes of nucleic acids using different types of synchrotron-based spectroscopies, such as X-ray absorption spectroscopy, X-ray emission spectroscopy, X-ray photoelectron spectroscopy, synchrotron radiation circular dichroism, X-ray footprinting and small-angle X-ray scattering. In the second section, the characterization of nucleic acid-based nanostructures, nucleic acid-functionalized nanomaterials and nucleic acid-lipid interactions using these spectroscopic techniques is summarized. Insights gained from these studies are described and future directions of this field are also discussed. PMID:25205057
Wu, Peiwen; Yu, Yang; McGhee, Claire E.; ...
2014-09-10
In this paper, we summarize recent progress in the application of synchrotron-based spectroscopic techniques for nucleic acid research that takes advantage of high-flux and high-brilliance electromagnetic radiation from synchrotron sources. The first section of the review focuses on the characterization of the structure and folding processes of nucleic acids using different types of synchrotron-based spectroscopies, such as X-ray absorption spectroscopy, X-ray emission spectroscopy, X-ray photoelectron spectroscopy, synchrotron radiation circular dichroism, X-ray footprinting and small-angle X-ray scattering. In the second section, the characterization of nucleic acid-based nanostructures, nucleic acid-functionalized nanomaterials and nucleic acid-lipid interactions using these spectroscopic techniques is summarized. Insightsmore » gained from these studies are described and future directions of this field are also discussed.« less
Wang, Dansheng; Wang, Qinghua; Wang, Hao; Zhu, Hongping
2016-01-01
In the electromechanical impedance (EMI) method, the PZT patch performs the functions of both sensor and exciter. Due to the high frequency actuation and non-model based characteristics, the EMI method can be utilized to detect incipient structural damage. In recent years EMI techniques have been widely applied to monitor the health status of concrete and steel materials, however, studies on application to timber are limited. This paper will explore the feasibility of using the EMI technique for damage detection in timber specimens. In addition, the conventional damage index, namely root mean square deviation (RMSD) is employed to evaluate the level of damage. On that basis, a new damage index, Mahalanobis distance based on RMSD, is proposed to evaluate the damage severity of timber specimens. Experimental studies are implemented to detect notch and hole damage in the timber specimens. Experimental results verify the availability and robustness of the proposed damage index and its superiority over the RMSD indexes. PMID:27782088
Wang, Dansheng; Wang, Qinghua; Wang, Hao; Zhu, Hongping
2016-10-22
In the electromechanical impedance (EMI) method, the PZT patch performs the functions of both sensor and exciter. Due to the high frequency actuation and non-model based characteristics, the EMI method can be utilized to detect incipient structural damage. In recent years EMI techniques have been widely applied to monitor the health status of concrete and steel materials, however, studies on application to timber are limited. This paper will explore the feasibility of using the EMI technique for damage detection in timber specimens. In addition, the conventional damage index, namely root mean square deviation (RMSD) is employed to evaluate the level of damage. On that basis, a new damage index, Mahalanobis distance based on RMSD, is proposed to evaluate the damage severity of timber specimens. Experimental studies are implemented to detect notch and hole damage in the timber specimens. Experimental results verify the availability and robustness of the proposed damage index and its superiority over the RMSD indexes.
Creating A Data Base For Design Of An Impeller
NASA Technical Reports Server (NTRS)
Prueger, George H.; Chen, Wei-Chung
1993-01-01
Report describes use of Taguchi method of parametric design to create data base facilitating optimization of design of impeller in centrifugal pump. Data base enables systematic design analysis covering all significant design parameters. Reduces time and cost of parametric optimization of design: for particular impeller considered, one can cover 4,374 designs by computational simulations of performance for only 18 cases.
Investigation and Taguchi Optimization of Microbial Fuel Cell Salt Bridge Dimensional Parameters
NASA Astrophysics Data System (ADS)
Sarma, Dhrupad; Barua, Parimal Bakul; Dey, Nabendu; Nath, Sumitro; Thakuria, Mrinmay; Mallick, Synthia
2018-01-01
One major problem of two chamber salt bridge microbial fuel cells (MFCs) is the high resistance offered by the salt bridge to anion flow. Many researchers who have studied and optimized various parameters related to salt bridge MFC, have not shed much light on the effect of salt bridge dimensional parameters on the MFC performance. Therefore, the main objective of this research is to investigate the effect of length and cross sectional area of salt bridge and the effect of solar radiation and atmospheric temperature on MFC current output. An experiment has been designed using Taguchi L9 orthogonal array, taking length and cross sectional area of salt bridge as factors having three levels. Nine MFCs were fabricated as per the nine trial conditions. Trials were conducted for 3 days and output current of each of the MFCs along with solar insolation and atmospheric temperature were recorded. Analysis of variance shows that salt bridge length has significant effect both on mean (with 53.90% contribution at 95% CL) and variance (with 56.46% contribution at 87% CL), whereas the effect of cross sectional area of the salt bridge and the interaction of these two factors is significant on mean only (with 95% CL). Optimum combination was found at 260 mm salt bridge length and 506.7 mm2 cross sectional area with 4.75 mA of mean output current. The temperature and solar insolation data when correlated with each of the MFCs average output current, revealed that both external factors have significant impact on MFC current output but the correlation coefficient varies from MFC to MFC depending on salt bridge dimensional parameters.
Singh, Gurmeet; Jain, Vivek; Gupta, Dheeraj; Ghai, Aman
2016-09-01
Orthopaedic surgery involves drilling of bones to get them fixed at their original position. The drilling process used in orthopaedic surgery is most likely to the mechanical drilling process and there is all likelihood that it may harm the already damaged bone, the surrounding bone tissue and nerves, and the peril is not limited at that. It is very much feared that the recovery of that part may be impeded so that it may not be able to sustain life long. To achieve sustainable orthopaedic surgery, a surgeon must try to control the drilling damage at the time of bone drilling. The area around the holes decides the life of bone joint and so, the contiguous area of drilled hole must be intact and retain its properties even after drilling. This study mainly focuses on optimization of drilling parameters like rotational speed, feed rate and the type of tool at three levels each used by Taguchi optimization for surface roughness and material removal rate. The confirmation experiments were also carried out and results found with the confidence interval. Scanning electrode microscopy (SEM) images assisted in getting the micro level information of bone damage. Copyright © 2016 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Giahi, M.; Farajpour, G.; Taghavi, H.; Shokri, S.
2014-07-01
In this study, ZnO nanoparticles were prepared by a sol-gel method for the first time. Taguchi method was used to identify the several factors that may affect degradation percentage of betamethasone sodium phosphate in wastewater in UV/K2S2O8/nano-ZnO system. Our experimental design consisted of testing five factors, i.e., dosage of K2S2O8, concentration of betamethasone sodium phosphate, amount of ZnO, irradiation time and initial pH. With four levels of each factor tested. It was found that, optimum parameters are irradiation time, 180 min; pH 9.0; betamethasone sodium phosphate, 30 mg/L; amount of ZnO, 13 mg; K2S2O8, 1 mM. The percentage contribution of each factor was determined by the analysis of variance (ANOVA). The results showed that irradiation time; pH; amount of ZnO; drug concentration and dosage of K2S2O8 contributed by 46.73, 28.56, 11.56, 6.70, and 6.44%, respectively. Finally, the kinetics process was studied and the photodegradation rate of betamethasone sodium phosphate was found to obey pseudo-first-order kinetics equation represented by the Langmuir-Hinshelwood model.
NASA Astrophysics Data System (ADS)
Muttalib, M. Firdaus A.; Chen, Ruiqi Y.; Pearce, S. J.; Charlton, Martin D. B.
2017-11-01
In this paper, we demonstrate the optimization of reactive-ion etching (RIE) parameters for the fabrication of tantalum pentoxide (Ta2O5) waveguide with chromium (Cr) hard mask in a commercial OIPT Plasmalab 80 RIE etcher. A design of experiment (DOE) using Taguchi method was implemented to find optimum RF power, mixture of CHF3 and Ar gas ratio, and chamber pressure for a high etch rate, good selectivity, and smooth waveguide sidewall. It was found that the optimized etch condition obtained in this work were RF power = 200 W, gas ratio = 80 %, and chamber pressure = 30 mTorr with an etch rate of 21.6 nm/min, Ta2O5/Cr selectivity ratio of 28, and smooth waveguide sidewall.
Effect of processing parameters on FDM process
NASA Astrophysics Data System (ADS)
Chari, V. Srinivasa; Venkatesh, P. R.; Krupashankar, Dinesh, Veena
2018-04-01
This paper focused on the process parameters on fused deposition modeling (FDM). Infill, resolution, temperature are the process variables considered for experimental studies. Compression strength, Hardness test microstructure are the outcome parameters, this experimental study done based on the taguchi's L9 orthogonal array is used. Taguchi array used to build the 9 different models and also to get the effective output results on the under taken parameters. The material used for this experimental study is Polylactic Acid (PLA).
NASA Astrophysics Data System (ADS)
Ermoline, Alexandre
The general objective of this work is to develop an experimental technique for studying the high-temperature phase compositions and phase equilibria in molten metal-based binary and ternary systems, such as Zr-O-N, B-N-O, Al-O, and others. A specific material system of Zr-O-N was selected for studying and testing this technique. The information about the high-temperature phase equilibria in reactive metal-based systems is scarce and their studying is difficult because of chemical reactions occurring between samples and essentially any container materials, and causing contamination of the system. Containerless microgravity experiments for studying equilibria in molten metal-gas systems were designed to be conducted onboard of a NASA KC-135 aircraft flying parabolic trajectories. A uniaxial apparatus suitable for acoustic levitation, laser heating, and splat quenching of small samples was developed and equipped with computer-based controller and optical diagnostics. Normal-gravity tests were conducted to determine the most suitable operating parameters of the levitator by direct observations of the levitated samples, as opposed to more traditional pressure mapping of the acoustic field. The size range of samples that could be reliably heated and quenched in this setup was determined to be on the order of 1--3 mm. In microgravity experiments, small spherical specimens (1--2 mm diameter), prepared as pressed, premixed solid components, ZrO2, ZrN, and Zr powders, were acoustically levitated inside an argon-filled chamber at one atmosphere and heated by a CO2 laser. The levitating samples could be continuously laser heated for about 1 sec, resulting in local sample melting. The sample stability in the vertical direction was undisturbed by simultaneous laser heating. Oscillations of the levitating sample in the horizontal direction increased while it was heated, which eventually resulted in the movement of the sample away from its stable levitation position and the laser
Nandal, Preeti; Ravella, Sreenivas Rao; Kuhad, Ramesh Chander
2013-01-01
Laccase production by Coriolopsis caperata RCK2011 under solid state fermentation was optimized following Taguchi design of experiment. An orthogonal array layout of L18 (21 × 37) was constructed using Qualitek-4 software with eight most influensive factors on laccase production. At individual level pH contributed higher influence, whereas, corn steep liquor (CSL) accounted for more than 50% of the severity index with biotin and KH2PO4 at the interactive level. The optimum conditions derived were; temperature 30°C, pH 5.0, wheat bran 5.0 g, inoculum size 0.5 ml (fungal cell mass = 0.015 g dry wt.), biotin 0.5% w/v, KH2PO4 0.013% w/v, CSL 0.1% v/v and 0.5 mM xylidine as an inducer. The validation experiments using optimized conditions confirmed an improvement in enzyme production by 58.01%. The laccase production to the level of 1623.55 Ugds−1 indicates that the fungus C. caperata RCK2011 has the commercial potential for laccase. PMID:23463372
Macready, Anna L; Fallaize, Rosalind; Butler, Laurie T; Ellis, Judi A; Kuznesof, Sharron; Frewer, Lynn J; Celis-Morales, Carlos; Livingstone, Katherine M; Araújo-Soares, Vera; Fischer, Arnout RH; Stewart-Knox, Barbara J; Mathers, John C
2018-01-01
Background To determine the efficacy of behavior change techniques applied in dietary and physical activity intervention studies, it is first necessary to record and describe techniques that have been used during such interventions. Published frameworks used in dietary and smoking cessation interventions undergo continuous development, and most are not adapted for Web-based delivery. The Food4Me study (N=1607) provided the opportunity to use existing frameworks to describe standardized Web-based techniques employed in a large-scale, internet-based intervention to change dietary behavior and physical activity. Objective The aims of this study were (1) to describe techniques embedded in the Food4Me study design and explain the selection rationale and (2) to demonstrate the use of behavior change technique taxonomies, develop standard operating procedures for training, and identify strengths and limitations of the Food4Me framework that will inform its use in future studies. Methods The 6-month randomized controlled trial took place simultaneously in seven European countries, with participants receiving one of four levels of personalized advice (generalized, intake-based, intake+phenotype–based, and intake+phenotype+gene–based). A three-phase approach was taken: (1) existing taxonomies were reviewed and techniques were identified a priori for possible inclusion in the Food4Me study, (2) a standard operating procedure was developed to maintain consistency in the use of methods and techniques across research centers, and (3) the Food4Me behavior change technique framework was reviewed and updated post intervention. An analysis of excluded techniques was also conducted. Results Of 46 techniques identified a priori as being applicable to Food4Me, 17 were embedded in the intervention design; 11 were from a dietary taxonomy, and 6 from a smoking cessation taxonomy. In addition, the four-category smoking cessation framework structure was adopted for clarity of
NASA Astrophysics Data System (ADS)
Fatrias, D.; Kamil, I.; Meilani, D.
2018-03-01
Coordinating business operation with suppliers becomes increasingly important to survive and prosper under the dynamic business environment. A good partnership with suppliers not only increase efficiency, but also strengthen corporate competitiveness. Associated with such concern, this study aims to develop a practical approach of multi-criteria supplier evaluation using combined methods of Taguchi loss function (TLF), best-worst method (BWM) and VIse Kriterijumska Optimizacija kompromisno Resenje (VIKOR). A new framework of integrative approach adopting these methods is our main contribution for supplier evaluation in literature. In this integrated approach, a compromised supplier ranking list based on the loss score of suppliers is obtained using efficient steps of a pairwise comparison based decision making process. Implemetation to the case problem with real data from crumb rubber industry shows the usefulness of the proposed approach. Finally, a suitable managerial implication is presented.
Davis, Stacy N; Govindaraju, Swapamthi; Jackson, Brittany; Williams, Kimberly R; Christy, Shannon M; Vadaparampil, Susan T; Quinn, Gwendolyn P; Shibata, David; Roetzheim, Richard; Meade, Cathy D; Gwede, Clement K
Recruiting ethnically diverse Black participants to an innovative, community-based research study to reduce colorectal cancer screening disparities requires multipronged recruitment techniques. This article describes active, passive, and snowball recruitment techniques, and challenges and lessons learned in recruiting a diverse sample of Black participants. For each of the three recruitment techniques, data were collected on strategies, enrollment efficiency (participants enrolled/participants evaluated), and reasons for ineligibility. Five hundred sixty individuals were evaluated, and 330 individuals were enrolled. Active recruitment yielded the highest number of enrolled participants, followed by passive and snowball. Snowball recruitment was the most efficient technique, with enrollment efficiency of 72.4%, followed by passive (58.1%) and active (55.7%) techniques. There were significant differences in gender, education, country of origin, health insurance, and having a regular physician by recruitment technique (p < .05). Multipronged recruitment techniques should be employed to increase reach, diversity, and study participation rates among Blacks. Although each recruitment technique had a variable enrollment efficiency, the use of multipronged recruitment techniques can lead to successful enrollment of diverse Blacks into cancer prevention and control interventions.
Spasojević Brkić, Vesna K; Veljković, Zorica A; Golubović, Tamara; Brkić, Aleksandar Dj; Kosić Šotić, Ivana
2016-01-01
Procedures in the development process of crane cabins are arbitrary and subjective. Since approximately 42% of incidents in the construction industry are linked to them, there is a need to collect fresh anthropometric data and provide additional recommendations for design. In this paper, dimensioning of the crane cabin interior space was carried out using a sample of 64 crane operators' anthropometric measurements, in the Republic of Serbia, by measuring workspace with 10 parameters using nine measured anthropometric data from each crane operator. This paper applies experiments run via full factorial designs using a combined traditional and Taguchi approach. The experiments indicated which design parameters are influenced by which anthropometric measurements and to what degree. The results are expected to be of use for crane cabin designers and should assist them to design a cabin that may lead to less strenuous sitting postures and fatigue for operators, thus improving safety and accident prevention.
NASA Astrophysics Data System (ADS)
Yudhanto, F.; Jamasri; Rochardjo, Heru S. B.
2018-05-01
The characterized agave cantala fiber in this research came from Sumenep, Madura, Indonesia was chemically processed using sodium hydroxide (NaOH) and hydrogen peroxide (H2O2) solution. The treatment with both solutions is called bleaching process. Tensile strength test of single fiber was used to get mechanical properties from selecting process of the various parameter are temperature, PH and concentration of H2O2 with an L9 orthogonal array by Taguchi method. The results indicate that PH is most significant parameter influencing the tensile strength followed by temperature and concentration H2O2. The influence of bleaching treatment on tensile strength showed increasing of crystallinity index of fiber by 21%. It showed by lost of hemicellulose and lignin layers of fiber can be seen from waveforms changes of 1735 (C=O), 1627 (OH), 1319 (CH2), 1250 (C-O) by FTIR graph. The photo SEM showed that the bleaching of fibers causes the fibers more roughly and clearly than untreated fibers.
Sutcu, Mucahit; Ozturk, Savas; Yalamac, Emre; Gencel, Osman
2016-10-01
Production of porous clay bricks lightened by adding olive mill waste as a pore making additive was investigated. Factors influencing the brick manufacturing process were analyzed by an experimental design, Taguchi method, to find out the most favorable conditions for the production of bricks. The optimum process conditions for brick preparation were investigated by studying the effects of mixture ratios (0, 5 and 10 wt%) and firing temperatures (850, 950 and 1050 °C) on the physical, thermal and mechanical properties of the bricks. Apparent density, bulk density, apparent porosity, water absorption, compressive strength, thermal conductivity, microstructure and crystalline phase formations of the fired brick samples were measured. It was found that the use of 10% waste addition reduced the bulk density of the samples up to 1.45 g/cm(3). As the porosities increased from 30.8 to 47.0%, the compressive strengths decreased from 36.9 to 10.26 MPa at firing temperature of 950 °C. The thermal conductivities of samples fired at the same temperature showed a decrease of 31% from 0.638 to 0.436 W/mK, which is hopeful for heat insulation in the buildings. Increasing of the firing temperature also affected their mechanical and physical properties. This study showed that the olive mill waste could be used as a pore maker in brick production. Copyright © 2016 Elsevier Ltd. All rights reserved.
Design and Evaluation of Perceptual-based Object Group Selection Techniques
NASA Astrophysics Data System (ADS)
Dehmeshki, Hoda
Selecting groups of objects is a frequent task in graphical user interfaces. It is required prior to many standard operations such as deletion, movement, or modification. Conventional selection techniques are lasso, rectangle selection, and the selection and de-selection of items through the use of modifier keys. These techniques may become time-consuming and error-prone when target objects are densely distributed or when the distances between target objects are large. Perceptual-based selection techniques can considerably improve selection tasks when targets have a perceptual structure, for example when arranged along a line. Current methods to detect such groups use ad hoc grouping algorithms that are not based on results from perception science. Moreover, these techniques do not allow selecting groups with arbitrary arrangements or permit modifying a selection. This dissertation presents two domain-independent perceptual-based systems that address these issues. Based on established group detection models from perception research, the proposed systems detect perceptual groups formed by the Gestalt principles of good continuation and proximity. The new systems provide gesture-based or click-based interaction techniques for selecting groups with curvilinear or arbitrary structures as well as clusters. Moreover, the gesture-based system is adapted for the graph domain to facilitate path selection. This dissertation includes several user studies that show the proposed systems outperform conventional selection techniques when targets form salient perceptual groups and are still competitive when targets are semi-structured.
Development of New Lipid-Based Paclitaxel Nanoparticles Using Sequential Simplex Optimization
Dong, Xiaowei; Mattingly, Cynthia A.; Tseng, Michael; Cho, Moo; Adams, Val R.; Mumper, Russell J.
2008-01-01
The objective of these studies was to develop Cremophor-free lipid-based paclitaxel (PX) nanoparticle formulations prepared from warm microemulsion precursors. To identify and optimize new nanoparticles, experimental design was performed combining Taguchi array and sequential simplex optimization. The combination of Taguchi array and sequential simplex optimization efficiently directed the design of paclitaxel nanoparticles. Two optimized paclitaxel nanoparticles (NPs) were obtained: G78 NPs composed of glyceryl tridodecanoate (GT) and polyoxyethylene 20-stearyl ether (Brij 78), and BTM NPs composed of Miglyol 812, Brij 78 and D-alpha-tocopheryl polyethylene glycol 1000 succinate (TPGS). Both nanoparticles successfully entrapped paclitaxel at a final concentration of 150 μg/ml (over 6% drug loading) with particle sizes less than 200 nm and over 85% of entrapment efficiency. These novel paclitaxel nanoparticles were stable at 4°C over three months and in PBS at 37°C over 102 hours as measured by physical stability. Release of paclitaxel was slow and sustained without initial burst release. Cytotoxicity studies in MDA-MB-231 cancer cells showed that both nanoparticles have similar anticancer activities compared to Taxol®. Interestingly, PX BTM nanocapsules could be lyophilized without cryoprotectants. The lyophilized powder comprised only of PX BTM NPs in water could be rapidly rehydrated with complete retention of original physicochemical properties, in-vitro release properties, and cytotoxicity profile. Sequential Simplex Optimization has been utilized to identify promising new lipid-based paclitaxel nanoparticles having useful attributes. PMID:19111929
Venkata Mohan, S; Chandrasekhara Rao, N; Krishna Prasad, K; Murali Krishna, P; Sreenivas Rao, R; Sarma, P N
2005-06-20
The Taguchi robust experimental design (DOE) methodology has been applied on a dynamic anaerobic process treating complex wastewater by an anaerobic sequencing batch biofilm reactor (AnSBBR). For optimizing the process as well as to evaluate the influence of different factors on the process, the uncontrollable (noise) factors have been considered. The Taguchi methodology adopting dynamic approach is the first of its kind for studying anaerobic process evaluation and process optimization. The designed experimental methodology consisted of four phases--planning, conducting, analysis, and validation connected sequence-wise to achieve the overall optimization. In the experimental design, five controllable factors, i.e., organic loading rate (OLR), inlet pH, biodegradability (BOD/COD ratio), temperature, and sulfate concentration, along with the two uncontrollable (noise) factors, volatile fatty acids (VFA) and alkalinity at two levels were considered for optimization of the anae robic system. Thirty-two anaerobic experiments were conducted with a different combination of factors and the results obtained in terms of substrate degradation rates were processed in Qualitek-4 software to study the main effect of individual factors, interaction between the individual factors, and signal-to-noise (S/N) ratio analysis. Attempts were also made to achieve optimum conditions. Studies on the influence of individual factors on process performance revealed the intensive effect of OLR. In multiple factor interaction studies, biodegradability with other factors, such as temperature, pH, and sulfate have shown maximum influence over the process performance. The optimum conditions for the efficient performance of the anaerobic system in treating complex wastewater by considering dynamic (noise) factors obtained are higher organic loading rate of 3.5 Kg COD/m3 day, neutral pH with high biodegradability (BOD/COD ratio of 0.5), along with mesophilic temperature range (40 degrees C), and
Manufacturing Research: Self-Directed Control
1991-01-01
reduce this sensitivity. SDO is performing Taguchi’s parameter design . 1-13 Statistical Process Control SPC techniques will be used to monitor the process...Florida,R.E. Krieger Pub. Co., 1988. Dehnad, Khowrow, Quality Control . Robust Design . and the Taguchi Method, Pacific Grove, California, Wadsworth... control system. This turns out to be a non -trivial exercise. A human operator can see an event occur (such as the vessel pressurizing above its setpoint
NASA Astrophysics Data System (ADS)
Lin, Yuan; Choudhury, Kingshuk R.; McAdams, H. Page; Foos, David H.; Samei, Ehsan
2014-03-01
We previously proposed a novel image-based quality assessment technique1 to assess the perceptual quality of clinical chest radiographs. In this paper, an observer study was designed and conducted to systematically validate this technique. Ten metrics were involved in the observer study, i.e., lung grey level, lung detail, lung noise, riblung contrast, rib sharpness, mediastinum detail, mediastinum noise, mediastinum alignment, subdiaphragm-lung contrast, and subdiaphragm area. For each metric, three tasks were successively presented to the observers. In each task, six ROI images were randomly presented in a row and observers were asked to rank the images only based on a designated quality and disregard the other qualities. A range slider on the top of the images was used for observers to indicate the acceptable range based on the corresponding perceptual attribute. Five boardcertificated radiologists from Duke participated in this observer study on a DICOM calibrated diagnostic display workstation and under low ambient lighting conditions. The observer data were analyzed in terms of the correlations between the observer ranking orders and the algorithmic ranking orders. Based on the collected acceptable ranges, quality consistency ranges were statistically derived. The observer study showed that, for each metric, the averaged ranking orders of the participated observers were strongly correlated with the algorithmic orders. For the lung grey level, the observer ranking orders completely accorded with the algorithmic ranking orders. The quality consistency ranges derived from this observer study were close to these derived from our previous study. The observer study indicates that the proposed image-based quality assessment technique provides a robust reflection of the perceptual image quality of the clinical chest radiographs. The derived quality consistency ranges can be used to automatically predict the acceptability of a clinical chest radiograph.
NASA Astrophysics Data System (ADS)
Ghani, Jaharah A.; Mohd Rodzi, Mohd Nor Azmi; Zaki Nuawi, Mohd; Othman, Kamal; Rahman, Mohd. Nizam Ab.; Haron, Che Hassan Che; Deros, Baba Md
2011-01-01
Machining is one of the most important manufacturing processes in these modern industries especially for finishing an automotive component after the primary manufacturing processes such as casting and forging. In this study the turning parameters of dry cutting environment (without air, normal air and chilled air), various cutting speed, and feed rate are evaluated using a Taguchi optimization methodology. An orthogonal array L27 (313), signal-to-noise (S/N) ratio and analysis of variance (ANOVA) are employed to analyze the effect of these turning parameters on the performance of a coated carbide tool. The results show that the tool life is affected by the cutting speed, feed rate and cutting environment with contribution of 38%, 32% and 27% respectively. Whereas for the surface roughness, the feed rate is significantly controlled the machined surface produced by 77%, followed by the cutting environment of 19%. The cutting speed is found insignificant in controlling the machined surface produced. The study shows that the dry cutting environment factor should be considered in order to produce longer tool life as well as for obtaining a good machined surface.
Macready, Anna L; Fallaize, Rosalind; Butler, Laurie T; Ellis, Judi A; Kuznesof, Sharron; Frewer, Lynn J; Celis-Morales, Carlos; Livingstone, Katherine M; Araújo-Soares, Vera; Fischer, Arnout Rh; Stewart-Knox, Barbara J; Mathers, John C; Lovegrove, Julie A
2018-04-09
To determine the efficacy of behavior change techniques applied in dietary and physical activity intervention studies, it is first necessary to record and describe techniques that have been used during such interventions. Published frameworks used in dietary and smoking cessation interventions undergo continuous development, and most are not adapted for Web-based delivery. The Food4Me study (N=1607) provided the opportunity to use existing frameworks to describe standardized Web-based techniques employed in a large-scale, internet-based intervention to change dietary behavior and physical activity. The aims of this study were (1) to describe techniques embedded in the Food4Me study design and explain the selection rationale and (2) to demonstrate the use of behavior change technique taxonomies, develop standard operating procedures for training, and identify strengths and limitations of the Food4Me framework that will inform its use in future studies. The 6-month randomized controlled trial took place simultaneously in seven European countries, with participants receiving one of four levels of personalized advice (generalized, intake-based, intake+phenotype-based, and intake+phenotype+gene-based). A three-phase approach was taken: (1) existing taxonomies were reviewed and techniques were identified a priori for possible inclusion in the Food4Me study, (2) a standard operating procedure was developed to maintain consistency in the use of methods and techniques across research centers, and (3) the Food4Me behavior change technique framework was reviewed and updated post intervention. An analysis of excluded techniques was also conducted. Of 46 techniques identified a priori as being applicable to Food4Me, 17 were embedded in the intervention design; 11 were from a dietary taxonomy, and 6 from a smoking cessation taxonomy. In addition, the four-category smoking cessation framework structure was adopted for clarity of communication. Smoking cessation texts were adapted
A gEUD-based inverse planning technique for HDR prostate brachytherapy: Feasibility study
DOE Office of Scientific and Technical Information (OSTI.GOV)
Giantsoudi, D.; Department of Radiation Oncology, Francis H. Burr Proton Therapy Center, Boston, Massachusetts 02114; Baltas, D.
2013-04-15
Purpose: The purpose of this work was to study the feasibility of a new inverse planning technique based on the generalized equivalent uniform dose for image-guided high dose rate (HDR) prostate cancer brachytherapy in comparison to conventional dose-volume based optimization. Methods: The quality of 12 clinical HDR brachytherapy implants for prostate utilizing HIPO (Hybrid Inverse Planning Optimization) is compared with alternative plans, which were produced through inverse planning using the generalized equivalent uniform dose (gEUD). All the common dose-volume indices for the prostate and the organs at risk were considered together with radiobiological measures. The clinical effectiveness of the differentmore » dose distributions was investigated by comparing dose volume histogram and gEUD evaluators. Results: Our results demonstrate the feasibility of gEUD-based inverse planning in HDR brachytherapy implants for prostate. A statistically significant decrease in D{sub 10} or/and final gEUD values for the organs at risk (urethra, bladder, and rectum) was found while improving dose homogeneity or dose conformity of the target volume. Conclusions: Following the promising results of gEUD-based optimization in intensity modulated radiation therapy treatment optimization, as reported in the literature, the implementation of a similar model in HDR brachytherapy treatment plan optimization is suggested by this study. The potential of improved sparing of organs at risk was shown for various gEUD-based optimization parameter protocols, which indicates the ability of this method to adapt to the user's preferences.« less
Taguchi Optimization of Cutting Parameters in Turning AISI 1020 MS with M2 HSS Tool
NASA Astrophysics Data System (ADS)
Sonowal, Dharindom; Sarma, Dhrupad; Bakul Barua, Parimal; Nath, Thuleswar
2017-08-01
In this paper the effect of three cutting parameters viz. Spindle speed, Feed and Depth of Cut on surface roughness of AISI 1020 mild steel bar in turning was investigated and optimized to obtain minimum surface roughness. All the experiments are conducted on HMT LB25 lathe machine using M2 HSS cutting tool. Ranges of parameters of interest have been decided through some preliminary experimentation (One Factor At a Time experiments). Finally a combined experiment has been carried out using Taguchi’s L27 Orthogonal Array (OA) to study the main effect and interaction effect of the all three parameters. The experimental results were analyzed with raw data ANOVA (Analysis of Variance) and S/N data (Signal to Noise ratio) ANOVA. Results show that Spindle speed, Feed and Depth of Cut have significant effects on both mean and variation of surface roughness in turning AISI 1020 mild steel. Mild two factors interactions are observed among the aforesaid factors with significant effects only on the mean of the output variable. From the Taguchi parameter optimization the optimum factor combination is found to be 630 rpm spindle speed, 0.05 mm/rev feed and 1.25 mm depth of cut with estimated surface roughness 2.358 ± 0.970 µm. A confirmatory experiment was conducted with the optimum factor combination to verify the results. In the confirmatory experiment the average value of surface roughness is found to be 2.408 µm which is well within the range (0.418 µm to 4.299 µm) predicted for confirmatory experiment.
Efficient techniques for wave-based sound propagation in interactive applications
NASA Astrophysics Data System (ADS)
Mehra, Ravish
-driven, rotating or time-varying directivity function at runtime. Unlike previous approaches, the listener directivity approach can be used to compute spatial audio (3D audio) for a moving, rotating listener at interactive rates. Lastly, we propose an efficient GPU-based time-domain solver for the wave equation that enables wave simulation up to the mid-frequency range in tens of minutes on a desktop computer. It is demonstrated that by carefully mapping all the components of the wave simulator to match the parallel processing capabilities of the graphics processors, significant improvement in performance can be achieved compared to the CPU-based simulators, while maintaining numerical accuracy. We validate these techniques with offline numerical simulations and measured data recorded in an outdoor scene. We present results of preliminary user evaluations conducted to study the impact of these techniques on user's immersion in virtual environment. We have integrated these techniques with the Half-Life 2 game engine, Oculus Rift head-mounted display, and Xbox game controller to enable users to experience high-quality acoustics effects and spatial audio in the virtual environment.
Data preparation techniques for a perinatal psychiatric study based on linked data.
Xu, Fenglian; Hilder, Lisa; Austin, Marie-Paule; Sullivan, Elizabeth A
2012-06-08
In recent years there has been an increase in the use of population-based linked data. However, there is little literature that describes the method of linked data preparation. This paper describes the method for merging data, calculating the statistical variable (SV), recoding psychiatric diagnoses and summarizing hospital admissions for a perinatal psychiatric study. The data preparation techniques described in this paper are based on linked birth data from the New South Wales (NSW) Midwives Data Collection (MDC), the Register of Congenital Conditions (RCC), the Admitted Patient Data Collection (APDC) and the Pharmaceutical Drugs of Addiction System (PHDAS). The master dataset is the meaningfully linked data which include all or major study data collections. The master dataset can be used to improve the data quality, calculate the SV and can be tailored for different analyses. To identify hospital admissions in the periods before pregnancy, during pregnancy and after birth, a statistical variable of time interval (SVTI) needs to be calculated. The methods and SPSS syntax for building a master dataset, calculating the SVTI, recoding the principal diagnoses of mental illness and summarizing hospital admissions are described. Linked data preparation, including building the master dataset and calculating the SV, can improve data quality and enhance data function.
Parameters optimization of laser brazing in crimping butt using Taguchi and BPNN-GA
NASA Astrophysics Data System (ADS)
Rong, Youmin; Zhang, Zhen; Zhang, Guojun; Yue, Chen; Gu, Yafei; Huang, Yu; Wang, Chunming; Shao, Xinyu
2015-04-01
The laser brazing (LB) is widely used in the automotive industry due to the advantages of high speed, small heat affected zone, high quality of welding seam, and low heat input. Welding parameters play a significant role in determining the bead geometry and hence quality of the weld joint. This paper addresses the optimization of the seam shape in LB process with welding crimping butt of 0.8 mm thickness using back propagation neural network (BPNN) and genetic algorithm (GA). A 3-factor, 5-level welding experiment is conducted by Taguchi L25 orthogonal array through the statistical design method. Then, the input parameters are considered here including welding speed, wire speed rate, and gap with 5 levels. The output results are efficient connection length of left side and right side, top width (WT) and bottom width (WB) of the weld bead. The experiment results are embed into the BPNN network to establish relationship between the input and output variables. The predicted results of the BPNN are fed to GA algorithm that optimizes the process parameters subjected to the objectives. Then, the effects of welding speed (WS), wire feed rate (WF), and gap (GAP) on the sum values of bead geometry is discussed. Eventually, the confirmation experiments are carried out to demonstrate the optimal values were effective and reliable. On the whole, the proposed hybrid method, BPNN-GA, can be used to guide the actual work and improve the efficiency and stability of LB process.
Subroto, Erna; Widjojokusumo, Edward; Veriansyah, Bambang; Tjandrawinata, Raymond R
2017-04-01
A series of experiments was conducted to determine optimum conditions for supercritical carbon dioxide extraction of candlenut oil. A Taguchi experimental design with L 9 orthogonal array (four factors in three levels) was employed to evaluate the effects of pressure of 25-35 MPa, temperature of 40-60 °C, CO 2 flow rate of 10-20 g/min and particle size of 0.3-0.8 mm on oil solubility. The obtained results showed that increase in particle size, pressure and temperature improved the oil solubility. The supercritical carbon dioxide extraction at optimized parameters resulted in oil yield extraction of 61.4% at solubility of 9.6 g oil/kg CO 2 . The obtained candlenut oil from supercritical carbon dioxide extraction has better oil quality than oil which was extracted by Soxhlet extraction using n-hexane. The oil contains high unsaturated oil (linoleic acid and linolenic acid), which have many beneficial effects on human health.
Eslami, Taban; Saeed, Fahad
2018-04-20
Functional magnetic resonance imaging (fMRI) is a non-invasive brain imaging technique, which has been regularly used for studying brain’s functional activities in the past few years. A very well-used measure for capturing functional associations in brain is Pearson’s correlation coefficient. Pearson’s correlation is widely used for constructing functional network and studying dynamic functional connectivity of the brain. These are useful measures for understanding the effects of brain disorders on connectivities among brain regions. The fMRI scanners produce huge number of voxels and using traditional central processing unit (CPU)-based techniques for computing pairwise correlations is very time consuming especially when large number of subjects are being studied. In this paper, we propose a graphics processing unit (GPU)-based algorithm called Fast-GPU-PCC for computing pairwise Pearson’s correlation coefficient. Based on the symmetric property of Pearson’s correlation, this approach returns N ( N − 1 ) / 2 correlation coefficients located at strictly upper triangle part of the correlation matrix. Storing correlations in a one-dimensional array with the order as proposed in this paper is useful for further usage. Our experiments on real and synthetic fMRI data for different number of voxels and varying length of time series show that the proposed approach outperformed state of the art GPU-based techniques as well as the sequential CPU-based versions. We show that Fast-GPU-PCC runs 62 times faster than CPU-based version and about 2 to 3 times faster than two other state of the art GPU-based methods.
Applying knowledge compilation techniques to model-based reasoning
NASA Technical Reports Server (NTRS)
Keller, Richard M.
1991-01-01
Researchers in the area of knowledge compilation are developing general purpose techniques for improving the efficiency of knowledge-based systems. In this article, an attempt is made to define knowledge compilation, to characterize several classes of knowledge compilation techniques, and to illustrate how some of these techniques can be applied to improve the performance of model-based reasoning systems.
NASA Astrophysics Data System (ADS)
'Ainullotfi, A. A.; Ibrahim, A. L.; Masron, T.
2014-02-01
This study is conducted to establish a community based flood management system that is integrated with remote sensing technique. To understand local knowledge, the demographic of the local society is obtained by using the survey approach. The local authorities are approached first to obtain information regarding the society in the study areas such as the population, the gender and the tabulation of settlement. The information about age, religion, ethnic, occupation, years of experience facing flood in the area, are recorded to understand more on how the local knowledge emerges. Then geographic data is obtained such as rainfall data, land use, land elevation, river discharge data. This information is used to establish a hydrological model of flood in the study area. Analysis were made from the survey approach to understand the pattern of society and how they react to floods while the analysis of geographic data is used to analyse the water extent and damage done by the flood. The final result of this research is to produce a flood mitigation method with a community based framework in the state of Kelantan. With the flood mitigation that involves the community's understanding towards flood also the techniques to forecast heavy rainfall and flood occurrence using remote sensing, it is hope that it could reduce the casualties and damage that might cause to the society and infrastructures in the study area.
NASA Technical Reports Server (NTRS)
Haste, Deepak; Ghoshal, Sudipto; Johnson, Stephen B.; Moore, Craig
2018-01-01
This paper describes the theory and considerations in the application of model-based techniques to assimilate information from disjoint knowledge sources for performing NASA's Fault Management (FM)-related activities using the TEAMS® toolset. FM consists of the operational mitigation of existing and impending spacecraft failures. NASA's FM directives have both design-phase and operational-phase goals. This paper highlights recent studies by QSI and DST of the capabilities required in the TEAMS® toolset for conducting FM activities with the aim of reducing operating costs, increasing autonomy, and conforming to time schedules. These studies use and extend the analytic capabilities of QSI's TEAMS® toolset to conduct a range of FM activities within a centralized platform.
Learning Physics through Project-Based Learning Game Techniques
ERIC Educational Resources Information Center
Baran, Medine; Maskan, Abdulkadir; Yasar, Seyma
2018-01-01
The aim of the present study, in which Project and game techniques are used together, is to examine the impact of project-based learning games on students' physics achievement. Participants of the study consist of 34 9th grade students (N = 34). The data were collected using achievement tests and a questionnaire. Throughout the applications, the…
Techniques for Enhancing Web-Based Education.
ERIC Educational Resources Information Center
Barbieri, Kathy; Mehringer, Susan
The Virtual Workshop is a World Wide Web-based set of modules on high performance computing developed at the Cornell Theory Center (CTC) (New York). This approach reaches a large audience, leverages staff effort, and poses challenges for developing interesting presentation techniques. This paper describes the following techniques with their…
Application of Optimization Techniques to Design of Unconventional Rocket Nozzle Configurations
NASA Technical Reports Server (NTRS)
Follett, W.; Ketchum, A.; Darian, A.; Hsu, Y.
1996-01-01
Several current rocket engine concepts such as the bell-annular tri-propellant engine, and the linear aerospike being proposed for the X-33 require unconventional three dimensional rocket nozzles which must conform to rectangular or sector shaped envelopes to meet integration constraints. These types of nozzles exist outside the current experience database, therefore, the application of efficient design methods for these propulsion concepts is critical to the success of launch vehicle programs. The objective of this work is to optimize several different nozzle configurations, including two- and three-dimensional geometries. Methodology includes coupling computational fluid dynamic (CFD) analysis to genetic algorithms and Taguchi methods as well as implementation of a streamline tracing technique. Results of applications are shown for several geometeries including: three dimensional thruster nozzles with round or super elliptic throats and rectangualar exits, two- and three-dimensional thrusters installed within a bell nozzle, and three dimensional thrusters with round throats and sector shaped exits. Due to the novel designs considered for this study, there is little experience which can be used to guide the effort and limit the design space. With a nearly infinite parameter space to explore, simple parametric design studies cannot possibly search the entire design space within the time frame required to impact the design cycle. For this reason, robust and efficient optimization methods are required to explore and exploit the design space to achieve high performance engine designs. Five case studies which examine the application of various techniques in the engineering environment are presented in this paper.
NASA Astrophysics Data System (ADS)
Soepangkat, Bobby O. P.; Suhardjono, Pramujati, Bambang
2017-06-01
Machining under minimum quantity lubrication (MQL) has drawn the attention of researchers as an alternative to the traditionally used wet and dry machining conditions with the purpose to minimize the cooling and lubricating cost, as well as to reduce cutting zone temperature, tool wear, and hole surface roughness. Drilling is one of the important operations to assemble machine components. The objective of this study was to optimize drilling parameters such as cutting feed and cutting speed, drill type and drill point angle on the thrust force, torque, hole surface roughness and tool flank wear in drilling EMS 45 tool steel using MQL. In this study, experiments were carried out as per Taguchi design of experiments while an L18 orthogonal array was used to study the influence of various combinations of drilling parameters and tool geometries on the thrust force, torque, hole surface roughness and tool flank wear. The optimum drilling parameters was determined by using grey relational grade obtained from grey relational analysis for multiple-performance characteristics. The drilling experiments were carried out by using twist drill and CNC machining center. This work is useful for optimum values selection of various drilling parameters and tool geometries that would not only minimize the thrust force and torque, but also reduce hole surface roughness and tool flank wear.
The efficacy of the 'mind map' study technique.
Farrand, Paul; Hussain, Fearzana; Hennessy, Enid
2002-05-01
To examine the effectiveness of using the 'mind map' study technique to improve factual recall from written information. To obtain baseline data, subjects completed a short test based on a 600-word passage of text prior to being randomly allocated to form two groups: 'self-selected study technique' and 'mind map'. After a 30-minute interval the self-selected study technique group were exposed to the same passage of text previously seen and told to apply existing study techniques. Subjects in the mind map group were trained in the mind map technique and told to apply it to the passage of text. Recall was measured after an interfering task and a week later. Measures of motivation were taken. Barts and the London School of Medicine and Dentistry, University of London. 50 second- and third-year medical students. Recall of factual material improved for both the mind map and self-selected study technique groups at immediate test compared with baseline. However this improvement was only robust after a week for those in the mind map group. At 1 week, the factual knowledge in the mind map group was greater by 10% (adjusting for baseline) (95% CI -1% to 22%). However motivation for the technique used was lower in the mind map group; if motivation could have been made equal in the groups, the improvement with mind mapping would have been 15% (95% CI 3% to 27%). Mind maps provide an effective study technique when applied to written material. However before mind maps are generally adopted as a study technique, consideration has to be given towards ways of improving motivation amongst users.
Study on key techniques for camera-based hydrological record image digitization
NASA Astrophysics Data System (ADS)
Li, Shijin; Zhan, Di; Hu, Jinlong; Gao, Xiangtao; Bo, Ping
2015-10-01
With the development of information technology, the digitization of scientific or engineering drawings has received more and more attention. In hydrology, meteorology, medicine and mining industry, the grid drawing sheet is commonly used to record the observations from sensors. However, these paper drawings may be destroyed and contaminated due to improper preservation or overuse. Further, it will be a heavy workload and prone to error if these data are manually transcripted into the computer. Hence, in order to digitize these drawings, establishing the corresponding data base will ensure the integrity of data and provide invaluable information for further research. This paper presents an automatic system for hydrological record image digitization, which consists of three key techniques, i.e., image segmentation, intersection point localization and distortion rectification. First, a novel approach to the binarization of the curves and grids in the water level sheet image has been proposed, which is based on the fusion of gradient and color information adaptively. Second, a fast search strategy for cross point location is invented and point-by-point processing is thus avoided, with the help of grid distribution information. And finally, we put forward a local rectification method through analyzing the central portions of the image and utilizing the domain knowledge of hydrology. The processing speed is accelerated, while the accuracy is still satisfying. Experiments on several real water level records show that our proposed techniques are effective and capable of recovering the hydrological observations accurately.
NASA Astrophysics Data System (ADS)
Mariajayaprakash, Arokiasamy; Senthilvelan, Thiyagarajan; Vivekananthan, Krishnapillai Ponnambal
2013-07-01
The various process parameters affecting the quality characteristics of the shock absorber during the process were identified using the Ishikawa diagram and by failure mode and effect analysis. The identified process parameters are welding process parameters (squeeze, heat control, wheel speed, and air pressure), damper sealing process parameters (load, hydraulic pressure, air pressure, and fixture height), washing process parameters (total alkalinity, temperature, pH value of rinsing water, and timing), and painting process parameters (flowability, coating thickness, pointage, and temperature). In this paper, the process parameters, namely, painting and washing process parameters, are optimized by Taguchi method. Though the defects are reasonably minimized by Taguchi method, in order to achieve zero defects during the processes, genetic algorithm technique is applied on the optimized parameters obtained by Taguchi method.
Emami, J; Mohiti, H; Hamishehkar, H; Varshosaz, J
2015-01-01
Budesonide is a potent non-halogenated corticosteroid with high anti-inflammatory effects. The lungs are an attractive route for non-invasive drug delivery with advantages for both systemic and local applications. The aim of the present study was to develop, characterize and optimize a solid lipid nanoparticle system to deliver budesonide to the lungs. Budesonide-loaded solid lipid nanoparticles were prepared by the emulsification-solvent diffusion method. The impact of various processing variables including surfactant type and concentration, lipid content organic and aqueous volume, and sonication time were assessed on the particle size, zeta potential, entrapment efficiency, loading percent and mean dissolution time. Taguchi design with 12 formulations along with Box-Behnken design with 17 formulations was developed. The impact of each factor upon the eventual responses was evaluated, and the optimized formulation was finally selected. The size and morphology of the prepared nanoparticles were studied using scanning electron microscope. Based on the optimization made by Design Expert 7(®) software, a formulation made of glycerol monostearate, 1.2 % polyvinyl alcohol (PVA), weight ratio of lipid/drug of 10 and sonication time of 90 s was selected. Particle size, zeta potential, entrapment efficiency, loading percent, and mean dissolution time of adopted formulation were predicted and confirmed to be 218.2 ± 6.6 nm, -26.7 ± 1.9 mV, 92.5 ± 0.52 %, 5.8 ± 0.3 %, and 10.4 ± 0.29 h, respectively. Since the preparation and evaluation of the selected formulation within the laboratory yielded acceptable results with low error percent, the modeling and optimization was justified. The optimized formulation co-spray dried with lactose (hybrid microparticles) displayed desirable fine particle fraction, mass median aerodynamic diameter (MMAD), and geometric standard deviation of 49.5%, 2.06 μm, and 2.98 μm; respectively. Our results provide fundamental data for the
Emami, J.; Mohiti, H.; Hamishehkar, H.; Varshosaz, J.
2015-01-01
Budesonide is a potent non-halogenated corticosteroid with high anti-inflammatory effects. The lungs are an attractive route for non-invasive drug delivery with advantages for both systemic and local applications. The aim of the present study was to develop, characterize and optimize a solid lipid nanoparticle system to deliver budesonide to the lungs. Budesonide-loaded solid lipid nanoparticles were prepared by the emulsification-solvent diffusion method. The impact of various processing variables including surfactant type and concentration, lipid content organic and aqueous volume, and sonication time were assessed on the particle size, zeta potential, entrapment efficiency, loading percent and mean dissolution time. Taguchi design with 12 formulations along with Box-Behnken design with 17 formulations was developed. The impact of each factor upon the eventual responses was evaluated, and the optimized formulation was finally selected. The size and morphology of the prepared nanoparticles were studied using scanning electron microscope. Based on the optimization made by Design Expert 7® software, a formulation made of glycerol monostearate, 1.2 % polyvinyl alcohol (PVA), weight ratio of lipid/drug of 10 and sonication time of 90 s was selected. Particle size, zeta potential, entrapment efficiency, loading percent, and mean dissolution time of adopted formulation were predicted and confirmed to be 218.2 ± 6.6 nm, -26.7 ± 1.9 mV, 92.5 ± 0.52 %, 5.8 ± 0.3 %, and 10.4 ± 0.29 h, respectively. Since the preparation and evaluation of the selected formulation within the laboratory yielded acceptable results with low error percent, the modeling and optimization was justified. The optimized formulation co-spray dried with lactose (hybrid microparticles) displayed desirable fine particle fraction, mass median aerodynamic diameter (MMAD), and geometric standard deviation of 49.5%, 2.06 μm, and 2.98 μm; respectively. Our results provide fundamental data for the
NASA Astrophysics Data System (ADS)
Mehedi, H.-A.; Baudrillart, B.; Alloyeau, D.; Mouhoub, O.; Ricolleau, C.; Pham, V. D.; Chacon, C.; Gicquel, A.; Lagoute, J.; Farhat, S.
2016-08-01
This article describes the significant roles of process parameters in the deposition of graphene films via cobalt-catalyzed decomposition of methane diluted in hydrogen using plasma-enhanced chemical vapor deposition (PECVD). The influence of growth temperature (700-850 °C), molar concentration of methane (2%-20%), growth time (30-90 s), and microwave power (300-400 W) on graphene thickness and defect density is investigated using Taguchi method which enables reaching the optimal parameter settings by performing reduced number of experiments. Growth temperature is found to be the most influential parameter in minimizing the number of graphene layers, whereas microwave power has the second largest effect on crystalline quality and minor role on thickness of graphene films. The structural properties of PECVD graphene obtained with optimized synthesis conditions are investigated with Raman spectroscopy and corroborated with atomic-scale characterization performed by high-resolution transmission electron microscopy and scanning tunneling microscopy, which reveals formation of continuous film consisting of 2-7 high quality graphene layers.
Construction of dynamic stochastic simulation models using knowledge-based techniques
NASA Technical Reports Server (NTRS)
Williams, M. Douglas; Shiva, Sajjan G.
1990-01-01
Over the past three decades, computer-based simulation models have proven themselves to be cost-effective alternatives to the more structured deterministic methods of systems analysis. During this time, many techniques, tools and languages for constructing computer-based simulation models have been developed. More recently, advances in knowledge-based system technology have led many researchers to note the similarities between knowledge-based programming and simulation technologies and to investigate the potential application of knowledge-based programming techniques to simulation modeling. The integration of conventional simulation techniques with knowledge-based programming techniques is discussed to provide a development environment for constructing knowledge-based simulation models. A comparison of the techniques used in the construction of dynamic stochastic simulation models and those used in the construction of knowledge-based systems provides the requirements for the environment. This leads to the design and implementation of a knowledge-based simulation development environment. These techniques were used in the construction of several knowledge-based simulation models including the Advanced Launch System Model (ALSYM).
Study the Effect of SiO2 Based Flux on Dilution in Submerged Arc Welding
NASA Astrophysics Data System (ADS)
kumar, Aditya; Maheshwari, Sachin
2017-08-01
This paper highlights the method for prediction of dilution in submerged arc welding (SAW). The most important factors of weld bead geometry are governed by the weld dilution which controls the chemical and mechanical properties. Submerged arc welding process is used generally due to its very easy control of process variables, good penetration, high weld quality, and smooth finish. Machining parameters, with suitable weld quality can be achieved with the different composition of the flux in the weld. In the present study Si02-Al2O3-CaO flux system was used. In SiO2 based flux NiO, MnO, MgO were mixed in various proportions. The paper investigates the relationship between the process parameters like voltage, % of flux constituents and dilution with the help of Taguchi’s method. The experiments were designed according to Taguchi L9 orthogonal array, while varying the voltage at two different levels in addition to alloying elements. Then the optimal results conditions were verified by confirmatory experiments.
NASA Astrophysics Data System (ADS)
Singh, Ramandeep
2018-04-01
The machining of materials on micro-meter and sub-micrometre is considered the technology of future. Due to challenging applications of biomedical and aerospace industries, the traditional manufacturing techniques lacks in dimensional accuracy. Thus for such industries, the technique that can control micron tolerances is Electrochemical Micromachining (EMM). Hard metals and alloys can also be machined by this technique. Thus to develop a novel EMM system setup and to investigate the effect of three different electrolytes i.e NaCl, NaNO3 and HCl with their different concentrations, the current study was conducted. Stainless Steel-304 and copper were chosen as the work piece material in the present study. Taguchi L18 orthogonal array was used for the best combination of experiment. According to the present investigation most prominent factor affecting the material removal (MR) comes out was electrolyte. HCl provides the better MR among other electrolytes i.e. NaNO3 and NaCl. The amount of MR increased with the increase in the concentration of electrolyte.
NASA Astrophysics Data System (ADS)
Hamedon, Zamzuri; Kuang, Shea Cheng; Jaafar, Hasnulhadi; Azhari, Azmir
2018-03-01
Incremental sheet forming is a versatile sheet metal forming process where a sheet metal is formed into its final shape by a series of localized deformation without a specialised die. However, it still has many shortcomings that need to be overcome such as geometric accuracy, surface roughness, formability, forming speed, and so on. This project focus on minimising the surface roughness of aluminium sheet and improving its thickness uniformity in incremental sheet forming via optimisation of wall angle, feed rate, and step size. Besides, the effect of wall angle, feed rate, and step size to the surface roughness and thickness uniformity of aluminium sheet was investigated in this project. From the results, it was observed that surface roughness and thickness uniformity were inversely varied due to the formation of surface waviness. Increase in feed rate and decrease in step size will produce a lower surface roughness, while uniform thickness reduction was obtained by reducing the wall angle and step size. By using Taguchi analysis, the optimum parameters for minimum surface roughness and uniform thickness reduction of aluminium sheet were determined. The finding of this project helps to reduce the time in optimising the surface roughness and thickness uniformity in incremental sheet forming.
Hooshmand, Sara; Es'haghi, Zarrin
2017-11-30
A number of four amino acids have been simultaneously determined at CdSe quantum dot-modified/multi-walled carbon nanotube hollow fiber pencil graphite electrode in different bodybuilding supplements. CdSe quantum dots were synthesized and applied to construct a modified carbon nanotube hollow fiber pencil graphite electrode. FT-IR, TEM, XRD and EDAX methods were applied for characterization of the synthesized CdSe QDs. The electro-oxidation of arginine (Arg), alanine (Ala), methionine (Met) and cysteine (Cys) at the surface of the modified electrode was studied. Then the Taguchi's method was applied using MINITAB 17 software to find out the optimum conditions for the amino acids determination. Under the optimized conditions, the differential pulse (DP) voltammetric peak currents of Arg, Ala, Met and Cys increased linearly with their concentrations in the ranges of 0.287-33670μM and detection limits of 0.081, 0.158, 0.094 and 0.116μM were obtained for them, respectively. Satisfactory results were achieved for calibration and validation sets. The prepared modified electrode represents a very good resolution between the voltammetric peaks of the four amino acids which makes it suitable for the detection of each in presence of others in real samples. Copyright © 2017. Published by Elsevier B.V.
The detection of bulk explosives using nuclear-based techniques
DOE Office of Scientific and Technical Information (OSTI.GOV)
Morgado, R.E.; Gozani, T.; Seher, C.C.
1988-01-01
In 1986 we presented a rationale for the detection of bulk explosives based on nuclear techniques that addressed the requirements of civil aviation security in the airport environment. Since then, efforts have intensified to implement a system based on thermal neutron activation (TNA), with new work developing in fast neutron and energetic photon reactions. In this paper we will describe these techniques and present new results from laboratory and airport testing. Based on preliminary results, we contended in our earlier paper that nuclear-based techniques did provide sufficiently penetrating probes and distinguishable detectable reaction products to achieve the FAA operational goals;more » new data have supported this contention. The status of nuclear-based techniques for the detection of bulk explosives presently under investigation by the US Federal Aviation Administration (FAA) is reviewed. These include thermal neutron activation (TNA), fast neutron activation (FNA), the associated particle technique, nuclear resonance absorption, and photoneutron activation. The results of comprehensive airport testing of the TNA system performed during 1987-88 are summarized. From a technical point of view, nuclear-based techniques now represent the most comprehensive and feasible approach for meeting the operational criteria of detection, false alarms, and throughput. 9 refs., 5 figs., 2 tabs.« less
Study on a novel laser target detection system based on software radio technique
NASA Astrophysics Data System (ADS)
Song, Song; Deng, Jia-hao; Wang, Xue-tian; Gao, Zhen; Sun, Ji; Sun, Zhi-hui
2008-12-01
This paper presents that software radio technique is applied to laser target detection system with the pseudo-random code modulation. Based on the theory of software radio, the basic framework of the system, hardware platform, and the implementation of the software system are detailed. Also, the block diagram of the system, DSP circuit, block diagram of the pseudo-random code generator, and soft flow diagram of signal processing are designed. Experimental results have shown that the application of software radio technique provides a novel method to realize the modularization, miniaturization and intelligence of the laser target detection system, and the upgrade and improvement of the system will become simpler, more convenient, and cheaper.
Laser-based direct-write techniques for cell printing
Schiele, Nathan R; Corr, David T; Huang, Yong; Raof, Nurazhani Abdul; Xie, Yubing; Chrisey, Douglas B
2016-01-01
Fabrication of cellular constructs with spatial control of cell location (±5 μm) is essential to the advancement of a wide range of applications including tissue engineering, stem cell and cancer research. Precise cell placement, especially of multiple cell types in co- or multi-cultures and in three dimensions, can enable research possibilities otherwise impossible, such as the cell-by-cell assembly of complex cellular constructs. Laser-based direct writing, a printing technique first utilized in electronics applications, has been adapted to transfer living cells and other biological materials (e.g., enzymes, proteins and bioceramics). Many different cell types have been printed using laser-based direct writing, and this technique offers significant improvements when compared to conventional cell patterning techniques. The predominance of work to date has not been in application of the technique, but rather focused on demonstrating the ability of direct writing to pattern living cells, in a spatially precise manner, while maintaining cellular viability. This paper reviews laser-based additive direct-write techniques for cell printing, and the various cell types successfully laser direct-written that have applications in tissue engineering, stem cell and cancer research are highlighted. A particular focus is paid to process dynamics modeling and process-induced cell injury during laser-based cell direct writing. PMID:20814088
ERIC Educational Resources Information Center
Seyihoglu, Aysegul; Kartal, Ayca
2010-01-01
The purpose of this study is to reveal the opinions of teachers on using the mind mapping technique in Life Science and Social Studies lessons. The participants of the study are 20 primary education teachers. In this study, a semi-structured interview technique was used. For content analysis, the themes and codes were defined, based on the views…
TOPSIS based parametric optimization of laser micro-drilling of TBC coated nickel based superalloy
NASA Astrophysics Data System (ADS)
Parthiban, K.; Duraiselvam, Muthukannan; Manivannan, R.
2018-06-01
The technique for order of preference by similarity ideal solution (TOPSIS) approach was used for optimizing the process parameters of laser micro-drilling of nickel superalloy C263 with Thermal Barrier Coating (TBC). Plasma spraying was used to deposit the TBC and a pico-second Nd:YAG pulsed laser was used to drill the specimens. Drilling angle, laser scan speed and number of passes were considered as input parameters. Based on the machining conditions, Taguchi L8 orthogonal array was used for conducting the experimental runs. The surface roughness and surface crack density (SCD) were considered as the output measures. The surface roughness was measured using 3D White Light Interferometer (WLI) and the crack density was measured using Scanning Electron Microscope (SEM). The optimized result achieved from this approach suggests reduced surface roughness and surface crack density. The holes drilled at an inclination angle of 45°, laser scan speed of 3 mm/s and 400 number of passes found to be optimum. From the Analysis of variance (ANOVA), inclination angle and number of passes were identified as the major influencing parameter. The optimized parameter combination exhibited a 19% improvement in surface finish and 12% reduction in SCD.
Developing a hybrid dictionary-based bio-entity recognition technique.
Song, Min; Yu, Hwanjo; Han, Wook-Shin
2015-01-01
Bio-entity extraction is a pivotal component for information extraction from biomedical literature. The dictionary-based bio-entity extraction is the first generation of Named Entity Recognition (NER) techniques. This paper presents a hybrid dictionary-based bio-entity extraction technique. The approach expands the bio-entity dictionary by combining different data sources and improves the recall rate through the shortest path edit distance algorithm. In addition, the proposed technique adopts text mining techniques in the merging stage of similar entities such as Part of Speech (POS) expansion, stemming, and the exploitation of the contextual cues to further improve the performance. The experimental results show that the proposed technique achieves the best or at least equivalent performance among compared techniques, GENIA, MESH, UMLS, and combinations of these three resources in F-measure. The results imply that the performance of dictionary-based extraction techniques is largely influenced by information resources used to build the dictionary. In addition, the edit distance algorithm shows steady performance with three different dictionaries in precision whereas the context-only technique achieves a high-end performance with three difference dictionaries in recall.
Developing a hybrid dictionary-based bio-entity recognition technique
2015-01-01
Background Bio-entity extraction is a pivotal component for information extraction from biomedical literature. The dictionary-based bio-entity extraction is the first generation of Named Entity Recognition (NER) techniques. Methods This paper presents a hybrid dictionary-based bio-entity extraction technique. The approach expands the bio-entity dictionary by combining different data sources and improves the recall rate through the shortest path edit distance algorithm. In addition, the proposed technique adopts text mining techniques in the merging stage of similar entities such as Part of Speech (POS) expansion, stemming, and the exploitation of the contextual cues to further improve the performance. Results The experimental results show that the proposed technique achieves the best or at least equivalent performance among compared techniques, GENIA, MESH, UMLS, and combinations of these three resources in F-measure. Conclusions The results imply that the performance of dictionary-based extraction techniques is largely influenced by information resources used to build the dictionary. In addition, the edit distance algorithm shows steady performance with three different dictionaries in precision whereas the context-only technique achieves a high-end performance with three difference dictionaries in recall. PMID:26043907
Experimental Study in Taguchi Method on Surface Quality Predication of HSM
NASA Astrophysics Data System (ADS)
Ji, Yan; Li, Yueen
2018-05-01
Based on the study of ball milling mechanism and machining surface formation mechanism, the formation of high speed ball-end milling surface is a time-varying and cumulative Thermos-mechanical coupling process. The nature of this problem is that the uneven stress field and temperature field affect the machined surface Process, the performance of the processing parameters in the processing interaction in the elastic-plastic materials produced by the elastic recovery and plastic deformation. The surface quality of machining surface is characterized by multivariable nonlinear system. It is still an indispensable and effective method to study the surface quality of high speed ball milling by experiments.
NASA Astrophysics Data System (ADS)
Giri Prasad, M. J.; Abhishek Raaj, A. S.; Rishi Kumar, R.; Gladson, Frank; M, Gautham
2016-09-01
The present study is concerned with resolving the problems pertaining to the conventional cutting fluids. Two samples of nano cutting fluids were prepared by dispersing 0.01 vol% of MWCNTs and a mixture of 0.01 vol% of MWCNTs and 0.01 vol% of nano ZnO in the soluble oil. The thermophysical properties such as the kinematic viscosity, density, flash point and the tribological properties of the prepared nano cutting fluid samples were experimentally investigated and were compared with those of plain soluble oil. In addition to this, a milling process was carried by varying the process parameters and by application of different samples of cutting fluids and an attempt was made to determine optimal cutting condition using the Taguchi optimization technique.
Hardcastle, Sarah J; Fortier, Michelle; Blake, Nicola; Hagger, Martin S
2017-03-01
Motivational interviewing (MI) is a complex intervention comprising multiple techniques aimed at changing health-related motivation and behaviour. However, MI techniques have not been systematically isolated and classified. This study aimed to identify the techniques unique to MI, classify them as content-related or relational, and evaluate the extent to which they overlap with techniques from the behaviour change technique taxonomy version 1 [BCTTv1; Michie, S., Richardson, M., Johnston, M., Abraham, C., Francis, J., Hardeman, W., … Wood, C. E. (2013). The behavior change technique taxonomy (v1) of 93 hierarchically clustered techniques: Building an international consensus for the reporting of behavior change interventions. Annals of Behavioral Medicine, 46, 81-95]. Behaviour change experts (n = 3) content-analysed MI techniques based on Miller and Rollnick's [(2013). Motivational interviewing: Preparing people for change (3rd ed.). New York: Guildford Press] conceptualisation. Each technique was then coded for independence and uniqueness by independent experts (n = 10). The experts also compared each MI technique to those from the BCTTv1. Experts identified 38 distinct MI techniques with high agreement on clarity, uniqueness, preciseness, and distinctiveness ratings. Of the identified techniques, 16 were classified as relational techniques. The remaining 22 techniques were classified as content based. Sixteen of the MI techniques were identified as having substantial overlap with techniques from the BCTTv1. The isolation and classification of MI techniques will provide researchers with the necessary tools to clearly specify MI interventions and test the main and interactive effects of the techniques on health behaviour. The distinction between relational and content-based techniques within MI is also an important advance, recognising that changes in motivation and behaviour in MI is a function of both intervention content and the interpersonal style
Spectroscopic techniques to study the immune response in human saliva
NASA Astrophysics Data System (ADS)
Nepomnyashchaya, E.; Savchenko, E.; Velichko, E.; Bogomaz, T.; Aksenov, E.
2018-01-01
Studies of the immune response dynamics by means of spectroscopic techniques, i.e., laser correlation spectroscopy and fluorescence spectroscopy, are described. The laser correlation spectroscopy is aimed at measuring sizes of particles in biological fluids. The fluorescence spectroscopy allows studying of the conformational and other structural changings in immune complex. We have developed a new scheme of a laser correlation spectrometer and an original signal processing algorithm. We have suggested a new fluorescence detection scheme based on a prism and an integrating pin diode. The developed system based on the spectroscopic techniques allows studies of complex process in human saliva and opens some prospects for an individual treatment of immune diseases.
Problem based learning with scaffolding technique on geometry
NASA Astrophysics Data System (ADS)
Bayuningsih, A. S.; Usodo, B.; Subanti, S.
2018-05-01
Geometry as one of the branches of mathematics has an important role in the study of mathematics. This research aims to explore the effectiveness of Problem Based Learning (PBL) with scaffolding technique viewed from self-regulation learning toward students’ achievement learning in mathematics. The research data obtained through mathematics learning achievement test and self-regulated learning (SRL) questionnaire. This research employed quasi-experimental research. The subjects of this research are students of the junior high school in Banyumas Central Java. The result of the research showed that problem-based learning model with scaffolding technique is more effective to generate students’ mathematics learning achievement than direct learning (DL). This is because in PBL model students are more able to think actively and creatively. The high SRL category student has better mathematic learning achievement than middle and low SRL categories, and then the middle SRL category has better than low SRL category. So, there are interactions between learning model with self-regulated learning in increasing mathematic learning achievement.
Molecular-Based Optical Measurement Techniques for Transition and Turbulence in High-Speed Flow
NASA Technical Reports Server (NTRS)
Bathel, Brett F.; Danehy, Paul M.; Cutler, Andrew D.
2013-01-01
High-speed laminar-to-turbulent transition and turbulence affect the control of flight vehicles, the heat transfer rate to a flight vehicle's surface, the material selected to protect such vehicles from high heating loads, the ultimate weight of a flight vehicle due to the presence of thermal protection systems, the efficiency of fuel-air mixing processes in high-speed combustion applications, etc. Gaining a fundamental understanding of the physical mechanisms involved in the transition process will lead to the development of predictive capabilities that can identify transition location and its impact on parameters like surface heating. Currently, there is no general theory that can completely describe the transition-to-turbulence process. However, transition research has led to the identification of the predominant pathways by which this process occurs. For a truly physics-based model of transition to be developed, the individual stages in the paths leading to the onset of fully turbulent flow must be well understood. This requires that each pathway be computationally modeled and experimentally characterized and validated. This may also lead to the discovery of new physical pathways. This document is intended to describe molecular based measurement techniques that have been developed, addressing the needs of the high-speed transition-to-turbulence and high-speed turbulence research fields. In particular, we focus on techniques that have either been used to study high speed transition and turbulence or techniques that show promise for studying these flows. This review is not exhaustive. In addition to the probe-based techniques described in the previous paragraph, several other classes of measurement techniques that are, or could be, used to study high speed transition and turbulence are excluded from this manuscript. For example, surface measurement techniques such as pressure and temperature paint, phosphor thermography, skin friction measurements and
Wang, Zhenhong; Luo, Zhuanxi; Yan, Changzhou; Xing, Baoshan
2017-07-01
Very limited information is available on how and to what extent environmental factors influence arsenic (As) biotransformation and release in freshwater algae. These factors include concentrations of arsenate (As(V)), dissolved inorganic nitrogen (N), phosphate (P), and ambient pH. This study conducted a series of experiments using Taguchi methods to determine optimum conditions for As biotransformation. We assessed principal effective factors of As(V), N, P, and pH and determined that As biotransformation and release actuate at 10.0 μM As(V) in dead alga cells, the As efflux ratio and organic As efflux content actuate at 1.0 mg/L P, algal growth and intracellular arsenite (As(III)) content actuate at 10.0 mg/L N, and the total sum of As(III) efflux from dead alga cells actuates at a pH level of 10. Moreover, N is the critical component for As(V) biotransformation in M. aeruginosa, specifically for As(III) transformation, because N can accelerate algal growth, subsequently improving As(III) accumulation and its efflux, which results in an As(V) to As(III) reduction. Furthermore, low P concentrations in combination with high N concentrations promote As accumulation. Following As(V), P was the primary impacting factor for As accumulation. In addition, small amounts of As accumulation under low concentrations of As and high P were securely stored in living algal cells and were easily released after cell death. Results from this study will help to assess practical applications and the overall control of key environmental factors, particularly those associated with algal bioremediation in As polluted water. Copyright © 2017 Elsevier Ltd. All rights reserved.
Neutron-based nonintrusive inspection techniques
NASA Astrophysics Data System (ADS)
Gozani, Tsahi
1997-02-01
Non-intrusive inspection of large objects such as trucks, sea-going shipping containers, air cargo containers and pallets is gaining attention as a vital tool in combating terrorism, drug smuggling and other violation of international and national transportation and Customs laws. Neutrons are the preferred probing radiation when material specificity is required, which is most often the case. Great strides have been made in neutron based inspection techniques. Fast and thermal neutrons, whether in steady state or in microsecond, or even nanosecond pulses are being employed to interrogate, at high speeds, for explosives, drugs, chemical agents, and nuclear and many other smuggled materials. Existing neutron techniques will be compared and their current status reported.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mehedi, H.-A.; Baudrillart, B.; Gicquel, A.
2016-08-14
This article describes the significant roles of process parameters in the deposition of graphene films via cobalt-catalyzed decomposition of methane diluted in hydrogen using plasma-enhanced chemical vapor deposition (PECVD). The influence of growth temperature (700–850 °C), molar concentration of methane (2%–20%), growth time (30–90 s), and microwave power (300–400 W) on graphene thickness and defect density is investigated using Taguchi method which enables reaching the optimal parameter settings by performing reduced number of experiments. Growth temperature is found to be the most influential parameter in minimizing the number of graphene layers, whereas microwave power has the second largest effect on crystalline qualitymore » and minor role on thickness of graphene films. The structural properties of PECVD graphene obtained with optimized synthesis conditions are investigated with Raman spectroscopy and corroborated with atomic-scale characterization performed by high-resolution transmission electron microscopy and scanning tunneling microscopy, which reveals formation of continuous film consisting of 2–7 high quality graphene layers.« less
A study on laser-based ultrasonic technique by the use of guided wave tomographic imaging
DOE Office of Scientific and Technical Information (OSTI.GOV)
Park, Junpil, E-mail: jpp@pusan.ac.kr; Lim, Juyoung, E-mail: jpp@pusan.ac.kr; Cho, Younho
2015-03-31
Guided wave tests are impractical for investigating specimens with limited accessibility and coarse surfaces or geometrically complicated features. A non-contact setup with a laser ultrasonic transmitter and receiver is the classic attractive for guided wave inspection. The present work was done to develop a non-contact guided-wave tomography technique by laser ultrasonic technique in a plate-like structure. A method for Lam wave generation and detection in an aluminum plate with a pulse laser ultrasonic transmitter and a Michelson interferometer receiver has been developed. In the images obtained by laser scanning, the defect shape and area showed good agreement with the actualmore » defect. The proposed approach can be used as a non-contact-based online inspection and monitoring technique.« less
Ultrabroadband Phased-Array Receivers Based on Optical Techniques
2016-02-26
AFRL-AFOSR-VA-TR-2016-0121 Ultrabroadband Phased- array Receivers Based on Optical Techniques Christopher Schuetz UNIVERSITY OF DELAWARE Final Report...Jul 15 4. TITLE AND SUBTITLE Ultrabroadband Phased- Array Receivers Based on Optical Techniques 5a. CONTRACT NUMBER 5b. GRANT NUMBER FA9550-12-1...receiver that enables us to capture and convert signals across an array using photonic modulators, routing these signals to a central location using
NASA Technical Reports Server (NTRS)
Ricks, Wendell R.; Abbott, Kathy H.
1987-01-01
To the software design community, the concern over the costs associated with a program's execution time and implementation is great. It is always desirable, and sometimes imperative, that the proper programming technique is chosen which minimizes all costs for a given application or type of application. A study is described that compared cost-related factors associated with traditional programming techniques to rule-based programming techniques for a specific application. The results of this study favored the traditional approach regarding execution efficiency, but favored the rule-based approach regarding programmer productivity (implementation ease). Although this study examined a specific application, the results should be widely applicable.
Bandwidth-Tunable Fiber Bragg Gratings Based on UV Glue Technique
NASA Astrophysics Data System (ADS)
Fu, Ming-Yue; Liu, Wen-Feng; Chen, Hsin-Tsang; Chuang, Chia-Wei; Bor, Sheau-Shong; Tien, Chuen-Lin
2007-07-01
In this study, we have demonstrated that a uniform fiber Bragg grating (FBG) can be transformed into a chirped fiber grating by a simple UV glue adhesive technique without shifting the reflection band with respect to the center wavelength of the FBG. The technique is based on the induced strain of an FBG due to the UV glue adhesive force on the fiber surface that causes a grating period variation and an effective index change. This technique can provide a fast and simple method of obtaining the required chirp value of a grating for applications in the dispersion compensators, gain flattening in erbium-doped fiber amplifiers (EDFAs) or optical filters.
Evaluation of a reproduction technique for the study of the enamel composite/bracket base area.
Wilner, F J; Oliver, R G
2000-09-01
The objective of the study was to evaluate a reproduction method that would enable the study of the enamel/ bracket/composite interface in vivo, and consisted of in vitro assessment of two different impression materials to compare reproduction of brackets bonded to extracted teeth followed by in vivo assessment of the superior material. In vitro standard edgewise brackets were bonded to two extracted teeth and impressions were taken using two different types of low viscosity silicone-based impression materials. A medium viscosity silicone impression material was used to support the original impression. Three impressions of both the gingival and occlusal aspect of the bracket base region were obtained using each of the impression materials. Replicas were then prepared for SEM viewing and these compared to SEMs of the real teeth for reproduction of detail. A 3-point Reproducibility Index was used to compare the SEM photographs of the comparable replicas. One impression material was clearly superior to the other and produced an acceptably accurate representation of the true clinical situation in three out of four samples. This material also performed well in the in vivo situation. The technique described is satisfactory for the production and analysis of SEM pictures of the enamel/composite/ bracket base interface in vivo.
NASA Astrophysics Data System (ADS)
Ginghtong, Thatchanok; Nakpathomkun, Natthapon; Pechyen, Chiravoot
2018-06-01
The parameters of the plastic injection molding process have been investigated for the manufacture of a 64 oz. ultra-thin polypropylene bucket. The 3 main parameters, such as injection speed, melting temperature, holding pressure, were investigated to study their effect on the physical appearance and compressive strength. The orthogonal array of Taguchi's L9 (33) was used to carry out the experimental plan. The physical properties were measured and the compressive strength was determined using linear regression analysis. The differential scanning calorimeter (DSC) was used to analyze the crystalline structure of the product. The optimization results show that the proposed approach can help engineers identify optimal process parameters and achieve competitive advantages of energy consumption and product quality. In addition, the injection molding of the product includes 24 mm of shot stroke, 1.47 mm position transfer, 268 rpm screw speed, injection speed 100 mm/s, 172 ton clamping force, 800 kgf holding pressure, 0.9 s holding time and 1.4 s cooling time, make the products in the shape and proportion of the product satisfactory. The parameters of influence are injection speed 71.07%, melting temperature 23.31% and holding pressure 5.62%, respectively. The compressive strength of the product was able to withstand a pressure of up to 839 N before the product became plastic. The low melting temperature was caused by the superior crystalline structure of the super-ultra-thin wall product which leads to a lower compressive strength.
Study of systems and techniques for data base management
NASA Technical Reports Server (NTRS)
1976-01-01
Data management areas were studied to identify pertinent problems and issues that will affect future NASA data users in terms of performance and cost. Specific topics discussed include the identifications of potential NASA data users other than those normally discussed, consideration affecting the clustering of minicomputers, low cost computer system for information retrieval and analysis, the testing of minicomputer based data base management systems, ongoing work related to the use of dedicated systems for data base management, and the problems of data interchange among a community of NASA data users.
Quantification of dental prostheses on cone‐beam CT images by the Taguchi method
Kuo, Rong‐Fu; Fang, Kwang‐Ming; TY, Wong
2016-01-01
The gray values accuracy of dental cone‐beam computed tomography (CBCT) is affected by dental metal prostheses. The distortion of dental CBCT gray values could lead to inaccuracies of orthodontic and implant treatment. The aim of this study was to quantify the effect of scanning parameters and dental metal prostheses on the accuracy of dental cone‐beam computed tomography (CBCT) gray values using the Taguchi method. Eight dental model casts of an upper jaw including prostheses, and a ninth prosthesis‐free dental model cast, were scanned by two dental CBCT devices. The mean gray value of the selected circular regions of interest (ROIs) were measured using dental CBCT images of eight dental model casts and were compared with those measured from CBCT images of the prosthesis‐free dental model cast. For each image set, four consecutive slices of gingiva were selected. The seven factors (CBCTs, occlusal plane canting, implant connection, prosthesis position, coping material, coping thickness, and types of dental restoration) were used to evaluate scanning parameter and dental prostheses effects. Statistical methods of signal to noise ratio (S/N) and analysis of variance (ANOVA) with 95% confidence were applied to quantify the effects of scanning parameters and dental prostheses on dental CBCT gray values accuracy. For ROIs surrounding dental prostheses, the accuracy of CBCT gray values were affected primarily by implant connection (42%), followed by type of restoration (29%), prostheses position (19%), coping material (4%), and coping thickness (4%). For a single crown prosthesis (without support of implants) placed in dental model casts, gray value differences for ROIs 1–9 were below 12% and gray value differences for ROIs 13–18 away from prostheses were below 10%. We found the gray value differences set to be between 7% and 8% for regions next to a single implant‐supported titanium prosthesis, and between 46% and 59% for regions between double implant
A technique based on droplet evaporation to recognize alcoholic drinks
NASA Astrophysics Data System (ADS)
González-Gutiérrez, Jorge; Pérez-Isidoro, Rosendo; Ruiz-Suárez, J. C.
2017-07-01
Chromatography is, at present, the most used technique to determine the purity of alcoholic drinks. This involves a careful separation of the components of the liquid elements. However, since this technique requires sophisticated instrumentation, there are alternative techniques such as conductivity measurements and UV-Vis and infrared spectrometries. We report here a method based on salt-induced crystallization patterns formed during the evaporation of alcoholic drops. We found that droplets of different samples form different structures upon drying, which we characterize by their radial density profiles. We prove that using the dried deposit of a spirit as a control sample, our method allows us to differentiate between pure and adulterated drinks. As a proof of concept, we study tequila.
Observation-Driven Configuration of Complex Software Systems
NASA Astrophysics Data System (ADS)
Sage, Aled
2010-06-01
The ever-increasing complexity of software systems makes them hard to comprehend, predict and tune due to emergent properties and non-deterministic behaviour. Complexity arises from the size of software systems and the wide variety of possible operating environments: the increasing choice of platforms and communication policies leads to ever more complex performance characteristics. In addition, software systems exhibit different behaviour under different workloads. Many software systems are designed to be configurable so that policies can be chosen to meet the needs of various stakeholders. For complex software systems it can be difficult to accurately predict the effects of a change and to know which configuration is most appropriate. This thesis demonstrates that it is useful to run automated experiments that measure a selection of system configurations. Experiments can find configurations that meet the stakeholders' needs, find interesting behavioural characteristics, and help produce predictive models of the system's behaviour. The design and use of ACT (Automated Configuration Tool) for running such experiments is described, in combination a number of search strategies for deciding on the configurations to measure. Design Of Experiments (DOE) is discussed, with emphasis on Taguchi Methods. These statistical methods have been used extensively in manufacturing, but have not previously been used for configuring software systems. The novel contribution here is an industrial case study, applying the combination of ACT and Taguchi Methods to DC-Directory, a product from Data Connection Ltd (DCL). The case study investigated the applicability of Taguchi Methods for configuring complex software systems. Taguchi Methods were found to be useful for modelling and configuring DC- Directory, making them a valuable addition to the techniques available to system administrators and developers.
Potential advantages of using synchrotron X-ray based techniques in pediatric research.
Pascolo, L; Esteve, F; Rizzardi, C; James, S; Menk, R H
2013-01-01
Synchrotron radiation (SR), which combines extremely high intensity, high collimation, tunability, and continuous energy spectrum, allows the development of advanced X-ray based techniques that are becoming a uniquely useful tool in life science research, along providing exciting opportunities in biomedical imaging and radiotherapy. This review summarize emerging techniques and their potential to greatly enhance the exploration of dynamical biological process occurring across various spatial and temporal regimes, from whole body physiology, down to the location of individual chemical species within single cells. In recent years pediatric research and clinic practice have started to profit from these new opportunities, particularly by extending the diagnostic and therapeutic capabilities of these X-ray based techniques. In diagnosis, technical advances in DEI and KES imaging modalities have been demonstrated as particularly valuable for children and women since SR allows dose minimization, with significant reductions compared to conventional approaches. However, the greatest expectations are in the field of SR based radiotherapy, increasingly studies are demonstrating SR radiotherapy provides improved chances of recovery; this is especially the case for pediatric patients. In addition, we report on the applicability of advanced X-ray microscopy techniques that offer exceptional spatial and quantitative resolution in elemental detection. These techniques, which are useful for in vitro studies, will be particularly advantageous where investigators seek deeper understanding of diseases where mismetabolism of metals, either physiological important (i.e. Cu, Zn) or outright toxic (i.e. Pb), underlies pathogenesis.
NASA Astrophysics Data System (ADS)
Mizamzul Mehat, Nik; Syuhada Zakarria, Noor; Kamaruddin, Shahrul
2018-03-01
The increase in demand for industrial gears has resulted in the increase in usage of plastic-matrix composites particularly glass fibre-reinforced plastics as the gear materials. The usage of these synthetic fibers is to enhance the mechanical strength and the thermal resistance of the plastic gears. Nevertheless, the production of large quantities of these synthetic fibre-reinforced composites poses a serious threat to the ecosystem. Comprehending to this fact, the present work aimed at investigating the effects of incorporating recycled glass fibre-reinforced plastics in various compositions particularly on dimensional stability and mechanical properties of gear produced with diverse injection moulding processing parameters setting. The integration of Grey relational analysis (GRA) and Taguchi method was adopted to evaluate the influence of recycled glass fibre-reinforced plastics and variation in processing parameters on gear quality. From the experimental results, the blending ratio was found as the most influential parameter of 56.0% contribution in both improving tensile properties as well as in minimizing shrinkage, followed by mould temperature of 24.1% contribution and cooling time of 10.6% contribution. The results obtained from the aforementioned work are expected to contribute to accessing the feasibility of using recycled glass fibre-reinforced plastics especially for gear application.
Wagemaker, Tais A L; Maia Campos, Patrícia M B G; Shimizu, Kenji; Kyotani, Daiki; Yoshida, Daisuke
2017-08-01
Cutaneous irritants exposure induces an excess of ROS in the skin and can ensue an inflammatory response. Topical antioxidant-based formulations can help to counteract ROS generation. This study evaluated the influence of antioxidant-based topical formulations on the inflammatory response of skin, using a combination of in vivo real-time non-invasive techniques. Nine test areas were defined on each volar forearm of the 25 Japanese volunteers. Measurements were performed before and after treatment with 15μL of a 5% sodium dodecyl sulfate solution and 15μL of the same based formulation or the vehicle with 1% of the antioxidants. Volunteers without antioxidant treatment showed more pronounced erythematous areas. Transepidermal water loss of areas treated with green tea polyphenol (GTP)-based formulation showed fully recovered skin. Skin barrier damage caused by repeated applications of SDS showed characteristic alterations, detectable by in vivo confocal microscopy such as desquamation, spongiosis and inflammatory infiltrates. The majority of confocal microscopy inflammation signs were found in skin without treatment followed by the vehicle. Ascorbyl tetraisopalmitate, Coenzyme Q 10 , GTP- and Resveratrol-based formulations reduced the anti-inflammatory cytokines release and attenuated inflammatory signs. The combination of techniques provides results that highlight the importance of antioxidant-based formulations for rapid skin recovery. Copyright © 2017 Elsevier B.V. All rights reserved.
Investigation of laser Doppler anemometry in developing a velocity-based measurement technique
NASA Astrophysics Data System (ADS)
Jung, Ki Won
2009-12-01
locations. Third, the power dissipated in a variable RLC load is measured. The three experiments validate the LDA technique proposed. The utility of the LDA method is then extended to the measurement of the complex propagation constant of the air inside a 100 ppi reticulated vitreous carbon (RVC) sample. Compared to measurements in the available studies, the measurement with the 100 ppi RVC sample supports the LDA technique in that it can achieve a low uncertainty in the determined quantity. This dissertation concludes with using the LDA technique for modal decomposition of the plane wave mode and the (1,1) mode that are driven simultaneously. This modal decomposition suggests that the LDA technique surpasses microphone-based techniques, because they are unable to determine the acoustic field based on an acoustic model with unconfined propagation constants for each modal component.
Sharon, Maheshwar; Apte, P R; Purandare, S C; Zacharia, Renju
2005-02-01
Seven variable parameters of the chemical vapor deposition system have been optimized with the help of the Taguchi analytical method for getting a desired product, e.g., carbon nanotubes or carbon nanobeads. It is observed that almost all selected parameters influence the growth of carbon nanotubes. However, among them, the nature of precursor (racemic, R or Technical grade camphor) and the carrier gas (hydrogen, argon and mixture of argon/hydrogen) seem to be more important parameters affecting the growth of carbon nanotubes. Whereas, for the growth of nanobeads, out of seven parameters, only two, i.e., catalyst (powder of iron, cobalt, and nickel) and temperature (1023 K, 1123 K, and 1273 K), are the most influential parameters. Systematic defects or islands on the substrate surface enhance nucleation of novel carbon materials. Quantitative contributions of process parameters as well as optimum factor levels are obtained by performing analysis of variance (ANOVA) and analysis of mean (ANOM), respectively.
Alexandre Teixeira, César; Direito, Bruno; Bandarabadi, Mojtaba; Le Van Quyen, Michel; Valderrama, Mario; Schelter, Bjoern; Schulze-Bonhage, Andreas; Navarro, Vincent; Sales, Francisco; Dourado, António
2014-05-01
The ability of computational intelligence methods to predict epileptic seizures is evaluated in long-term EEG recordings of 278 patients suffering from pharmaco-resistant partial epilepsy, also known as refractory epilepsy. This extensive study in seizure prediction considers the 278 patients from the European Epilepsy Database, collected in three epilepsy centres: Hôpital Pitié-là-Salpêtrière, Paris, France; Universitätsklinikum Freiburg, Germany; Centro Hospitalar e Universitário de Coimbra, Portugal. For a considerable number of patients it was possible to find a patient specific predictor with an acceptable performance, as for example predictors that anticipate at least half of the seizures with a rate of false alarms of no more than 1 in 6 h (0.15 h⁻¹). We observed that the epileptic focus localization, data sampling frequency, testing duration, number of seizures in testing, type of machine learning, and preictal time influence significantly the prediction performance. The results allow to face optimistically the feasibility of a patient specific prospective alarming system, based on machine learning techniques by considering the combination of several univariate (single-channel) electroencephalogram features. We envisage that this work will serve as benchmark data that will be of valuable importance for future studies based on the European Epilepsy Database. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
Dimensional changes of acrylic resin denture bases: conventional versus injection-molding technique.
Gharechahi, Jafar; Asadzadeh, Nafiseh; Shahabian, Foad; Gharechahi, Maryam
2014-07-01
Acrylic resin denture bases undergo dimensional changes during polymerization. Injection molding techniques are reported to reduce these changes and thereby improve physical properties of denture bases. The aim of this study was to compare dimensional changes of specimens processed by conventional and injection-molding techniques. SR-Ivocap Triplex Hot resin was used for conventional pressure-packed and SR-Ivocap High Impact was used for injection-molding techniques. After processing, all the specimens were stored in distilled water at room temperature until measured. For dimensional accuracy evaluation, measurements were recorded at 24-hour, 48-hour and 12-day intervals using a digital caliper with an accuracy of 0.01 mm. Statistical analysis was carried out by SPSS (SPSS Inc., Chicago, IL, USA) using t-test and repeated-measures ANOVA. Statistical significance was defined at P<0.05. After each water storage period, the acrylic specimens produced by injection exhibited less dimensional changes compared to those produced by the conventional technique. Curing shrinkage was compensated by water sorption with an increase in water storage time decreasing dimensional changes. Within the limitations of this study, dimensional changes of acrylic resin specimens were influenced by the molding technique used and SR-Ivocap injection procedure exhibited higher dimensional accuracy compared to conventional molding.
Development of a Pulsed Pressure-Based Technique for Cavitation Damage Study
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ren, Fei; Wang, Jy-An John; Liu, Yun
2012-01-01
Cavitation occurs in many fluid systems and can lead to severe material damage. To assist the study of cavitation damage, a novel testing method utilizing pulsed pressure was developed. In this talk, the scientific background and the technical approach of this development are present and preliminary testing results are discussed. It is expected that this technique can be used to evaluate cavitation damage under various testing conditions including harsh environments such as those relevant to geothermal power generation.
A Study on the Characteristics of Design Variables for IRSS Diffuser
NASA Astrophysics Data System (ADS)
Cho, Yong-Jin; Ko, Dae-Eun
2017-11-01
In modern naval ships, infrared signature suppression systems (IRSS) are installed to decrease the temperature of waste gas generated in propulsion engine and the metallic surface temperature of heated exhaust pipes. Generally, IRSS is composed of eductor, mixing tube, and diffuser. Diffuser serves to reduce the temperature by creating an air film using the pressure difference between internal gas and external air. In this study, design variables were selected by analyzing the diffuser and the characteristics of design variables that affect the performance of diffuser were examined using Taguchi experiment method. For the diffuser performance analysis, a heat flow analysis technique established in previous research was used. The IRSS performance evaluation was carried out based on the average area value of the metal surface temperature and the temperature of the exhaust gas at the outlet of the diffuser, which are variables directly related to the intensity of infrared signature in naval ships. It was verified that the exhaust gas temperature is greatly affected by changes in the diameter of the diffuser outlet, and the metal surface temperature of diffuser is greatly affected by changes in the number of diffuser rings.
Noninvasive in vivo glucose sensing using an iris based technique
NASA Astrophysics Data System (ADS)
Webb, Anthony J.; Cameron, Brent D.
2011-03-01
Physiological glucose monitoring is important aspect in the treatment of individuals afflicted with diabetes mellitus. Although invasive techniques for glucose monitoring are widely available, it would be very beneficial to make such measurements in a noninvasive manner. In this study, a New Zealand White (NZW) rabbit animal model was utilized to evaluate a developed iris-based imaging technique for the in vivo measurement of physiological glucose concentration. The animals were anesthetized with isoflurane and an insulin/dextrose protocol was used to control blood glucose concentration. To further help restrict eye movement, a developed ocular fixation device was used. During the experimental time frame, near infrared illuminated iris images were acquired along with corresponding discrete blood glucose measurements taken with a handheld glucometer. Calibration was performed using an image based Partial Least Squares (PLS) technique. Independent validation was also performed to assess model performance along with Clarke Error Grid Analysis (CEGA). Initial validation results were promising and show that a high percentage of the predicted glucose concentrations are within 20% of the reference values.
A study of trends and techniques for space base electronics
NASA Technical Reports Server (NTRS)
Trotter, J. D.; Wade, T. E.; Gassaway, J. D.
1978-01-01
Furnaces and photolithography related equipment were applied to experiments on double layer metal. The double layer metal activity emphasized wet chemistry techniques. By incorporating the following techniques: (1) ultrasonic etching of the vias; (2) premetal clean using a modified buffered hydrogen fluoride; (3) phosphorus doped vapor; and (4) extended sintering, yields of 98 percent were obtained using the standard test pattern. The two dimensional modeling problems have stemmed from, alternately, instability and too much computation time to achieve convergence.
An image registration-based technique for noninvasive vascular elastography
NASA Astrophysics Data System (ADS)
Valizadeh, Sina; Makkiabadi, Bahador; Mirbagheri, Alireza; Soozande, Mehdi; Manwar, Rayyan; Mozaffarzadeh, Moein; Nasiriavanaki, Mohammadreza
2018-02-01
Non-invasive vascular elastography is an emerging technique in vascular tissue imaging. During the past decades, several techniques have been suggested to estimate the tissue elasticity by measuring the displacement of the Carotid vessel wall. Cross correlation-based methods are the most prevalent approaches to measure the strain exerted in the wall vessel by the blood pressure. In the case of a low pressure, the displacement is too small to be apparent in ultrasound imaging, especially in the regions far from the center of the vessel, causing a high error of displacement measurement. On the other hand, increasing the compression leads to a relatively large displacement in the regions near the center, which reduces the performance of the cross correlation-based methods. In this study, a non-rigid image registration-based technique is proposed to measure the tissue displacement for a relatively large compression. The results show that the error of the displacement measurement obtained by the proposed method is reduced by increasing the amount of compression while the error of the cross correlationbased method rises for a relatively large compression. We also used the synthetic aperture imaging method, benefiting the directivity diagram, to improve the image quality, especially in the superficial regions. The best relative root-mean-square error (RMSE) of the proposed method and the adaptive cross correlation method were 4.5% and 6%, respectively. Consequently, the proposed algorithm outperforms the conventional method and reduces the relative RMSE by 25%.
Comparison of denture base adaptation between CAD-CAM and conventional fabrication techniques.
Goodacre, Brian J; Goodacre, Charles J; Baba, Nadim Z; Kattadiyil, Mathew T
2016-08-01
Currently no data comparing the denture base adaptation of CAD-CAM and conventional denture processing techniques have been reported. The purpose of this in vitro study was to compare the denture base adaptation of pack and press, pour, injection, and CAD-CAM techniques for fabricating dentures to determine which process produces the most accurate and reproducible adaptation. A definitive cast was duplicated to create 40 gypsum casts that were laser scanned before any fabrication procedures were initiated. A master denture was made using the CAD-CAM process and was then used to create a putty mold for the fabrication of 30 standardized wax festooned dentures, 10 for each of the conventional processing techniques (pack and press, pour, injection). Scan files from 10 casts were sent to Global Dental Science, LLC for fabrication of the CAD-CAM test specimens. After specimens for each of the 4 techniques had been fabricated, they were hydrated for 24 hours and the intaglio surface laser scanned. The scan file of each denture was superimposed on the scan file of the corresponding preprocessing cast using surface matching software. Measurements were made at 60 locations, providing evaluation of fit discrepancies at the following areas: apex of the denture border, 6 mm from the denture border, crest of the ridge, palate, and posterior palatal seal. The use of median and interquartile range was used to assess accuracy and reproducibility. The Levine and Kruskal-Wallis analysis of variance was used to evaluate differences between processing techniques at the 5 specified locations (α=.05). The ranking of results based on median and interquartile range determined that the accuracy and reproducibility of the CAD-CAM technique was more consistently localized around zero at 3 of the 5 locations. Therefore, the CAD-CAM technique showed the best combination of accuracy and reproducibility among the tested fabrication techniques. The pack and press technique was more accurate at
Intramuscular injection technique: an evidence-based approach.
Ogston-Tuck, Sherri
2014-09-30
Intramuscular injections require a thorough and meticulous approach to patient assessment and injection technique. This article, the second in a series of two, reviews the evidence base to inform safer practice and to consider the evidence for nursing practice in this area. A framework for safe practice is included, identifying important points for safe technique, patient care and clinical decision making. It also highlights the ongoing debate in selection of intramuscular injection sites, predominately the ventrogluteal and dorsogluteal muscles.
NASA Astrophysics Data System (ADS)
Zhang, Huibin; Wang, Yuqiao; Chen, Haoran; Zhao, Yongli; Zhang, Jie
2017-12-01
In software defined optical networks (SDON), the centralized control plane may encounter numerous intrusion threatens which compromise the security level of provisioned services. In this paper, the issue of control plane security is studied and two machine-learning-based control plane intrusion detection techniques are proposed for SDON with properly selected features such as bandwidth, route length, etc. We validate the feasibility and efficiency of the proposed techniques by simulations. Results show an accuracy of 83% for intrusion detection can be achieved with the proposed machine-learning-based control plane intrusion detection techniques.
Simulation verification techniques study: Simulation self test hardware design and techniques report
NASA Technical Reports Server (NTRS)
1974-01-01
The final results are presented of the hardware verification task. The basic objectives of the various subtasks are reviewed along with the ground rules under which the overall task was conducted and which impacted the approach taken in deriving techniques for hardware self test. The results of the first subtask and the definition of simulation hardware are presented. The hardware definition is based primarily on a brief review of the simulator configurations anticipated for the shuttle training program. The results of the survey of current self test techniques are presented. The data sources that were considered in the search for current techniques are reviewed, and results of the survey are presented in terms of the specific types of tests that are of interest for training simulator applications. Specifically, these types of tests are readiness tests, fault isolation tests and incipient fault detection techniques. The most applicable techniques were structured into software flows that are then referenced in discussions of techniques for specific subsystems.
Empirical study on neural network based predictive techniques for automatic number plate recognition
NASA Astrophysics Data System (ADS)
Shashidhara, M. S.; Indrakumar, S. S.
2011-10-01
The objective of this study is to provide an easy, accurate and effective technology for the Bangalore city traffic control. This is based on the techniques of image processing and laser beam technology. The core concept chosen here is an image processing technology by the method of automatic number plate recognition system. First number plate is recognized if any vehicle breaks the traffic rules in the signals. The number is fetched from the database of the RTO office by the process of automatic database fetching. Next this sends the notice and penalty related information to the vehicle owner email-id and an SMS sent to vehicle owner. In this paper, we use of cameras with zooming options & laser beams to get accurate pictures further applied image processing techniques such as Edge detection to understand the vehicle, Identifying the location of the number plate, Identifying the number plate for further use, Plain plate number, Number plate with additional information, Number plates in the different fonts. Accessing the database of the vehicle registration office to identify the name and address and other information of the vehicle number. The updates to be made to the database for the recording of the violation and penalty issues. A feed forward artificial neural network is used for OCR. This procedure is particularly important for glyphs that are visually similar such as '8' and '9' and results in training sets of between 25,000 and 40,000 training samples. Over training of the neural network is prevented by Bayesian regularization. The neural network output value is set to 0.05 when the input is not desired glyph, and 0.95 for correct input.
SU-F-T-185: Study of the Robustness of a Proton Arc Technique Based On PBS Beams
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Z; Zheng, Y
Purpose: One potential technique to realize proton arc is through using PBS beams from many directions to form overlaid Bragg peak (OBP) spots and placing these OBP spots throughout the target volume to achieve desired dose distribution. In this study, we analyzed the robustness of this proton arc technique. Methods: We used a cylindrical water phantom of 20 cm in radius in our robustness analysis. To study the range uncertainty effect, we changed the density of the phantom by ±3%. To study the setup uncertainty effect, we shifted the phantom by 3 & 5 mm. We also combined the rangemore » and setup uncertainties (3mm/±3%). For each test plan, we performed dose calculation for the nominal and 6 disturbed scenarios. Two test plans were used, one with single OBP spot and the other consisting of 121 OBP spots covering a 10×10cm{sup 2} area. We compared the dose profiles between the nominal and disturbed scenarios to estimate the impact of the uncertainties. Dose calculation was performed with Gate/GEANT based Monte Carlo software in cloud computing environment. Results: For each of the 7 scenarios, we simulated 100k & 10M events for plans consisting of single OBP spot and 121 OBP spots respectively. For single OBP spot, the setup uncertainty had minimum impact on the spot’s dose profile while range uncertainty had significant impact on the dose profile. For plan consisting of 121 OBP spots, similar effect was observed but the extent of disturbance was much less compared to single OBP spot. Conclusion: For PBS arc technique, range uncertainty has significantly more impact than setup uncertainty. Although single OBP spot can be severely disturbed by the range uncertainty, the overall effect is much less when a large number of OBP spots are used. Robustness optimization for PBS arc technique should consider range uncertainty with priority.« less
Dimensional Changes of Acrylic Resin Denture Bases: Conventional Versus Injection-Molding Technique
Gharechahi, Jafar; Asadzadeh, Nafiseh; Shahabian, Foad; Gharechahi, Maryam
2014-01-01
Objective: Acrylic resin denture bases undergo dimensional changes during polymerization. Injection molding techniques are reported to reduce these changes and thereby improve physical properties of denture bases. The aim of this study was to compare dimensional changes of specimens processed by conventional and injection-molding techniques. Materials and Methods: SR-Ivocap Triplex Hot resin was used for conventional pressure-packed and SR-Ivocap High Impact was used for injection-molding techniques. After processing, all the specimens were stored in distilled water at room temperature until measured. For dimensional accuracy evaluation, measurements were recorded at 24-hour, 48-hour and 12-day intervals using a digital caliper with an accuracy of 0.01 mm. Statistical analysis was carried out by SPSS (SPSS Inc., Chicago, IL, USA) using t-test and repeated-measures ANOVA. Statistical significance was defined at P<0.05. Results: After each water storage period, the acrylic specimens produced by injection exhibited less dimensional changes compared to those produced by the conventional technique. Curing shrinkage was compensated by water sorption with an increase in water storage time decreasing dimensional changes. Conclusion: Within the limitations of this study, dimensional changes of acrylic resin specimens were influenced by the molding technique used and SR-Ivocap injection procedure exhibited higher dimensional accuracy compared to conventional molding. PMID:25584050
A Randomized Controlled Trial of the Group-Based Modified Story Memory Technique in TBI
2017-10-01
AWARD NUMBER: W81XWH-16-1-0726 TITLE: A Randomized Controlled Trial of the Group -Based Modified Story Memory Technique in TBI PRINCIPAL...2017 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER A Randomized Controlled Trial of the Group -Based Modified Story Memory Technique in TBI 5b. GRANT...forthcoming, The current study addresses this need through a double blind, placebo- controlled , randomized clinical trial (RCT) of a group
Analyzer-based imaging technique in tomography of cartilage and metal implants: a study at the ESRF
COAN, Paola; MOLLENHAUER, Juergen; WAGNER, Andreas; Muehleman, Carol; BRAVIN, Alberto
2009-01-01
Monitoring the progression of osteoarthritis (OA) and the effects of therapy during clinical trials is still a challenge for present clinical imaging techniques since they present intrinsic limitations and can be sensitive only in case of advanced OA stages. In very severe cases, partial or complete joint replacement surgery is the only solution for reducing pain and restoring the joint functions. Poor imaging quality in practically all medical imaging technologies with respect to joint surfaces and to metal implant imaging calls for the development of new techniques that are sensitive to stages preceding the point of irreversible damage of the cartilage tissue. In this scenario, X-ray phase contrast modalities could play an important role since they can provide improved contrast compared to conventional absorption radiography, with a similar or even reduced tissue radiation dose. In this study, the Analyzer-based imaging (ABI), a technique sensitive to the X-ray refraction and permitting a high scatter rejection, has been successfully applied in-vitro on excised human synovial joints and sheep implants. Pathological and healthy joints as well as metal implants have been imaged in projection and computed tomography ABI mode at high resolution and clinically compatible doses (< 10 mGy). Volume rendering and segmentation permitted visualization of the cartilage from volumetric CT-scans. Results demonstrate that ABI can provide an unequivocal non-invasive diagnosis of the state of disease of the joint and be considered a new tool in orthopaedic research. PMID:18584983
Wear Detection of Drill Bit by Image-based Technique
NASA Astrophysics Data System (ADS)
Sukeri, Maziyah; Zulhilmi Paiz Ismadi, Mohd; Rahim Othman, Abdul; Kamaruddin, Shahrul
2018-03-01
Image processing for computer vision function plays an essential aspect in the manufacturing industries for the tool condition monitoring. This study proposes a dependable direct measurement method to measure the tool wear using image-based analysis. Segmentation and thresholding technique were used as the means to filter and convert the colour image to binary datasets. Then, the edge detection method was applied to characterize the edge of the drill bit. By using cross-correlation method, the edges of original and worn drill bits were correlated to each other. Cross-correlation graphs were able to detect the difference of the worn edge despite small difference between the graphs. Future development will focus on quantifying the worn profile as well as enhancing the sensitivity of the technique.
Gaaz, Tayser Sumer; Sulong, Abu Bakar; Kadhum, Abdul Amir H.; Nassir, Mohamed H.; Al-Amiery, Ahmed A.
2016-01-01
Halloysite nanotubes-thermoplastic polyurethane (HNTs-TPU) nanocomposites are attractive products due to increasing demands for specialized materials. This study attempts to optimize the parameters for injection just before marketing. The study shows the importance of the preparation of the samples and how well these parameters play their roles in the injection. The control parameters for injection are carefully determined to examine the mechanical properties and the density of the HNTs-TPU nanocomposites. Three types of modified HNTs were used as untreated HNTs (uHNTs), sulfuric acid treated (aHNTs) and a combined treatment of polyvinyl alcohol (PVA)-sodium dodecyl sulfate (SDS)-malonic acid (MA) (treatment (mHNTs)). It was found that mHNTs have the most influential effect of producing HNTs-TPU nanocomposites with the best qualities. One possible reason for this extraordinary result is the effect of SDS as a disperser and MA as a crosslinker between HNTs and PVA. For the highest tensile strength, the control parameters are demonstrated at 150 °C (injection temperature), 8 bar (injection pressure), 30 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and mHNT (HNTs type). Meanwhile, the optimized combination of the levels for all six control parameters that provide the highest Young’s modulus and highest density was found to be 150 °C (injection temperature), 8 bar (injection pressure), 32 °C (mold temperature), 8 min (injection time), 3 wt % (HNTs loading) and mHNT (HNTs type). For the best tensile strain, the six control parameters are found to be 160 °C (injection temperature), 8 bar (injection pressure), 32 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and mHNT (HNTs type). For the highest hardness, the best parameters are 140 °C (injection temperature), 6 bar (injection pressure), 30 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and mHNT (HNTs type). The analyses are carried out by
Gaaz, Tayser Sumer; Sulong, Abu Bakar; Kadhum, Abdul Amir H; Nassir, Mohamed H; Al-Amiery, Ahmed A
2016-11-22
Halloysite nanotubes-thermoplastic polyurethane (HNTs-TPU) nanocomposites are attractive products due to increasing demands for specialized materials. This study attempts to optimize the parameters for injection just before marketing. The study shows the importance of the preparation of the samples and how well these parameters play their roles in the injection. The control parameters for injection are carefully determined to examine the mechanical properties and the density of the HNTs-TPU nanocomposites. Three types of modified HNTs were used as untreated HNTs ( u HNTs), sulfuric acid treated ( a HNTs) and a combined treatment of polyvinyl alcohol (PVA)-sodium dodecyl sulfate (SDS)-malonic acid (MA) (treatment ( m HNTs)). It was found that m HNTs have the most influential effect of producing HNTs-TPU nanocomposites with the best qualities. One possible reason for this extraordinary result is the effect of SDS as a disperser and MA as a crosslinker between HNTs and PVA. For the highest tensile strength, the control parameters are demonstrated at 150 °C (injection temperature), 8 bar (injection pressure), 30 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and m HNT (HNTs type). Meanwhile, the optimized combination of the levels for all six control parameters that provide the highest Young's modulus and highest density was found to be 150 °C (injection temperature), 8 bar (injection pressure), 32 °C (mold temperature), 8 min (injection time), 3 wt % (HNTs loading) and m HNT (HNTs type). For the best tensile strain, the six control parameters are found to be 160 °C (injection temperature), 8 bar (injection pressure), 32 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and m HNT (HNTs type). For the highest hardness, the best parameters are 140 °C (injection temperature), 6 bar (injection pressure), 30 °C (mold temperature), 8 min (injection time), 2 wt % (HNTs loading) and m HNT (HNTs type). The analyses are carried
NASA Astrophysics Data System (ADS)
Babagowda; Kadadevara Math, R. S.; Goutham, R.; Srinivas Prasad, K. R.
2018-02-01
Fused deposition modeling is a rapidly growing additive manufacturing technology due to its ability to build functional parts having complex geometry. The mechanical properties of the build part is depends on several process parameters and build material of the printed specimen. The aim of this study is to characterize and optimize the parameters such as layer thickness and PLA build material which is mixed with recycled PLA material. Tensile and flexural or bending test are carried out to determine the mechanical response characteristics of the printed specimen. Taguchi method is used for number of experiments and Taguchi S/N ratio is used to identify the set of parameters which give good results for respective response characteristics, effectiveness of each parameters is investigated by using analysis of variance (ANOVA).
NASA Technical Reports Server (NTRS)
Young, S. G.
1973-01-01
The NASA nickel-base alloy WAZ-20 was analyzed by advanced metallographic techniques to qualitatively and quantitatively characterize its phases and stability. The as-cast alloy contained primary gamma-prime, a coarse gamma-gamma prime eutectic, a gamma-fine gamma prime matrix, and MC carbides. A specimen aged at 870 C for 1000 hours contained these same constituents and a few widely scattered high W particles. No detrimental phases (such as sigma or mu) were observed. Scanning electron microscope, light metallography, and replica electron microscope methods are compared. The value of quantitative electron microprobe techniques such as spot and area analysis is demonstrated.
Experiments on Adaptive Techniques for Host-Based Intrusion Detection
DOE Office of Scientific and Technical Information (OSTI.GOV)
DRAELOS, TIMOTHY J.; COLLINS, MICHAEL J.; DUGGAN, DAVID P.
2001-09-01
This research explores four experiments of adaptive host-based intrusion detection (ID) techniques in an attempt to develop systems that can detect novel exploits. The technique considered to have the most potential is adaptive critic designs (ACDs) because of their utilization of reinforcement learning, which allows learning exploits that are difficult to pinpoint in sensor data. Preliminary results of ID using an ACD, an Elman recurrent neural network, and a statistical anomaly detection technique demonstrate an ability to learn to distinguish between clean and exploit data. We used the Solaris Basic Security Module (BSM) as a data source and performed considerablemore » preprocessing on the raw data. A detection approach called generalized signature-based ID is recommended as a middle ground between signature-based ID, which has an inability to detect novel exploits, and anomaly detection, which detects too many events including events that are not exploits. The primary results of the ID experiments demonstrate the use of custom data for generalized signature-based intrusion detection and the ability of neural network-based systems to learn in this application environment.« less
C-MOS array design techniques: SUMC multiprocessor system study
NASA Technical Reports Server (NTRS)
Clapp, W. A.; Helbig, W. A.; Merriam, A. S.
1972-01-01
The current capabilities of LSI techniques for speed and reliability, plus the possibilities of assembling large configurations of LSI logic and storage elements, have demanded the study of multiprocessors and multiprocessing techniques, problems, and potentialities. Evaluated are three previous systems studies for a space ultrareliable modular computer multiprocessing system, and a new multiprocessing system is proposed that is flexibly configured with up to four central processors, four 1/0 processors, and 16 main memory units, plus auxiliary memory and peripheral devices. This multiprocessor system features a multilevel interrupt, qualified S/360 compatibility for ground-based generation of programs, virtual memory management of a storage hierarchy through 1/0 processors, and multiport access to multiple and shared memory units.
NASA Astrophysics Data System (ADS)
Guía-Tello, J. C.; Pech-Canul, M. A.; Trujillo-Vázquez, E.; Pech-Canul, M. I.
2017-08-01
Controlled atmosphere brazing has a widespread industrial use in the production of aluminum automotive heat exchangers. Good-quality joints between the components depend on the initial condition of materials as well as on the brazing process parameters. In this work, the Taguchi method was used to optimize the brazing parameters with respect to corrosion performance for tube-fin mini-assemblies of an automotive condenser. The experimental design consisted of five factors (micro-channel tube type, flux type, peak temperature, heating rate and dwell time), with two levels each. The corrosion behavior in acidified seawater solution pH 2.8 was evaluated through potentiodynamic polarization and electrochemical impedance spectroscopy (EIS) measurements. Scanning electron microscope (SEM) and energy-dispersive x-ray spectroscopy (EDS) were used to analyze the microstructural features in the joint zone. The results showed that the parameters that most significantly affect the corrosion rate are the type of flux and the peak temperature. The optimal conditions were: micro-channel tube with 4.2 g/m2 of zinc coating, standard flux, 610 °C peak temperature, 5 °C/min heating rate and 4 min dwell time. The corrosion current density value of the confirmation experiment is in excellent agreement with the predicted value. The electrochemical characterization for selected samples gave indication that the brazing conditions had a more significant effect on the kinetics of the hydrogen evolution reaction than on the kinetics of the metal dissolution reaction.
Classification of the Regional Ionospheric Disturbance Based on Machine Learning Techniques
NASA Astrophysics Data System (ADS)
Terzi, Merve Begum; Arikan, Orhan; Karatay, Secil; Arikan, Feza; Gulyaeva, Tamara
2016-08-01
In this study, Total Electron Content (TEC) estimated from GPS receivers is used to model the regional and local variability that differs from global activity along with solar and geomagnetic indices. For the automated classification of regional disturbances, a classification technique based on a robust machine learning technique that have found wide spread use, Support Vector Machine (SVM) is proposed. Performance of developed classification technique is demonstrated for midlatitude ionosphere over Anatolia using TEC estimates generated from GPS data provided by Turkish National Permanent GPS Network (TNPGN-Active) for solar maximum year of 2011. As a result of implementing developed classification technique to Global Ionospheric Map (GIM) TEC data, which is provided by the NASA Jet Propulsion Laboratory (JPL), it is shown that SVM can be a suitable learning method to detect anomalies in TEC variations.
Fabrication of an Optical Fiber Micro-Sphere with a Diameter of Several Tens of Micrometers.
Yu, Huijuan; Huang, Qiangxian; Zhao, Jian
2014-06-25
A new method to fabricate an integrated optical fiber micro-sphere with a diameter within 100 µm, based on the optical fiber tapering technique and the Taguchi method is proposed. Using a 125 µm diameter single-mode (SM) optical fiber, an optical fiber taper with a cone angle is formed with the tapering technique, and the fabrication optimization of a micro-sphere with a diameter of less than 100 µm is achieved using the Taguchi method. The optimum combination of process factors levels is obtained, and the signal-to-noise ratio (SNR) of three quality evaluation parameters and the significance of each process factors influencing them are selected as the two standards. Using the minimum zone method (MZM) to evaluate the quality of the fabricated optical fiber micro-sphere, a three-dimensional (3D) numerical fitting image of its surface profile and the true sphericity are subsequently realized. From the results, an optical fiber micro-sphere with a two-dimensional (2D) diameter less than 80 µm, 2D roundness error less than 0.70 µm, 2D offset distance between the micro-sphere center and the fiber stylus central line less than 0.65 µm, and true sphericity of about 0.5 µm, is fabricated.
Multiresponse Optimization of Process Parameters in Turning of GFRP Using TOPSIS Method
Parida, Arun Kumar; Routara, Bharat Chandra
2014-01-01
Taguchi's design of experiment is utilized to optimize the process parameters in turning operation with dry environment. Three parameters, cutting speed (v), feed (f), and depth of cut (d), with three different levels are taken for the responses like material removal rate (MRR) and surface roughness (R a). The machining is conducted with Taguchi L9 orthogonal array, and based on the S/N analysis, the optimal process parameters for surface roughness and MRR are calculated separately. Considering the larger-the-better approach, optimal process parameters for material removal rate are cutting speed at level 3, feed at level 2, and depth of cut at level 3, that is, v 3-f 2-d 3. Similarly for surface roughness, considering smaller-the-better approach, the optimal process parameters are cutting speed at level 1, feed at level 1, and depth of cut at level 3, that is, v 1-f 1-d 3. Results of the main effects plot indicate that depth of cut is the most influencing parameter for MRR but cutting speed is the most influencing parameter for surface roughness and feed is found to be the least influencing parameter for both the responses. The confirmation test is conducted for both MRR and surface roughness separately. Finally, an attempt has been made to optimize the multiresponses using technique for order preference by similarity to ideal solution (TOPSIS) with Taguchi approach. PMID:27437503
An ionospheric occultation inversion technique based on epoch difference
NASA Astrophysics Data System (ADS)
Lin, Jian; Xiong, Jing; Zhu, Fuying; Yang, Jian; Qiao, Xuejun
2013-09-01
Of the ionospheric radio occultation (IRO) electron density profile (EDP) retrievals, the Abel based calibrated TEC inversion (CTI) is the most widely used technique. In order to eliminate the contribution from the altitude above the RO satellite, it is necessary to utilize the calibrated TEC to retrieve the EDP, which introduces the error due to the coplanar assumption. In this paper, a new technique based on the epoch difference inversion (EDI) is firstly proposed to eliminate this error. The comparisons between CTI and EDI have been done, taking advantage of the simulated and real COSMIC data. The following conclusions can be drawn: the EDI technique can successfully retrieve the EDPs without non-occultation side measurements and shows better performance than the CTI method, especially for lower orbit mission; no matter which technique is used, the inversion results at the higher altitudes are better than those at the lower altitudes, which could be explained theoretically.
GIS-based bivariate statistical techniques for groundwater potential analysis (an example of Iran)
NASA Astrophysics Data System (ADS)
Haghizadeh, Ali; Moghaddam, Davoud Davoudi; Pourghasemi, Hamid Reza
2017-12-01
Groundwater potential analysis prepares better comprehension of hydrological settings of different regions. This study shows the potency of two GIS-based data driven bivariate techniques namely statistical index (SI) and Dempster-Shafer theory (DST) to analyze groundwater potential in Broujerd region of Iran. The research was done using 11 groundwater conditioning factors and 496 spring positions. Based on the ground water potential maps (GPMs) of SI and DST methods, 24.22% and 23.74% of the study area is covered by poor zone of groundwater potential, and 43.93% and 36.3% of Broujerd region is covered by good and very good potential zones, respectively. The validation of outcomes displayed that area under the curve (AUC) of SI and DST techniques are 81.23% and 79.41%, respectively, which shows SI method has slightly a better performance than the DST technique. Therefore, SI and DST methods are advantageous to analyze groundwater capacity and scrutinize the complicated relation between groundwater occurrence and groundwater conditioning factors, which permits investigation of both systemic and stochastic uncertainty. Finally, it can be realized that these techniques are very beneficial for groundwater potential analyzing and can be practical for water-resource management experts.
NASA Astrophysics Data System (ADS)
Chakravarty, Subhas
Since IGY period (1957-58), natural and artificially produced Very Low Frequency (VLF) elec-tromagnetic radiations are being recorded at large number of ground stations all over the world and on-board satellites to study various radio wave-thermal/energetic plasma interactive pro-cesses related to earth's ionosphere-plasmasphere-magnetosphere environment. The terrestrial propagation of these VLF radio waves are primarily enabled through the earth ionosphere wave guide (EIWG) mode to long horizontal distances around the globe and ducted along the ge-omagnetic field lines into the conjugate hemisphere through the plasmasphere-magnetosphere regions. The time frequency spectra of the received signals indicate presence of dispersion (wave/group velocities changing with frequency) and various cut-off frequencies based on the width of the EIWG, electron gyro and plasma frequencies etc., providing several types of received signals like whistlers, chorus, tweeks, hiss and hisslers which can be heard on loud-speakers/earphones with distinguishing audio structures. While the VLF technique has been a very effective tool for studying middle and high latitude phenomena, the importance of the similar and anomalous observations over the Indian low latitude stations provide potentially new challenges for their scientific interpretation and modelling. The ducted and non-ducted magnetospheric propagation, pro-longitudinal (PL) mode, low latitude TRIMPI/TLE (Tran-sient Luminous Emissions) or other effects of wave-particle/wave-wave interactions, effects due to ionospheric irregularities and electric fields, full wave solutions to D-region ionisation per-turbations due to solar and stellar energetic X-and γ ray emissions during normal and flaring conditions are a few problems which have been addressed in these low latitude studies over India. Since the conjugate points of Indian stations lie over the Indian oceanic region, the VLF propagation effects would be relatively free from
NASA Astrophysics Data System (ADS)
Moosavi, Saeideh Sadat; Norouzbeigi, Reza; Velayi, Elmira
2017-11-01
In the present work, copper oxide superhydrophobic surface is fabricated on a copper foil via the chemical bath deposition (CBD) method. The effects of some influential factors such as initial concentrations of Cu (II) ions and the surface energy modifier, solution pH, reaction and modification steps time on the wettability property of copper oxide surface were evaluated using Taguchi L16 experimental design. Results showed that the initial concentration of Cu (II) has the most significant impact on the water contact angle and wettability characteristics. The XRD, SEM, AFM and FTIR analyses were used to characterize the copper oxide surfaces. The Water contact angle (WCA) and contact angle hysteresis (CAH) were also measured. The SEM results indicated the formation of a flower-like micro/nano dual-scale structure of copper oxide on the substrate. This structure composed of numerous nano-petals with a thickness of about 50 nm. As a result, a copper oxide hierarchical surface with WCA of 168.4°± 3.5° and CAH of 2.73° exhibited the best superhydrophobicity under proposed optimum condition. This result has been obtained just by 10 min hydrolysis reaction. Besides, this surface showed a good stability under acidic and saline conditions.
Powder metallurgy: Solid and liquid phase sintering of copper
NASA Technical Reports Server (NTRS)
Sheldon, Rex; Weiser, Martin W.
1993-01-01
Basic powder metallurgy (P/M) principles and techniques are presented in this laboratory experiment. A copper based system is used since it is relatively easy to work with and is commercially important. In addition to standard solid state sintering, small quantities of low melting metals such as tin, zinc, lead, and aluminum can be added to demonstrate liquid phase sintering and alloy formation. The Taguchi Method of experimental design was used to study the effect of particle size, pressing force, sintering temperature, and sintering time. These parameters can be easily changed to incorporate liquid phase sintering effects and some guidelines for such substitutions are presented. The experiment is typically carried out over a period of three weeks.
Biosensor-based microRNA detection: techniques, design, performance, and challenges.
Johnson, Blake N; Mutharasan, Raj
2014-04-07
The current state of biosensor-based techniques for amplification-free microRNA (miRNA) detection is critically reviewed. Comparison with non-sensor and amplification-based molecular techniques (MTs), such as polymerase-based methods, is made in terms of transduction mechanism, associated protocol, and sensitivity. Challenges associated with miRNA hybridization thermodynamics which affect assay selectivity and amplification bias are briefly discussed. Electrochemical, electromechanical, and optical classes of miRNA biosensors are reviewed in terms of transduction mechanism, limit of detection (LOD), time-to-results (TTR), multiplexing potential, and measurement robustness. Current trends suggest that biosensor-based techniques (BTs) for miRNA assay will complement MTs due to the advantages of amplification-free detection, LOD being femtomolar (fM)-attomolar (aM), short TTR, multiplexing capability, and minimal sample preparation requirement. Areas of future importance in miRNA BT development are presented which include focus on achieving high measurement confidence and multiplexing capabilities.
Nitrous oxide-based techniques versus nitrous oxide-free techniques for general anaesthesia.
Sun, Rao; Jia, Wen Qin; Zhang, Peng; Yang, KeHu; Tian, Jin Hui; Ma, Bin; Liu, Yali; Jia, Run H; Luo, Xiao F; Kuriyama, Akira
2015-11-06
anaesthesia (or both) with any general anaesthesia using a volatile anaesthetic or propofol-based maintenance of anaesthesia but no nitrous oxide for adults undergoing surgery. Our primary outcome was inhospital case fatality rate. Secondary outcomes were complications and length of stay. Two review authors independently assessed trial quality and extracted the outcome data. We used meta-analysis for data synthesis. Heterogeneity was examined with the Chi² test and by calculating the I² statistic. We used a fixed-effect model if the measure of inconsistency was low for all comparisons (I² statistic < 50%); otherwise we used a random-effects model for measures with high inconsistency. We undertook subgroup analyses to explore inconsistency and sensitivity analyses to evaluate whether the results were robust. We assessed the quality of evidence of the main outcomes using the Grading of Recommendations, Assessment, Development and Evaluation (GRADE) system. We included 35 trials (13,872 adult participants). Seven included studies were at low risk of bias. We identified eight studies as awaiting classification since we could not obtain the full texts, and had insufficient information to include or exclude them. We included data from 24 trials for quantitative synthesis. The results of meta-analyses showed that nitrous oxide-based techniques increased the incidence of pulmonary atelectasis (odds ratio (OR) 1.57, 95% confidence interval (CI) 1.18 to 2.10, P = 0.002), but had no effects on the inhospital case fatality rate, the incidence of pneumonia, myocardial infarction, stroke, severe nausea and vomiting, venous thromboembolism, wound infection, or the length of hospital stay. The sensitivity analyses suggested that the results of the meta-analyses were all robust except for the outcomes of pneumonia, and severe nausea and vomiting. Two trials reported length of intensive care unit (ICU) stay but the data were skewed so were not pooled. Both trials reported that nitrous oxide-based
A review of the different techniques for solid surface acid-base characterization.
Sun, Chenhang; Berg, John C
2003-09-18
In this work, various techniques for solid surface acid-base (AB) characterization are reviewed. Different techniques employ different scales to rank acid-base properties. Based on the results from literature and the authors' own investigations for mineral oxides, these scales are compared. The comparison shows that Isoelectric Point (IEP), the most commonly used AB scale, is not a description of the absolute basicity or acidity of a surface, but a description of their relative strength. That is, a high IEP surface shows more basic functionality comparing with its acidic functionality, whereas a low IEP surface shows less basic functionality comparing with its acidic functionality. The choice of technique and scale for AB characterization depends on the specific application. For the cases in which the overall AB property is of interest, IEP (by electrokinetic titration) and H(0,max) (by indicator dye adsorption) are appropriate. For the cases in which the absolute AB property is of interest such as in the study of adhesion, it is more pertinent to use chemical shift (by XPS) and the heat of adsorption of probe gases (by calorimetry or IGC).
NASA Astrophysics Data System (ADS)
Chowdhury, D. P.; Pal, Sujit; Parthasarathy, R.; Mathur, P. K.; Kohli, A. K.; Limaye, P. K.
1998-09-01
Thin layer activation (TLA) technique has been developed in Zr based alloy materials, e.g., zircaloy II, using 40 MeV α-particles from Variable Energy Cyclotron Centre at Calcutta. A brief description of the methodology of TLA technique is presented to determine the surface wear. The sensitivity of the measurement of surface wear in zircaloy material is found to be 0.22±0.05 μm. The surface wear is determined by TLA technique in zircaloy material which is used in pressurised heavy water reactor and the values have been compared with that obtained by conventional technique for the analytical validation of the TLA technique.
Berry, J.; Nesbit, M.; Saberi, S.; Petridis, H.
2014-01-01
Aim The aim of this study was to identify the communication methods and production techniques used by dentists and dental technicians for the fabrication of fixed prostheses within the UK from the dental technicians' perspective. This second paper reports on the production techniques utilised. Materials and methods Seven hundred and eighty-two online questionnaires were distributed to the Dental Laboratories Association membership and included a broad range of topics, such as demographics, impression disinfection and suitability, and various production techniques. Settings were managed in order to ensure anonymity of respondents. Statistical analysis was undertaken to test the influence of various demographic variables such as the source of information, the location, and the size of the dental laboratory. Results The number of completed responses totalled 248 (32% response rate). Ninety percent of the respondents were based in England and the majority of dental laboratories were categorised as small sized (working with up to 25 dentists). Concerns were raised regarding inadequate disinfection protocols between dentists and dental laboratories and the poor quality of master impressions. Full arch plastic trays were the most popular impression tray used by dentists in the fabrication of crowns (61%) and bridgework (68%). The majority (89%) of jaw registration records were considered inaccurate. Forty-four percent of dental laboratories preferred using semi-adjustable articulators. Axial and occlusal under-preparation of abutment teeth was reported as an issue in about 25% of cases. Base metal alloy was the most (52%) commonly used alloy material. Metal-ceramic crowns were the most popular choice for anterior (69%) and posterior (70%) cases. The various factors considered did not have any statistically significant effect on the answers provided. The only notable exception was the fact that more methods of communicating the size and shape of crowns were utilised for
NASA Astrophysics Data System (ADS)
Nammi, Srinagalakshmi; Vasa, Nilesh J.; Gurusamy, Balaganesan; Mathur, Anil C.
2017-09-01
A plasma shielding phenomenon and its influence on micromachining is studied experimentally and theoretically for laser wavelengths of 355 nm, 532 nm and 1064 nm. A time resolved pump-probe technique is proposed and demonstrated by splitting a single nanosecond Nd3+:YAG laser into an ablation laser (pump laser) and a probe laser to understand the influence of plasma shielding on laser ablation of copper (Cu) clad on polyimide thin films. The proposed nanosecond pump-probe technique allows simultaneous measurement of the absorption characteristics of plasma produced during Cu film ablation by the pump laser. Experimental measurements of the probe intensity distinctly show that the absorption by the ablated plume increases with increase in the pump intensity, as a result of plasma shielding. Theoretical estimation of the intensity of the transmitted pump beam based on the thermo-temporal modeling is in qualitative agreement with the pump-probe based experimental measurements. The theoretical estimate of the depth attained for a single pulse with high pump intensity value on a Cu thin film is limited by the plasma shielding of the incident laser beam, similar to that observed experimentally. Further, the depth of micro-channels produced shows a similar trend for all three wavelengths, however, the channel depth achieved is lesser at the wavelength of 1064 nm.
A New Femtosecond Laser-Based Three-Dimensional Tomography Technique
NASA Astrophysics Data System (ADS)
Echlin, McLean P.
2011-12-01
Tomographic imaging has dramatically changed science, most notably in the fields of medicine and biology, by producing 3D views of structures which are too complex to understand in any other way. Current tomographic techniques require extensive time both for post-processing and data collection. Femtosecond laser based tomographic techniques have been developed in both standard atmosphere (femtosecond laser-based serial sectioning technique - FSLSS) and in vacuum (Tri-Beam System) for the fast collection (10 5mum3/s) of mm3 sized 3D datasets. Both techniques use femtosecond laser pulses to selectively remove layer-by-layer areas of material with low collateral damage and a negligible heat affected zone. To the authors knowledge, femtosecond lasers have never been used to serial section and these techniques have been entirely and uniquely developed by the author and his collaborators at the University of Michigan and University of California Santa Barbara. The FSLSS was applied to measure the 3D distribution of TiN particles in a 4330 steel. Single pulse ablation morphologies and rates were measured and collected from literature. Simultaneous two-phase ablation of TiN and steel matrix was shown to occur at fluences of 0.9-2 J/cm2. Laser scanning protocols were developed minimizing surface roughness to 0.1-0.4 mum for laser-based sectioning. The FSLSS technique was used to section and 3D reconstruct titanium nitride (TiN) containing 4330 steel. Statistical analysis of 3D TiN particle sizes, distribution parameters, and particle density were measured. A methodology was developed to use the 3D datasets to produce statistical volume elements (SVEs) for toughness modeling. Six FSLSS TiN datasets were sub-sampled into 48 SVEs for statistical analysis and toughness modeling using the Rice-Tracey and Garrison-Moody models. A two-parameter Weibull analysis was performed and variability in the toughness data agreed well with Ruggieri et al. bulk toughness measurements. The Tri
Elsyad, Moustafa Abdou; El-Waseef, Fatma Ahmad; Al-Mahdy, Yasmeen Fathy; Fouad, Mohammed Mohammed
2013-08-01
This study aimed to evaluate mandibular denture base deformation along with three impression techniques used for implant-retained overdenture. Ten edentulous patients (five men and five women) received two implants in the canine region of the mandible and three duplicate mandibular overdentures which were constructed with mucostatic, selective pressure, and definitive pressure impression techniques. Ball abutments and respective gold matrices were used to connect the overdentures to the implants. Six linear strain gauges were bonded to the lingual polished surface of each duplicate overdenture at midline and implant areas to measure strain during maximal clenching and gum chewing. The strains recorded at midline were compressive while strains at implant areas were tensile. Clenching recorded significant higher strain when compared with gum chewing for all techniques. The mucostatic technique recorded the highest strain and the definite pressure technique recorded the lowest. There was no significant difference between the strain recorded with mucostatic technique and that registered with selective pressure technique. The highest strain was recorded at the level of ball abutment's top with the mucostatic technique during clenching. Definite pressure impression technique for implant-retained mandibular overdenture is associated with minimal denture deformation during function when compared with mucostatic and selective pressure techniques. Reinforcement of the denture base over the implants may be recommended to increase resistance of fracture when mucostatic or selective pressure impression technique is used. © 2012 John Wiley & Sons A/S.
A VIKOR Technique with Applications Based on DEMATEL and ANP
NASA Astrophysics Data System (ADS)
Ou Yang, Yu-Ping; Shieh, How-Ming; Tzeng, Gwo-Hshiung
In multiple criteria decision making (MCDM) methods, the compromise ranking method (named VIKOR) was introduced as one applicable technique to implement within MCDM. It was developed for multicriteria optimization of complex systems. However, few papers discuss conflicting (competing) criteria with dependence and feedback in the compromise solution method. Therefore, this study proposes and provides applications for a novel model using the VIKOR technique based on DEMATEL and the ANP to solve the problem of conflicting criteria with dependence and feedback. In addition, this research also uses DEMATEL to normalize the unweighted supermatrix of the ANP to suit the real world. An example is also presented to illustrate the proposed method with applications thereof. The results show the proposed method is suitable and effective in real-world applications.
Cell culture-based biosensing techniques for detecting toxicity in water.
Tan, Lu; Schirmer, Kristin
2017-06-01
The significant increase of contaminants entering fresh water bodies calls for the development of rapid and reliable methods to monitor the aquatic environment and to detect water toxicity. Cell culture-based biosensing techniques utilise the overall cytotoxic response to external stimuli, mediated by a transduced signal, to specify the toxicity of aqueous samples. These biosensing techniques can effectively indicate water toxicity for human safety and aquatic organism health. In this review we account for the recent developments of the mainstream cell culture-based biosensing techniques for water quality evaluation, discuss their key features, potentials and limitations, and outline the future prospects of their development. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.
Khogli, Ahmed Eltigani; Cauwels, Rita; Vercruysse, Chris; Verbeeck, Ronald; Martens, Luc
2013-01-01
Optimal pit and fissure sealing is determined by surface preparation techniques and choice of materials. This study aimed (i) to compare the microleakage and penetration depth of a hydrophilic sealant and a conventional resin-based sealant using one of the following preparation techniques: acid etching (AE) only, a diamond bur + AE, and Er:YAG laser combined with AE, and (ii) to evaluate the microleakage and penetration depth of the hydrophilic pit and fissure sealant on different surface conditions. Eighty recently extracted 3rd molars were randomly assigned to eight groups of ten teeth according to the material, preparation technique, and surface condition. For saliva contamination, 0.1 mL of fresh whole human saliva was used. All samples were submitted to 1000 thermal cycles and immersed in 2% methylene blue dye for 4 h. Sections were examined by a light microscope and analysed using image analysis software (Sigmascan(®)). The combination of Er:YAG + AE + conventional sealant showed the least microleakage. The sealing ability of the hydrophilic sealant was influenced by the surface condition. Er:YAG ablation significantly decreased the microleakage at the tooth-sealant interface compared to the non-invasive technique. The hydrophilic sealant applied on different surface conditions showed comparable result to the conventional resin-based sealant. © 2012 The Authors. International Journal of Paediatric Dentistry © 2012 BSPD, IAPD and Blackwell Publishing Ltd.
NASA Astrophysics Data System (ADS)
Idrees, Mohammed Oludare; Pradhan, Biswajeet; Buchroithner, Manfred F.; Shafri, Helmi Zulhaidi Mohd; Khairunniza Bejo, Siti
2016-07-01
As far back as early 15th century during the reign of the Ming Dynasty (1368 to 1634 AD), Gomantong cave in Sabah (Malaysia) has been known as one of the largest roosting sites for wrinkle-lipped bats (Chaerephon plicata) and swiftlet birds (Aerodramus maximus and Aerodramus fuciphagus) in very large colonies. Until recently, no study has been done to quantify or estimate the colony sizes of these inhabitants in spite of the grave danger posed to this avifauna by human activities and potential habitat loss to postspeleogenetic processes. This paper evaluates the transferability of a hybrid optimization image analysis-based method developed to detect and count cave roosting birds. The method utilizes high-resolution terrestrial laser scanning intensity image. First, segmentation parameters were optimized by integrating objective function and the statistical Taguchi methods. Thereafter, the optimized parameters were used as input into the segmentation and classification processes using two images selected from Simud Hitam (lower cave) and Simud Putih (upper cave) of the Gomantong cave. The result shows that the method is capable of detecting birds (and bats) from the image for accurate population censusing. A total number of 9998 swiftlet birds were counted from the first image while 1132 comprising of both bats and birds were obtained from the second image. Furthermore, the transferability evaluation yielded overall accuracies of 0.93 and 0.94 (area under receiver operating characteristic curve) for the first and second image, respectively, with p value of <0.0001 at 95% confidence level. The findings indicate that the method is not only efficient for the detection and counting cave birds for which it was developed for but also useful for counting bats; thus, it can be adopted in any cave.
Model-checking techniques based on cumulative residuals.
Lin, D Y; Wei, L J; Ying, Z
2002-03-01
Residuals have long been used for graphical and numerical examinations of the adequacy of regression models. Conventional residual analysis based on the plots of raw residuals or their smoothed curves is highly subjective, whereas most numerical goodness-of-fit tests provide little information about the nature of model misspecification. In this paper, we develop objective and informative model-checking techniques by taking the cumulative sums of residuals over certain coordinates (e.g., covariates or fitted values) or by considering some related aggregates of residuals, such as moving sums and moving averages. For a variety of statistical models and data structures, including generalized linear models with independent or dependent observations, the distributions of these stochastic processes tinder the assumed model can be approximated by the distributions of certain zero-mean Gaussian processes whose realizations can be easily generated by computer simulation. Each observed process can then be compared, both graphically and numerically, with a number of realizations from the Gaussian process. Such comparisons enable one to assess objectively whether a trend seen in a residual plot reflects model misspecification or natural variation. The proposed techniques are particularly useful in checking the functional form of a covariate and the link function. Illustrations with several medical studies are provided.
Choi, Kihwan; Li, Ruijiang; Nam, Haewon; Xing, Lei
2014-06-21
As a solution to iterative CT image reconstruction, first-order methods are prominent for the large-scale capability and the fast convergence rate [Formula: see text]. In practice, the CT system matrix with a large condition number may lead to slow convergence speed despite the theoretically promising upper bound. The aim of this study is to develop a Fourier-based scaling technique to enhance the convergence speed of first-order methods applied to CT image reconstruction. Instead of working in the projection domain, we transform the projection data and construct a data fidelity model in Fourier space. Inspired by the filtered backprojection formalism, the data are appropriately weighted in Fourier space. We formulate an optimization problem based on weighted least-squares in the Fourier space and total-variation (TV) regularization in image space for parallel-beam, fan-beam and cone-beam CT geometry. To achieve the maximum computational speed, the optimization problem is solved using a fast iterative shrinkage-thresholding algorithm with backtracking line search and GPU implementation of projection/backprojection. The performance of the proposed algorithm is demonstrated through a series of digital simulation and experimental phantom studies. The results are compared with the existing TV regularized techniques based on statistics-based weighted least-squares as well as basic algebraic reconstruction technique. The proposed Fourier-based compressed sensing (CS) method significantly improves both the image quality and the convergence rate compared to the existing CS techniques.
Memristor-Based Computing Architecture: Design Methodologies and Circuit Techniques
2013-03-01
MEMRISTOR-BASED COMPUTING ARCHITECTURE : DESIGN METHODOLOGIES AND CIRCUIT TECHNIQUES POLYTECHNIC INSTITUTE OF NEW YORK UNIVERSITY...TECHNICAL REPORT 3. DATES COVERED (From - To) OCT 2010 – OCT 2012 4. TITLE AND SUBTITLE MEMRISTOR-BASED COMPUTING ARCHITECTURE : DESIGN METHODOLOGIES...schemes for a memristor-based reconfigurable architecture design have not been fully explored yet. Therefore, in this project, we investigated
NASA Astrophysics Data System (ADS)
Vidya Sagar, R.; Raghu Prasad, B. K.
2012-03-01
This article presents a review of recent developments in parametric based acoustic emission (AE) techniques applied to concrete structures. It recapitulates the significant milestones achieved by previous researchers including various methods and models developed in AE testing of concrete structures. The aim is to provide an overview of the specific features of parametric based AE techniques of concrete structures carried out over the years. Emphasis is given to traditional parameter-based AE techniques applied to concrete structures. A significant amount of research on AE techniques applied to concrete structures has already been published and considerable attention has been given to those publications. Some recent studies such as AE energy analysis and b-value analysis used to assess damage of concrete bridge beams have also been discussed. The formation of fracture process zone and the AE energy released during the fracture process in concrete beam specimens have been summarised. A large body of experimental data on AE characteristics of concrete has accumulated over the last three decades. This review of parametric based AE techniques applied to concrete structures may be helpful to the concerned researchers and engineers to better understand the failure mechanism of concrete and evolve more useful methods and approaches for diagnostic inspection of structural elements and failure prediction/prevention of concrete structures.
Modern Micro and Nanoparticle-Based Imaging Techniques
Ryvolova, Marketa; Chomoucka, Jana; Drbohlavova, Jana; Kopel, Pavel; Babula, Petr; Hynek, David; Adam, Vojtech; Eckschlager, Tomas; Hubalek, Jaromir; Stiborova, Marie; Kaiser, Jozef; Kizek, Rene
2012-01-01
The requirements for early diagnostics as well as effective treatment of insidious diseases such as cancer constantly increase the pressure on development of efficient and reliable methods for targeted drug/gene delivery as well as imaging of the treatment success/failure. One of the most recent approaches covering both the drug delivery as well as the imaging aspects is benefitting from the unique properties of nanomaterials. Therefore a new field called nanomedicine is attracting continuously growing attention. Nanoparticles, including fluorescent semiconductor nanocrystals (quantum dots) and magnetic nanoparticles, have proven their excellent properties for in vivo imaging techniques in a number of modalities such as magnetic resonance and fluorescence imaging, respectively. In this article, we review the main properties and applications of nanoparticles in various in vitro imaging techniques, including microscopy and/or laser breakdown spectroscopy and in vivo methods such as magnetic resonance imaging and/or fluorescence-based imaging. Moreover the advantages of the drug delivery performed by nanocarriers such as iron oxides, gold, biodegradable polymers, dendrimers, lipid based carriers such as liposomes or micelles are also highlighted. PMID:23202187
Montero, C; Cristescu, S M; Jiménez, J B; Orea, J M; te Lintel Hekkert, S; Harren, F J M; González Ureña, A
2003-01-01
Two modern laser-based techniques were synchronously applied to study the dynamics of the trans-resveratrol activity in Botrytis cinerea-infected grapes. Direct analysis of trans-resveratrol in both infected and noninfected grapes (Vitis vinifera, Aledo variety) was performed by using an analytical technique incorporating laser desorption coupled with laser resonant ionization and time-of-flight mass spectrometry. On the other hand, one of the most sensitive on-line methods for trace gas detection, laser photoacoustic spectroscopy, was used to investigate the involvement of the plant hormone ethylene (C(2)H(4)) in the B. cinerea grapes interaction and its temporal relationship with the trans-resveratrol content upon infection. The trans-resveratrol content and the ethylene released by noninfected grapes showed an opposite behavior. In this case, a high trans-resveratrol content corresponds to a low ethylene emission. For the B. cinerea-infected grapes, ethylene emission rises up after 48 h when the analogous content of trans-resveratrol started to decrease irreversibly. Moreover, the activity of trans-resveratrol as natural pesticide has been investigated by exogenous application on grapes. A short submerge (5 s) of the grapes in 1.6 x 10(-4) M solution of trans-resveratrol delays the increase of C(2)H(4) emission with about 48 h and produces a decrease of the C(2)H(4) concentration and its emission rate. The treatment has positive effects on fruit conservation during storage; it doubled the normal shelf-life of grapes at room temperature, maintaining their post-harvest quality within 10 d.
NASA Astrophysics Data System (ADS)
Bian, Qiang; Song, Zhangqi; Zhang, Xueliang; Yu, Yang; Chen, Yuzhong
2018-03-01
We proposed a refractive index sensor based on optical fiber end face using pulse reference-based compensation technique. With good compensation effect of this compensation technique, the power fluctuation of light source, the change of optic components transmission loss and coupler splitting ratio can be compensated, which largely reduces the background noise. The refractive index resolutions can achieve 3.8 × 10-6 RIU and1.6 × 10-6 RIU in different refractive index regions.
ERIC Educational Resources Information Center
Taguchi, Naoko; Naganuma, Naeko; Budding, Carlos
2015-01-01
This study examined the effects of explicit instruction on the development of pragmatic competence in L2 English. The study is based on Taguchi's (2012) study conducted in an English-medium university in Japan, which revealed patterns of change in Japanese EFL students' production of requests in high- and low-imposition situations. Students showed…
Rasta, Seyed Hossein; Partovi, Mahsa Eisazadeh; Seyedarabi, Hadi; Javadzadeh, Alireza
2015-01-01
To investigate the effect of preprocessing techniques including contrast enhancement and illumination correction on retinal image quality, a comparative study was carried out. We studied and implemented a few illumination correction and contrast enhancement techniques on color retinal images to find out the best technique for optimum image enhancement. To compare and choose the best illumination correction technique we analyzed the corrected red and green components of color retinal images statistically and visually. The two contrast enhancement techniques were analyzed using a vessel segmentation algorithm by calculating the sensitivity and specificity. The statistical evaluation of the illumination correction techniques were carried out by calculating the coefficients of variation. The dividing method using the median filter to estimate background illumination showed the lowest Coefficients of variations in the red component. The quotient and homomorphic filtering methods after the dividing method presented good results based on their low Coefficients of variations. The contrast limited adaptive histogram equalization increased the sensitivity of the vessel segmentation algorithm up to 5% in the same amount of accuracy. The contrast limited adaptive histogram equalization technique has a higher sensitivity than the polynomial transformation operator as a contrast enhancement technique for vessel segmentation. Three techniques including the dividing method using the median filter to estimate background, quotient based and homomorphic filtering were found as the effective illumination correction techniques based on a statistical evaluation. Applying the local contrast enhancement technique, such as CLAHE, for fundus images presented good potentials in enhancing the vasculature segmentation.
NASA Astrophysics Data System (ADS)
Robins, Marthony; Solomon, Justin; Sahbaee, Pooyan; Sedlmair, Martin; Choudhury, Kingshuk Roy; Pezeshk, Aria; Sahiner, Berkman; Samei, Ehsan
2017-09-01
Virtual nodule insertion paves the way towards the development of standardized databases of hybrid CT images with known lesions. The purpose of this study was to assess three methods (an established and two newly developed techniques) for inserting virtual lung nodules into CT images. Assessment was done by comparing virtual nodule volume and shape to the CT-derived volume and shape of synthetic nodules. 24 synthetic nodules (three sizes, four morphologies, two repeats) were physically inserted into the lung cavity of an anthropomorphic chest phantom (KYOTO KAGAKU). The phantom was imaged with and without nodules on a commercial CT scanner (SOMATOM Definition Flash, Siemens) using a standard thoracic CT protocol at two dose levels (1.4 and 22 mGy CTDIvol). Raw projection data were saved and reconstructed with filtered back-projection and sinogram affirmed iterative reconstruction (SAFIRE, strength 5) at 0.6 mm slice thickness. Corresponding 3D idealized, virtual nodule models were co-registered with the CT images to determine each nodule’s location and orientation. Virtual nodules were voxelized, partial volume corrected, and inserted into nodule-free CT data (accounting for system imaging physics) using two methods: projection-based Technique A, and image-based Technique B. Also a third Technique C based on cropping a region of interest from the acquired image of the real nodule and blending it into the nodule-free image was tested. Nodule volumes were measured using a commercial segmentation tool (iNtuition, TeraRecon, Inc.) and deformation was assessed using the Hausdorff distance. Nodule volumes and deformations were compared between the idealized, CT-derived and virtual nodules using a linear mixed effects regression model which utilized the mean, standard deviation, and coefficient of variation (Mea{{n}RHD} , ST{{D}RHD} and C{{V}RHD}{) }~ of the regional Hausdorff distance. Overall, there was a close concordance between the volumes of the CT-derived and
Robins, Marthony; Solomon, Justin; Sahbaee, Pooyan; Sedlmair, Martin; Choudhury, Kingshuk Roy; Pezeshk, Aria; Sahiner, Berkman; Samei, Ehsan
2017-01-01
Virtual nodule insertion paves the way towards the development of standardized databases of hybrid CT images with known lesions. The purpose of this study was to assess three methods (an established and two newly developed techniques) for inserting virtual lung nodules into CT images. Assessment was done by comparing virtual nodule volume and shape to the CT-derived volume and shape of synthetic nodules. 24 synthetic nodules (three sizes, four morphologies, two repeats) were physically inserted into the lung cavity of an anthropomorphic chest phantom (KYOTO KAGAKU). The phantom was imaged with and without nodules on a commercial CT scanner (SOMATOM Definition Flash, Siemens) using a standard thoracic CT protocol at two dose levels (1.4 and 22 mGy CTDIvol). Raw projection data were saved and reconstructed with filtered back-projection and sinogram affirmed iterative reconstruction (SAFIRE, strength 5) at 0.6 mm slice thickness. Corresponding 3D idealized, virtual nodule models were co-registered with the CT images to determine each nodule’s location and orientation. Virtual nodules were voxelized, partial volume corrected, and inserted into nodule-free CT data (accounting for system imaging physics) using two methods: projection-based Technique A, and image-based Technique B. Also a third Technique C based on cropping a region of interest from the acquired image of the real nodule and blending it into the nodule-free image was tested. Nodule volumes were measured using a commercial segmentation tool (iNtuition, TeraRecon, Inc.) and deformation was assessed using the Hausdorff distance. Nodule volumes and deformations were compared between the idealized, CT-derived and virtual nodules using a linear mixed effects regression model which utilized the mean, standard deviation, and coefficient of variation (MeanRHD, and STDRHD CVRHD) of the regional Hausdorff distance. Overall, there was a close concordance between the volumes of the CT-derived and virtual nodules
Nuclear techniques in studies of condensed matter
NASA Technical Reports Server (NTRS)
Singh, Jag J.
1987-01-01
Nuclear techniques have played an important role in the studies of materials over the past several decades. For example, X-ray diffraction, neutron diffraction, neutron activation, and particle- or photon-induced X-ray emission techniques have been used extensively for the elucidation of structural and compositional details of materials. Several new techniques have been developed recently. Four such techniques are briefly reviewed which have great potential in the study and development of new materials. Of these four, Mossbauer spectroscopy, muon spin rotation, and positron annihilation spectroscopy techniques exploit their great sensitivity to the local atomic environments in the test materials. Interest in synchrotron radiation, on the other hand, stems from its special properties, such as high intensity, high degree of polarization, and high monochromaticity. It is hoped that this brief review will stimulate interest in the exploitation of these newer techniques for the development of improved materials.
Studies of industrial emissions by accelerator-based techniques: A review of applications at CEDAD
NASA Astrophysics Data System (ADS)
Calcagnile, L.; Quarta, G.
2012-04-01
Different research activities are in progress at the Centre for Dating and Diagnostics (CEDAD), University of Salento, in the field of environmental monitoring by exploiting the potentialities given by the different experimental beam lines implemented on the 3 MV Tande-tron accelerator and dedicated to AMS (Accelerator Mass Spectrome-try) radiocarbon dating and IB A (Ion Beam Analysis). An overview of these activities is presented by showing how accelerator-based analytical techniques can be a powerful tool for monitoring the anthropogenic carbon dioxide emissions from industrial sources and for the assessment of the biogenic content in SRF (Solid Recovered Fuel) burned in WTE (Waste to Energy) plants.
Comparative study of shear wave-based elastography techniques in optical coherence tomography
NASA Astrophysics Data System (ADS)
Zvietcovich, Fernando; Rolland, Jannick P.; Yao, Jianing; Meemon, Panomsak; Parker, Kevin J.
2017-03-01
We compare five optical coherence elastography techniques able to estimate the shear speed of waves generated by one and two sources of excitation. The first two techniques make use of one piezoelectric actuator in order to produce a continuous shear wave propagation or a tone-burst propagation (TBP) of 400 Hz over a gelatin tissue-mimicking phantom. The remaining techniques utilize a second actuator located on the opposite side of the region of interest in order to create three types of interference patterns: crawling waves, swept crawling waves, and standing waves, depending on the selection of the frequency difference between the two actuators. We evaluated accuracy, contrast to noise ratio, resolution, and acquisition time for each technique during experiments. Numerical simulations were also performed in order to support the experimental findings. Results suggest that in the presence of strong internal reflections, single source methods are more accurate and less variable when compared to the two-actuator methods. In particular, TBP reports the best performance with an accuracy error <4.1%. Finally, the TBP was tested in a fresh chicken tibialis anterior muscle with a localized thermally ablated lesion in order to evaluate its performance in biological tissue.
Feathering effect detection and artifact agglomeration index-based video deinterlacing technique
NASA Astrophysics Data System (ADS)
Martins, André Luis; Rodrigues, Evandro Luis Linhari; de Paiva, Maria Stela Veludo
2018-03-01
Several video deinterlacing techniques have been developed, and each one presents a better performance in certain conditions. Occasionally, even the most modern deinterlacing techniques create frames with worse quality than primitive deinterlacing processes. This paper validates that the final image quality can be improved by combining different types of deinterlacing techniques. The proposed strategy is able to select between two types of deinterlaced frames and, if necessary, make the local correction of the defects. This decision is based on an artifact agglomeration index obtained from a feathering effect detection map. Starting from a deinterlaced frame produced by the "interfield average" method, the defective areas are identified, and, if deemed appropriate, these areas are replaced by pixels generated through the "edge-based line average" method. Test results have proven that the proposed technique is able to produce video frames with higher quality than applying a single deinterlacing technique through getting what is good from intra- and interfield methods.
Analyzing the effectiveness of a frame-level redundancy scrubbing technique for SRAM-based FPGAs
Tonfat, Jorge; Lima Kastensmidt, Fernanda; Rech, Paolo; ...
2015-12-17
Radiation effects such as soft errors are the major threat to the reliability of SRAM-based FPGAs. This work analyzes the effectiveness in correcting soft errors of a novel scrubbing technique using internal frame redundancy called Frame-level Redundancy Scrubbing (FLR-scrubbing). This correction technique can be implemented in a coarse grain TMR design. The FLR-scrubbing technique was implemented on a mid-size Xilinx Virtex-5 FPGA device used as a case study. The FLR-scrubbing technique was tested under neutron radiation and fault injection. Implementation results demonstrated minimum area and energy consumption overhead when compared to other techniques. The time to repair the fault ismore » also improved by using the Internal Configuration Access Port (ICAP). Lastly, neutron radiation test results demonstrated that the proposed technique is suitable for correcting accumulated SEUs and MBUs.« less
Analyzing the effectiveness of a frame-level redundancy scrubbing technique for SRAM-based FPGAs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tonfat, Jorge; Lima Kastensmidt, Fernanda; Rech, Paolo
Radiation effects such as soft errors are the major threat to the reliability of SRAM-based FPGAs. This work analyzes the effectiveness in correcting soft errors of a novel scrubbing technique using internal frame redundancy called Frame-level Redundancy Scrubbing (FLR-scrubbing). This correction technique can be implemented in a coarse grain TMR design. The FLR-scrubbing technique was implemented on a mid-size Xilinx Virtex-5 FPGA device used as a case study. The FLR-scrubbing technique was tested under neutron radiation and fault injection. Implementation results demonstrated minimum area and energy consumption overhead when compared to other techniques. The time to repair the fault ismore » also improved by using the Internal Configuration Access Port (ICAP). Lastly, neutron radiation test results demonstrated that the proposed technique is suitable for correcting accumulated SEUs and MBUs.« less
Performance analysis of clustering techniques over microarray data: A case study
NASA Astrophysics Data System (ADS)
Dash, Rasmita; Misra, Bijan Bihari
2018-03-01
Handling big data is one of the major issues in the field of statistical data analysis. In such investigation cluster analysis plays a vital role to deal with the large scale data. There are many clustering techniques with different cluster analysis approach. But which approach suits a particular dataset is difficult to predict. To deal with this problem a grading approach is introduced over many clustering techniques to identify a stable technique. But the grading approach depends on the characteristic of dataset as well as on the validity indices. So a two stage grading approach is implemented. In this study the grading approach is implemented over five clustering techniques like hybrid swarm based clustering (HSC), k-means, partitioning around medoids (PAM), vector quantization (VQ) and agglomerative nesting (AGNES). The experimentation is conducted over five microarray datasets with seven validity indices. The finding of grading approach that a cluster technique is significant is also established by Nemenyi post-hoc hypothetical test.
Retraction techniques in laparoscopic colorectal surgery: a literature-based review.
Ladwa, N; Sajid, M S; Pankhania, N K; Sains, P; Baig, M K
2013-08-01
To systematically review the published literature and describe the various techniques of bowel and mesentery retraction available for use in laparoscopic colorectal resection. A comprehensive search of the literature was undertaken using MESH terms 'retraction', 'laparoscopic' and 'colorectal'. All articles describing methods of retraction in laparoscopic colorectal surgery were included. Twelve methods of retraction in laparoscopic colorectal surgery were described. Five case-based series and three case studies were reported on 108 patients. Techniques were classified into those offering retraction of the small or large bowel or according to the mode of retraction. Many retraction methods are available to the surgeon varying in cost, invasiveness and complexity. Adequate retraction remains a challenge for optimal exposure and dissection during laparoscopic colorectal surgery. © 2013 The Authors Colorectal Disease © 2013 The Association of Coloproctology of Great Britain and Ireland.
A new imaging technique based on resonance for arterial vessels
NASA Astrophysics Data System (ADS)
Zhang, Xiaoming; Fatemi, Mostafa; Greenleaf, James F.
2003-04-01
Vibro-acoustography is a new noncontact imaging method based on the radiation force of ultrasound. We extend this technique for imaging of arterial vessels based on vibration resonance. The arterial vessel is excited remotely by ultrasound at a resonant frequency, at which the vibration of the vessel as well as its transmission to the body surface are large enough to be measured. By scanning the ultrasound beam across the vessel plane and measuring the vibration at one single point on the body or vessel surface, an image of the interior artery can be mapped. Theory is developed that predicts the measured velocity is proportional to the value of the mode shape at resonance. Experimental studies were carried out on a silicone tube embedded in a cylindrical gel phantom of large radius, which simulates a large artery and the surrounding body. The fundamental frequency was measured at which the ultrasound transducer scanned across the tube plane with velocity measurement at one single point on the tube or on the phantom by laser. The images obtained show clearly the interior tube and the modal shape of the tube. The present technique offers a new imaging method for arterial vessels.
NASA Technical Reports Server (NTRS)
Size, P.; Takeuchi, Esther S.
1993-01-01
The purpose of this contract is to evaluate parametrically the effects of various factors including the electrolyte type, electrolyte concentration, depolarizer type, and cell configuration on lithium cell electrical performance and safety. This effort shall allow for the selection and optimization of cell design for future NASA applications while maintaining close ties with WGL's continuous improvements in manufacturing processes and lithium cell design. Taguchi experimental design techniques are employed in this task, and allow for a maximum amount of information to be obtained while requiring significantly less cells than if a full factorial design were employed. Acceptance testing for this task is modeled after the NASA Document EP5-83-025, Revision C, for cell weights, OCV's and load voltages. The performance attributes that are studied in this effort are fresh capacity and start-up characteristics evaluated at two rates and two temperatures, shelf-life characteristics including start-up and capacity retention, and iterative microcalorimetry measurements. Abuse testing includes forced over discharge at two rates with and without diode protection, temperature tolerance testing, and shorting tests at three rates with the measurement of heat generated during shorting conditions.
Rasta, Seyed Hossein; Partovi, Mahsa Eisazadeh; Seyedarabi, Hadi; Javadzadeh, Alireza
2015-01-01
To investigate the effect of preprocessing techniques including contrast enhancement and illumination correction on retinal image quality, a comparative study was carried out. We studied and implemented a few illumination correction and contrast enhancement techniques on color retinal images to find out the best technique for optimum image enhancement. To compare and choose the best illumination correction technique we analyzed the corrected red and green components of color retinal images statistically and visually. The two contrast enhancement techniques were analyzed using a vessel segmentation algorithm by calculating the sensitivity and specificity. The statistical evaluation of the illumination correction techniques were carried out by calculating the coefficients of variation. The dividing method using the median filter to estimate background illumination showed the lowest Coefficients of variations in the red component. The quotient and homomorphic filtering methods after the dividing method presented good results based on their low Coefficients of variations. The contrast limited adaptive histogram equalization increased the sensitivity of the vessel segmentation algorithm up to 5% in the same amount of accuracy. The contrast limited adaptive histogram equalization technique has a higher sensitivity than the polynomial transformation operator as a contrast enhancement technique for vessel segmentation. Three techniques including the dividing method using the median filter to estimate background, quotient based and homomorphic filtering were found as the effective illumination correction techniques based on a statistical evaluation. Applying the local contrast enhancement technique, such as CLAHE, for fundus images presented good potentials in enhancing the vasculature segmentation. PMID:25709940
NASA Astrophysics Data System (ADS)
Zeqiri, F.; Alkan, M.; Kaya, B.; Toros, S.
2018-01-01
In this paper, the effects of cutting parameters on cutting forces and surface roughness based on Taguchi experimental design method are determined. Taguchi L9 orthogonal array is used to investigate the effects of machining parameters. Optimal cutting conditions are determined using the signal/noise (S/N) ratio which is calculated by average surface roughness and cutting force. Using results of analysis, effects of parameters on both average surface roughness and cutting forces are calculated on Minitab 17 using ANOVA method. The material that was investigated is Inconel 625 steel for two cases with heat treatment and without heat treatment. The predicted and calculated values with measurement are very close to each other. Confirmation test of results showed that the Taguchi method was very successful in the optimization of machining parameters for maximum surface roughness and cutting forces in the CNC turning process.
Allen, Edwin B; Walls, Richard T; Reilly, Frank D
2008-02-01
This study investigated the effects of interactive instructional techniques in a web-based peripheral nervous system (PNS) component of a first year medical school human anatomy course. Existing data from 9 years of instruction involving 856 students were used to determine (1) the effect of web-based interactive instructional techniques on written exam item performance and (2) differences between student opinions of the benefit level of five different types of interactive learning objects used. The interactive learning objects included Patient Case studies, review Games, Simulated Interactive Patients (SIP), Flashcards, and unit Quizzes. Exam item analysis scores were found to be significantly higher (p < 0.05) for students receiving the instructional treatment incorporating the web-based interactive learning objects than for students not receiving this treatment. Questionnaires using a five-point Likert scale were analysed to determine student opinion ratings of the interactive learning objects. Students reported favorably on the benefit level of all learning objects. Students rated the benefit level of the Simulated Interactive Patients (SIP) highest, and this rating was significantly higher (p < 0.05) than all other learning objects. This study suggests that web-based interactive instructional techniques improve student exam performance. Students indicated a strong acceptance of Simulated Interactive Patient learning objects.
DCT-based cyber defense techniques
NASA Astrophysics Data System (ADS)
Amsalem, Yaron; Puzanov, Anton; Bedinerman, Anton; Kutcher, Maxim; Hadar, Ofer
2015-09-01
With the increasing popularity of video streaming services and multimedia sharing via social networks, there is a need to protect the multimedia from malicious use. An attacker may use steganography and watermarking techniques to embed malicious content, in order to attack the end user. Most of the attack algorithms are robust to basic image processing techniques such as filtering, compression, noise addition, etc. Hence, in this article two novel, real-time, defense techniques are proposed: Smart threshold and anomaly correction. Both techniques operate at the DCT domain, and are applicable for JPEG images and H.264 I-Frames. The defense performance was evaluated against a highly robust attack, and the perceptual quality degradation was measured by the well-known PSNR and SSIM quality assessment metrics. A set of defense techniques is suggested for improving the defense efficiency. For the most aggressive attack configuration, the combination of all the defense techniques results in 80% protection against cyber-attacks with PSNR of 25.74 db.
Theoretical Bound of CRLB for Energy Efficient Technique of RSS-Based Factor Graph Geolocation
NASA Astrophysics Data System (ADS)
Kahar Aziz, Muhammad Reza; Heriansyah; Saputra, EfaMaydhona; Musa, Ardiansyah
2018-03-01
To support the increase of wireless geolocation development as the key of the technology in the future, this paper proposes theoretical bound derivation, i.e., Cramer Rao lower bound (CRLB) for energy efficient of received signal strength (RSS)-based factor graph wireless geolocation technique. The theoretical bound derivation is crucially important to evaluate whether the energy efficient technique of RSS-based factor graph wireless geolocation is effective as well as to open the opportunity to further innovation of the technique. The CRLB is derived in this paper by using the Fisher information matrix (FIM) of the main formula of the RSS-based factor graph geolocation technique, which is lied on the Jacobian matrix. The simulation result shows that the derived CRLB has the highest accuracy as a bound shown by its lowest root mean squared error (RMSE) curve compared to the RMSE curve of the RSS-based factor graph geolocation technique. Hence, the derived CRLB becomes the lower bound for the efficient technique of RSS-based factor graph wireless geolocation.
Structural damage diagnostics via wave propagation-based filtering techniques
NASA Astrophysics Data System (ADS)
Ayers, James T., III
Structural health monitoring (SHM) of aerospace components is a rapidly emerging field due in part to commercial and military transport vehicles remaining in operation beyond their designed life cycles. Damage detection strategies are sought that provide real-time information of the structure's integrity. One approach that has shown promise to accurately identify and quantify structural defects is based on guided ultrasonic wave (GUW) inspections, where low amplitude attenuation properties allow for long range and large specimen evaluation. One drawback to GUWs is that they exhibit a complex multi-modal response, such that each frequency corresponds to at least two excited modes, and thus intelligent signal processing is required for even the simplest of structures. In addition, GUWs are dispersive, whereby the wave velocity is a function of frequency, and the shape of the wave packet changes over the spatial domain, requiring sophisticated detection algorithms. Moreover, existing damage quantification measures are typically formulated as a comparison of the damaged to undamaged response, which has proven to be highly sensitive to changes in environment, and therefore often unreliable. As a response to these challenges inherent to GUW inspections, this research develops techniques to locate and estimate the severity of the damage. Specifically, a phase gradient based localization algorithm is introduced to identify the defect position independent of excitation frequency and damage size. Mode separation through the filtering technique is central in isolating and extracting single mode components, such as reflected, converted, and transmitted modes that may arise from the incident wave impacting a damage. Spatially-integrated single and multiple component mode coefficients are also formulated with the intent to better characterize wave reflections and conversions and to increase the signal to noise ratios. The techniques are applied to damaged isotropic finite
Knowledge based systems: A critical survey of major concepts, issues and techniques. Visuals
NASA Technical Reports Server (NTRS)
Dominick, Wayne D. (Editor); Kavi, Srinu
1984-01-01
This Working Paper Series entry represents a collection of presentation visuals associated with the companion report entitled, Knowledge Based Systems: A Critical Survey of Major Concepts, Issues, and Techniques, USL/DBMS NASA/RECON Working Paper Series report number DBMS.NASA/RECON-9. The objectives of the report are to: examine various techniques used to build the KBS; to examine at least one KBS in detail, i.e., a case study; to list and identify limitations and problems with the KBS; to suggest future areas of research; and to provide extensive reference materials.
Electrocardiogram signal denoising based on empirical mode decomposition technique: an overview
NASA Astrophysics Data System (ADS)
Han, G.; Lin, B.; Xu, Z.
2017-03-01
Electrocardiogram (ECG) signal is nonlinear and non-stationary weak signal which reflects whether the heart is functioning normally or abnormally. ECG signal is susceptible to various kinds of noises such as high/low frequency noises, powerline interference and baseline wander. Hence, the removal of noises from ECG signal becomes a vital link in the ECG signal processing and plays a significant role in the detection and diagnosis of heart diseases. The review will describe the recent developments of ECG signal denoising based on Empirical Mode Decomposition (EMD) technique including high frequency noise removal, powerline interference separation, baseline wander correction, the combining of EMD and Other Methods, EEMD technique. EMD technique is a quite potential and prospective but not perfect method in the application of processing nonlinear and non-stationary signal like ECG signal. The EMD combined with other algorithms is a good solution to improve the performance of noise cancellation. The pros and cons of EMD technique in ECG signal denoising are discussed in detail. Finally, the future work and challenges in ECG signal denoising based on EMD technique are clarified.
Web image retrieval using an effective topic and content-based technique
NASA Astrophysics Data System (ADS)
Lee, Ching-Cheng; Prabhakara, Rashmi
2005-03-01
There has been an exponential growth in the amount of image data that is available on the World Wide Web since the early development of Internet. With such a large amount of information and image available and its usefulness, an effective image retrieval system is thus greatly needed. In this paper, we present an effective approach with both image matching and indexing techniques that improvise on existing integrated image retrieval methods. This technique follows a two-phase approach, integrating query by topic and query by example specification methods. In the first phase, The topic-based image retrieval is performed by using an improved text information retrieval (IR) technique that makes use of the structured format of HTML documents. This technique consists of a focused crawler that not only provides for the user to enter the keyword for the topic-based search but also, the scope in which the user wants to find the images. In the second phase, we use query by example specification to perform a low-level content-based image match in order to retrieve smaller and relatively closer results of the example image. From this, information related to the image feature is automatically extracted from the query image. The main objective of our approach is to develop a functional image search and indexing technique and to demonstrate that better retrieval results can be achieved.
Prediction of drug synergy in cancer using ensemble-based machine learning techniques
NASA Astrophysics Data System (ADS)
Singh, Harpreet; Rana, Prashant Singh; Singh, Urvinder
2018-04-01
Drug synergy prediction plays a significant role in the medical field for inhibiting specific cancer agents. It can be developed as a pre-processing tool for therapeutic successes. Examination of different drug-drug interaction can be done by drug synergy score. It needs efficient regression-based machine learning approaches to minimize the prediction errors. Numerous machine learning techniques such as neural networks, support vector machines, random forests, LASSO, Elastic Nets, etc., have been used in the past to realize requirement as mentioned above. However, these techniques individually do not provide significant accuracy in drug synergy score. Therefore, the primary objective of this paper is to design a neuro-fuzzy-based ensembling approach. To achieve this, nine well-known machine learning techniques have been implemented by considering the drug synergy data. Based on the accuracy of each model, four techniques with high accuracy are selected to develop ensemble-based machine learning model. These models are Random forest, Fuzzy Rules Using Genetic Cooperative-Competitive Learning method (GFS.GCCL), Adaptive-Network-Based Fuzzy Inference System (ANFIS) and Dynamic Evolving Neural-Fuzzy Inference System method (DENFIS). Ensembling is achieved by evaluating the biased weighted aggregation (i.e. adding more weights to the model with a higher prediction score) of predicted data by selected models. The proposed and existing machine learning techniques have been evaluated on drug synergy score data. The comparative analysis reveals that the proposed method outperforms others in terms of accuracy, root mean square error and coefficient of correlation.
AFM-Based Single Molecule Techniques: Unraveling the Amyloid Pathogenic Species
Ruggeri, Francesco Simone; Habchi, Johnny; Cerreta, Andrea; Dietler, Giovanni
2016-01-01
Background A wide class of human diseases and neurodegenerative disorders, such as Alzheimer’s disease, is due to the failure of a specific peptide or protein to keep its native functional conformational state and to undergo a conformational change into a misfolded state, triggering the formation of fibrillar cross-β sheet amyloid aggregates. During the fibrillization, several coexisting species are formed, giving rise to a highly heterogeneous mixture. Despite its fundamental role in biological function and malfunction, the mechanism of protein self-assembly and the fundamental origins of the connection between aggregation, cellular toxicity and the biochemistry of neurodegeneration remains challenging to elucidate in molecular detail. In particular, the nature of the specific state of proteins that is most prone to cause cytotoxicity is not established. Methods: In the present review, we present the latest advances obtained by Atomic Force Microscopy (AFM) based techniques to unravel the biophysical properties of amyloid aggregates at the nanoscale. Unraveling amyloid single species biophysical properties still represents a formidable experimental challenge, mainly because of their nanoscale dimensions and heterogeneous nature. Bulk techniques, such as circular dichroism or infrared spectroscopy, are not able to characterize the heterogeneity and inner properties of amyloid aggregates at the single species level, preventing a profound investigation of the correlation between the biophysical properties and toxicity of the individual species. Conclusion: The information delivered by AFM based techniques could be central to study the aggregation pathway of proteins and to design molecules that could interfere with amyloid aggregation delaying the onset of misfolding diseases. PMID:27189600
Validating Ultrasound-based HIFU Lesion-size Monitoring Technique with MR Thermometry and Histology
NASA Astrophysics Data System (ADS)
Zhou, Shiwei; Petruzzello, John; Anand, Ajay; Sethuraman, Shriram; Azevedo, Jose
2010-03-01
In order to control and monitor HIFU lesions accurately and cost-effectively in real-time, we have developed an ultrasound-based therapy monitoring technique using acoustic radiation force to track the change in tissue mechanical properties. We validate our method with concurrent MR thermometry and histology. Comparison studies have been completed on in-vitro bovine liver samples. A single-element 1.1 MHz focused transducer was used to deliver HIFU and produce acoustic radiation force (ARF). A 5 MHz single-element transducer was placed co-axially with the HIFU transducer to acquire the RF data, and track the tissue displacement induced by ARF. During therapy, the monitoring procedure was interleaved with HIFU. MR thermometry (Philips Panorama 1T system) and ultrasound monitoring were performed simultaneously. The tissue temperature and thermal dose (CEM43 = 240 mins) were computed from the MR thermometry data. The tissue displacement induced by the acoustic radiation force was calculated from the ultrasound RF data in real-time using a cross-correlation based method. A normalized displacement difference (NDD) parameter was developed and calibrated to estimate the lesion size. The lesion size estimated by the NDD was compared with both MR thermometry prediction and the histology analysis. For lesions smaller than 8mm, the NDD-based lesion monitoring technique showed very similar performance as MR thermometry. The standard deviation of lesion size error is 0.66 mm, which is comparable to MR thermal dose contour prediction (0.42 mm). A phased array is needed for tracking displacement in 2D and monitoring lesion larger than 8 mm. The study demonstrates the potential of our ultrasound based technique to achieve precise HIFU lesion control through real-time monitoring. The results compare well with histology and an established technique like MR Thermometry. This approach provides feedback control in real-time to terminate therapy when a pre-determined lesion size is
Choi, Hyun Ho; Lee, Ju Hwan; Kim, Sung Min; Park, Sung Yun
2015-01-01
Here, the speckle noise in ultrasonic images is removed using an image fusion-based denoising method. To optimize the denoising performance, each discrete wavelet transform (DWT) and filtering technique was analyzed and compared. In addition, the performances were compared in order to derive the optimal input conditions. To evaluate the speckle noise removal performance, an image fusion algorithm was applied to the ultrasound images, and comparatively analyzed with the original image without the algorithm. As a result, applying DWT and filtering techniques caused information loss and noise characteristics, and did not represent the most significant noise reduction performance. Conversely, an image fusion method applying SRAD-original conditions preserved the key information in the original image, and the speckle noise was removed. Based on such characteristics, the input conditions of SRAD-original had the best denoising performance with the ultrasound images. From this study, the best denoising technique proposed based on the results was confirmed to have a high potential for clinical application.
Validation of motion correction techniques for liver CT perfusion studies
Chandler, A; Wei, W; Anderson, E F; Herron, D H; Ye, Z; Ng, C S
2012-01-01
Objectives Motion in images potentially compromises the evaluation of temporally acquired CT perfusion (CTp) data; image registration should mitigate this, but first requires validation. Our objective was to compare the relative performance of manual, rigid and non-rigid registration techniques to correct anatomical misalignment in acquired liver CTp data sets. Methods 17 data sets in patients with liver tumours who had undergone a CTp protocol were evaluated. Each data set consisted of a cine acquisition during a breath-hold (Phase 1), followed by six further sets of cine scans (each containing 11 images) acquired during free breathing (Phase 2). Phase 2 images were registered to a reference image from Phase 1 cine using two semi-automated intensity-based registration techniques (rigid and non-rigid) and a manual technique (the only option available in the relevant vendor CTp software). The performance of each technique to align liver anatomy was assessed by four observers, independently and blindly, on two separate occasions, using a semi-quantitative visual validation study (employing a six-point score). The registration techniques were statistically compared using an ordinal probit regression model. Results 306 registrations (2448 observer scores) were evaluated. The three registration techniques were significantly different from each other (p=0.03). On pairwise comparison, the semi-automated techniques were significantly superior to the manual technique, with non-rigid significantly superior to rigid (p<0.0001), which in turn was significantly superior to manual registration (p=0.04). Conclusion Semi-automated registration techniques achieved superior alignment of liver anatomy compared with the manual technique. We hope this will translate into more reliable CTp analyses. PMID:22374283
Application of Petri net based analysis techniques to signal transduction pathways.
Sackmann, Andrea; Heiner, Monika; Koch, Ina
2006-11-02
Signal transduction pathways are usually modelled using classical quantitative methods, which are based on ordinary differential equations (ODEs). However, some difficulties are inherent in this approach. On the one hand, the kinetic parameters involved are often unknown and have to be estimated. With increasing size and complexity of signal transduction pathways, the estimation of missing kinetic data is not possible. On the other hand, ODEs based models do not support any explicit insights into possible (signal-) flows within the network. Moreover, a huge amount of qualitative data is available due to high-throughput techniques. In order to get information on the systems behaviour, qualitative analysis techniques have been developed. Applications of the known qualitative analysis methods concern mainly metabolic networks. Petri net theory provides a variety of established analysis techniques, which are also applicable to signal transduction models. In this context special properties have to be considered and new dedicated techniques have to be designed. We apply Petri net theory to model and analyse signal transduction pathways first qualitatively before continuing with quantitative analyses. This paper demonstrates how to build systematically a discrete model, which reflects provably the qualitative biological behaviour without any knowledge of kinetic parameters. The mating pheromone response pathway in Saccharomyces cerevisiae serves as case study. We propose an approach for model validation of signal transduction pathways based on the network structure only. For this purpose, we introduce the new notion of feasible t-invariants, which represent minimal self-contained subnets being active under a given input situation. Each of these subnets stands for a signal flow in the system. We define maximal common transition sets (MCT-sets), which can be used for t-invariant examination and net decomposition into smallest biologically meaningful functional units. The
Ooi, Chia Huey; Chetty, Madhu; Teng, Shyh Wei
2006-06-23
Due to the large number of genes in a typical microarray dataset, feature selection looks set to play an important role in reducing noise and computational cost in gene expression-based tissue classification while improving accuracy at the same time. Surprisingly, this does not appear to be the case for all multiclass microarray datasets. The reason is that many feature selection techniques applied on microarray datasets are either rank-based and hence do not take into account correlations between genes, or are wrapper-based, which require high computational cost, and often yield difficult-to-reproduce results. In studies where correlations between genes are considered, attempts to establish the merit of the proposed techniques are hampered by evaluation procedures which are less than meticulous, resulting in overly optimistic estimates of accuracy. We present two realistically evaluated correlation-based feature selection techniques which incorporate, in addition to the two existing criteria involved in forming a predictor set (relevance and redundancy), a third criterion called the degree of differential prioritization (DDP). DDP functions as a parameter to strike the balance between relevance and redundancy, providing our techniques with the novel ability to differentially prioritize the optimization of relevance against redundancy (and vice versa). This ability proves useful in producing optimal classification accuracy while using reasonably small predictor set sizes for nine well-known multiclass microarray datasets. For multiclass microarray datasets, especially the GCM and NCI60 datasets, DDP enables our filter-based techniques to produce accuracies better than those reported in previous studies which employed similarly realistic evaluation procedures.
Self-Alignment MEMS IMU Method Based on the Rotation Modulation Technique on a Swing Base
Chen, Zhiyong; Yang, Haotian; Wang, Chengbin; Lin, Zhihui; Guo, Meifeng
2018-01-01
The micro-electro-mechanical-system (MEMS) inertial measurement unit (IMU) has been widely used in the field of inertial navigation due to its small size, low cost, and light weight, but aligning MEMS IMUs remains a challenge for researchers. MEMS IMUs have been conventionally aligned on a static base, requiring other sensors, such as magnetometers or satellites, to provide auxiliary information, which limits its application range to some extent. Therefore, improving the alignment accuracy of MEMS IMU as much as possible under swing conditions is of considerable value. This paper proposes an alignment method based on the rotation modulation technique (RMT), which is completely self-aligned, unlike the existing alignment techniques. The effect of the inertial sensor errors is mitigated by rotating the IMU. Then, inertial frame-based alignment using the rotation modulation technique (RMT-IFBA) achieved coarse alignment on the swing base. The strong tracking filter (STF) further improved the alignment accuracy. The performance of the proposed method was validated with a physical experiment, and the results of the alignment showed that the standard deviations of pitch, roll, and heading angle were 0.0140°, 0.0097°, and 0.91°, respectively, which verified the practicality and efficacy of the proposed method for the self-alignment of the MEMS IMU on a swing base. PMID:29649150
Sridhara Rao, Duggi V; Sankarasubramanian, Ramachandran; Muraleedharan, Kuttanellore; Mehrtens, Thorsten; Rosenauer, Andreas; Banerjee, Dipankar
2014-08-01
In GaAs-based pseudomorphic high-electron mobility transistor device structures, strain and composition of the In x Ga1-x As channel layer are very important as they influence the electronic properties of these devices. In this context, transmission electron microscopy techniques such as (002) dark-field imaging, high-resolution transmission electron microscopy (HRTEM) imaging, scanning transmission electron microscopy-high angle annular dark field (STEM-HAADF) imaging and selected area diffraction, are useful. A quantitative comparative study using these techniques is relevant for assessing the merits and limitations of the respective techniques. In this article, we have investigated strain and composition of the In x Ga1-x As layer with the mentioned techniques and compared the results. The HRTEM images were investigated with strain state analysis. The indium content in this layer was quantified by HAADF imaging and correlated with STEM simulations. The studies showed that the In x Ga1-x As channel layer was pseudomorphically grown leading to tetragonal strain along the [001] growth direction and that the average indium content (x) in the epilayer is ~0.12. We found consistency in the results obtained using various methods of analysis.
Kuiper, Rebecca M; Nederhoff, Tim; Klugkist, Irene
2015-05-01
In this paper, the performance of six types of techniques for comparisons of means is examined. These six emerge from the distinction between the method employed (hypothesis testing, model selection using information criteria, or Bayesian model selection) and the set of hypotheses that is investigated (a classical, exploration-based set of hypotheses containing equality constraints on the means, or a theory-based limited set of hypotheses with equality and/or order restrictions). A simulation study is conducted to examine the performance of these techniques. We demonstrate that, if one has specific, a priori specified hypotheses, confirmation (i.e., investigating theory-based hypotheses) has advantages over exploration (i.e., examining all possible equality-constrained hypotheses). Furthermore, examining reasonable order-restricted hypotheses has more power to detect the true effect/non-null hypothesis than evaluating only equality restrictions. Additionally, when investigating more than one theory-based hypothesis, model selection is preferred over hypothesis testing. Because of the first two results, we further examine the techniques that are able to evaluate order restrictions in a confirmatory fashion by examining their performance when the homogeneity of variance assumption is violated. Results show that the techniques are robust to heterogeneity when the sample sizes are equal. When the sample sizes are unequal, the performance is affected by heterogeneity. The size and direction of the deviations from the baseline, where there is no heterogeneity, depend on the effect size (of the means) and on the trend in the group variances with respect to the ordering of the group sizes. Importantly, the deviations are less pronounced when the group variances and sizes exhibit the same trend (e.g., are both increasing with group number). © 2014 The British Psychological Society.
Video multiple watermarking technique based on image interlacing using DWT.
Ibrahim, Mohamed M; Abdel Kader, Neamat S; Zorkany, M
2014-01-01
Digital watermarking is one of the important techniques to secure digital media files in the domains of data authentication and copyright protection. In the nonblind watermarking systems, the need of the original host file in the watermark recovery operation makes an overhead over the system resources, doubles memory capacity, and doubles communications bandwidth. In this paper, a robust video multiple watermarking technique is proposed to solve this problem. This technique is based on image interlacing. In this technique, three-level discrete wavelet transform (DWT) is used as a watermark embedding/extracting domain, Arnold transform is used as a watermark encryption/decryption method, and different types of media (gray image, color image, and video) are used as watermarks. The robustness of this technique is tested by applying different types of attacks such as: geometric, noising, format-compression, and image-processing attacks. The simulation results show the effectiveness and good performance of the proposed technique in saving system resources, memory capacity, and communications bandwidth.
Low Power LDPC Code Decoder Architecture Based on Intermediate Message Compression Technique
NASA Astrophysics Data System (ADS)
Shimizu, Kazunori; Togawa, Nozomu; Ikenaga, Takeshi; Goto, Satoshi
Reducing the power dissipation for LDPC code decoder is a major challenging task to apply it to the practical digital communication systems. In this paper, we propose a low power LDPC code decoder architecture based on an intermediate message-compression technique which features as follows: (i) An intermediate message compression technique enables the decoder to reduce the required memory capacity and write power dissipation. (ii) A clock gated shift register based intermediate message memory architecture enables the decoder to decompress the compressed messages in a single clock cycle while reducing the read power dissipation. The combination of the above two techniques enables the decoder to reduce the power dissipation while keeping the decoding throughput. The simulation results show that the proposed architecture improves the power efficiency up to 52% and 18% compared to that of the decoder based on the overlapped schedule and the rapid convergence schedule without the proposed techniques respectively.
NASA Technical Reports Server (NTRS)
Ricks, Wendell R.; Abbott, Kathy H.
1987-01-01
A traditional programming technique for controlling the display of optional flight information in a civil transport cockpit is compared to a rule-based technique for the same function. This application required complex decision logic and a frequently modified rule base. The techniques are evaluated for execution efficiency and implementation ease; the criterion used to calculate the execution efficiency is the total number of steps required to isolate hypotheses that were true and the criteria used to evaluate the implementability are ease of modification and verification and explanation capability. It is observed that the traditional program is more efficient than the rule-based program; however, the rule-based programming technique is more applicable for improving programmer productivity.
Kawasaki, Yoshiteru; Hirano, Tetsuya; Miyatake, Katsutoshi; Fujii, Koji; Takeda, Yoshitsugu
2014-07-01
Coracoid base fracture accompanied by acromioclavicular joint dislocation with intact coracoclavicular ligaments is a rare injury. Generally, an open reduction with screw fixation is the first treatment choice, as it protects the important structures around the coracoid process. This report presents a new technique of screw fixation for coracoid base fracture and provides anatomic information on cross-sectional size of the coracoid base obtained by computed tomography (CT). An axial image of the coracoid base was visualized over the neck of the scapula, and a guidewire was inserted into this circle under fluoroscopic guidance. The wire was inserted easily into the neck of scapula across the coracoid base fracture with imaging in only 1 plane. In addition, 25 measurements of the coracoid base were made in 25 subjects on axial CT images. Average length of the long and short axes at the thinnest part of the coracoid base was 13.9 ± 2.0 mm (range 10.6-17.0) and 10.5 ± 2.2 mm (6.6-15.1), respectively. This new screw fixation technique and measurement data on the coracoid base may be beneficial for safety screw fixation of coracoid base fracture.
A Different Web-Based Geocoding Service Using Fuzzy Techniques
NASA Astrophysics Data System (ADS)
Pahlavani, P.; Abbaspour, R. A.; Zare Zadiny, A.
2015-12-01
Geocoding - the process of finding position based on descriptive data such as address or postal code - is considered as one of the most commonly used spatial analyses. Many online map providers such as Google Maps, Bing Maps and Yahoo Maps present geocoding as one of their basic capabilities. Despite the diversity of geocoding services, users usually face some limitations when they use available online geocoding services. In existing geocoding services, proximity and nearness concept is not modelled appropriately as well as these services search address only by address matching based on descriptive data. In addition there are also some limitations in display searching results. Resolving these limitations can enhance efficiency of the existing geocoding services. This paper proposes the idea of integrating fuzzy technique with geocoding process to resolve these limitations. In order to implement the proposed method, a web-based system is designed. In proposed method, nearness to places is defined by fuzzy membership functions and multiple fuzzy distance maps are created. Then these fuzzy distance maps are integrated using fuzzy overlay technique for obtain the results. Proposed methods provides different capabilities for users such as ability to search multi-part addresses, searching places based on their location, non-point representation of results as well as displaying search results based on their priority.
Montero, C.; Cristescu, S.M.; Jiménez, J.B.; Orea, J.M.; te Lintel Hekkert, S.; Harren, F.J.M.; González Ureña, A.
2003-01-01
Two modern laser-based techniques were synchronously applied to study the dynamics of the trans-resveratrol activity in Botrytis cinerea-infected grapes. Direct analysis of trans-resveratrol in both infected and noninfected grapes (Vitis vinifera, Aledo variety) was performed by using an analytical technique incorporating laser desorption coupled with laser resonant ionization and time-of-flight mass spectrometry. On the other hand, one of the most sensitive on-line methods for trace gas detection, laser photoacoustic spectroscopy, was used to investigate the involvement of the plant hormone ethylene (C2H4) in the B. cinerea grapes interaction and its temporal relationship with the trans-resveratrol content upon infection. The trans-resveratrol content and the ethylene released by noninfected grapes showed an opposite behavior. In this case, a high trans-resveratrol content corresponds to a low ethylene emission. For the B. cinerea-infected grapes, ethylene emission rises up after 48 h when the analogous content of trans-resveratrol started to decrease irreversibly. Moreover, the activity of trans-resveratrol as natural pesticide has been investigated by exogenous application on grapes. A short submerge (5 s) of the grapes in 1.6 × 10−4 m solution of trans-resveratrol delays the increase of C2H4 emission with about 48 h and produces a decrease of the C2H4 concentration and its emission rate. The treatment has positive effects on fruit conservation during storage; it doubled the normal shelf-life of grapes at room temperature, maintaining their post-harvest quality within 10 d. PMID:12529521
Statistics and Machine Learning based Outlier Detection Techniques for Exoplanets
NASA Astrophysics Data System (ADS)
Goel, Amit; Montgomery, Michele
2015-08-01
Architectures of planetary systems are observable snapshots in time that can indicate formation and dynamic evolution of planets. The observable key parameters that we consider are planetary mass and orbital period. If planet masses are significantly less than their host star masses, then Keplerian Motion is defined as P^2 = a^3 where P is the orbital period in units of years and a is the orbital period in units of Astronomical Units (AU). Keplerian motion works on small scales such as the size of the Solar System but not on large scales such as the size of the Milky Way Galaxy. In this work, for confirmed exoplanets of known stellar mass, planetary mass, orbital period, and stellar age, we analyze Keplerian motion of systems based on stellar age to seek if Keplerian motion has an age dependency and to identify outliers. For detecting outliers, we apply several techniques based on statistical and machine learning methods such as probabilistic, linear, and proximity based models. In probabilistic and statistical models of outliers, the parameters of a closed form probability distributions are learned in order to detect the outliers. Linear models use regression analysis based techniques for detecting outliers. Proximity based models use distance based algorithms such as k-nearest neighbour, clustering algorithms such as k-means, or density based algorithms such as kernel density estimation. In this work, we will use unsupervised learning algorithms with only the proximity based models. In addition, we explore the relative strengths and weaknesses of the various techniques by validating the outliers. The validation criteria for the outliers is if the ratio of planetary mass to stellar mass is less than 0.001. In this work, we present our statistical analysis of the outliers thus detected.
Studies of EGRET sources with a novel image restoration technique
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tajima, Hiroyasu; Cohen-Tanugi, Johann; Kamae, Tuneyoshi
2007-07-12
We have developed an image restoration technique based on the Richardson-Lucy algorithm optimized for GLAST-LAT image analysis. Our algorithm is original since it utilizes the PSF (point spread function) that is calculated for each event. This is critical for EGRET and GLAST-LAT image analysis since the PSF depends on the energy and angle of incident gamma-rays and varies by more than one order of magnitude. EGRET and GLAST-LAT image analysis also faces Poisson noise due to low photon statistics. Our technique incorporates wavelet filtering to minimize noise effects. We present studies of EGRET sources using this novel image restoration techniquemore » for possible identification of extended gamma-ray sources.« less
Experimental study on internal cooling system in hard turning of HCWCI using CBN tools
NASA Astrophysics Data System (ADS)
Ravi, A. M.; Murigendrappa, S. M.
2018-04-01
In recent times, hard turning became most emerging technique in manufacturing processes, especially to cut high hard materials like high chrome white cast iron (HCWCI). Use of Cubic boron nitride (CBN), pCBN and Carbide tools are most appropriate to shear the metals but are uneconomical. Since hard turning carried out in dry condition, lowering the tool wear by minimizing tool temperature is the only solution. Study reveals, no effective cooling systems are available so for in order to enhance the tool life of the cutting tools and to improve machinability characteristics. The detrimental effect of cutting parameters on cutting temperature is generally controlled by proper selections. The objective of this paper is to develop a new cooling system to control tool tip temperature, thereby minimizing the cutting forces and the tool wear rates. The materials chosen for this work was HCWCI and cutting tools are CBN inserts. Intricate cavities were made on the periphery of the tool holder for easy flow of cold water. Taguchi techniques were adopted to carry out the experimentations. The experimental results confirm considerable reduction in the cutting forces and tool wear rates.
Using data mining techniques to characterize participation in observational studies.
Linden, Ariel; Yarnold, Paul R
2016-12-01
Data mining techniques are gaining in popularity among health researchers for an array of purposes, such as improving diagnostic accuracy, identifying high-risk patients and extracting concepts from unstructured data. In this paper, we describe how these techniques can be applied to another area in the health research domain: identifying characteristics of individuals who do and do not choose to participate in observational studies. In contrast to randomized studies where individuals have no control over their treatment assignment, participants in observational studies self-select into the treatment arm and therefore have the potential to differ in their characteristics from those who elect not to participate. These differences may explain part, or all, of the difference in the observed outcome, making it crucial to assess whether there is differential participation based on observed characteristics. As compared to traditional approaches to this assessment, data mining offers a more precise understanding of these differences. To describe and illustrate the application of data mining in this domain, we use data from a primary care-based medical home pilot programme and compare the performance of commonly used classification approaches - logistic regression, support vector machines, random forests and classification tree analysis (CTA) - in correctly classifying participants and non-participants. We find that CTA is substantially more accurate than the other models. Moreover, unlike the other models, CTA offers transparency in its computational approach, ease of interpretation via the decision rules produced and provides statistical results familiar to health researchers. Beyond their application to research, data mining techniques could help administrators to identify new candidates for participation who may most benefit from the intervention. © 2016 John Wiley & Sons, Ltd.
Jäger, Jessica
2013-07-01
This article reports on a follow-up study exploring the use of play-based evaluation methods to facilitate children's views of therapy. The development and piloting of these techniques, with 12 children in the author's own practice, was previously reported in this journal. It was argued that play-based evaluation methods reduce the power imbalance inherent in adult researcher/interviewer-child relationships and provide children with meaningful ways to share their views. In this article, follow-up research into play-based evaluations with 20 children and 7 different play therapists is drawn upon to explore in greater depth the strengths and weaknesses of these techniques. The study shows that play-based evaluation techniques are important and flexible methods for facilitating children's views of child therapy. It is argued that those play therapists who incorporate their therapeutic skills effectively, maintain flexibility and sensitively attune to the child during the evaluation session, enable the child to explore their views most fully.
Perineal hypospadias: the Bilateral Based (BILAB) skin flap technique.
Hadidi, Ahmed T
2014-01-01
The purpose of this study was to present the "bilateral based skin flap (BILAB) technique" for patients with perineal hypospadias associated with severe deep chordee and report the follow up. The short urethral plate is incised, the hypoplastic corpus spongiosum excised, the glans split in the midline, and a healthy urethral plate is reconstructed from the lateral penile skin and foreskin to the tip of the glans. Between January 2005 and December 2011, the BILAB technique was performed in 68 patients with perineal hypospadias associated with severe chordee as a one or two stage repair. The records of 63 patients who maintained regular follow up were reviewed. The new urethra was reconstructed from the new urethral plate during the same setting in 26 patients. The remaining 37 patients had the urethroplasty performed 3-6 months later as a two stage repair. Patient age ranged from 8 months and 2 years (mean 10 months). Follow-up period ranged from 18 months to 8 years (mean 43 months). Satisfactory results were obtained in 54 patients (85%). Two children developed diverticula, two patients developed glans dehiscence, one child developed fistula, and one child had complete wound dehiscence, and urethroplasty was successfully reconstructed 6 months later. Three patients developed scar contraction 6-12 months after surgery. This was corrected by excision of the scarred intermediate layer. The BILAB technique is a reliable technique for patients with perineal hypospadias. It produces slit like meatus at the tip of the glans. The surgeon may decide intra-operatively whether to complete the urethroplasty in one or two stages. Long term follow up until puberty is being carried out. Copyright © 2014 Elsevier Inc. All rights reserved.
Graphene-based terahertz photodetector by noise thermometry technique
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Ming-Jye, E-mail: mingjye@asiss.sinica.edu.tw; Institute of Physics, Academia Sinica, Taipei 11529, Taiwan; Wang, Ji-Wun
2014-01-20
We report the characteristics of graphene-based terahertz (THz) photodetector based on noise thermometry technique by measuring its noise power at frequency from 4 to 6 GHz. Hot electron system in graphene microbridge is generated after THz photon pumping and creates extra noise power. The equivalent noise temperature and electron temperature increase rapidly in low THz pumping regime and saturate gradually in high THz power regime which is attributed to a faster energy relaxation process involved by stronger electron-phonon interaction. Based on this detector, a conversion efficiency around 0.15 from THz power to noise power in 4–6 GHz span has been achieved.
Yang, Chih-Cheng; Liu, Chang-Lun
2016-08-12
Cold forging is often applied in the fastener industry. Wires in coil form are used as semi-finished products for the production of billets. This process usually requires preliminarily drawing wire coil in order to reduce the diameter of products. The wire usually has to be annealed to improve its cold formability. The quality of spheroidizing annealed wire affects the forming quality of screws. In the fastener industry, most companies use a subcritical process for spheroidized annealing. Various parameters affect the spheroidized annealing quality of steel wire, such as the spheroidized annealing temperature, prolonged heating time, furnace cooling time and flow rate of nitrogen (protective atmosphere). The effects of the spheroidized annealing parameters affect the quality characteristics of steel wire, such as the tensile strength and hardness. A series of experimental tests on AISI 1022 low carbon steel wire are carried out and the Taguchi method is used to obtain optimum spheroidized annealing conditions to improve the mechanical properties of steel wires for cold forming. The results show that the spheroidized annealing temperature and prolonged heating time have the greatest effect on the mechanical properties of steel wires. A comparison between the results obtained using the optimum spheroidizing conditions and the measures using the original settings shows the new spheroidizing parameter settings effectively improve the performance measures over their value at the original settings. The results presented in this paper could be used as a reference for wire manufacturers.
Vision-based system identification technique for building structures using a motion capture system
NASA Astrophysics Data System (ADS)
Oh, Byung Kwan; Hwang, Jin Woo; Kim, Yousok; Cho, Tongjun; Park, Hyo Seon
2015-11-01
This paper presents a new vision-based system identification (SI) technique for building structures by using a motion capture system (MCS). The MCS with outstanding capabilities for dynamic response measurements can provide gage-free measurements of vibrations through the convenient installation of multiple markers. In this technique, from the dynamic displacement responses measured by MCS, the dynamic characteristics (natural frequency, mode shape, and damping ratio) of building structures are extracted after the processes of converting the displacement from MCS to acceleration and conducting SI by frequency domain decomposition. A free vibration experiment on a three-story shear frame was conducted to validate the proposed technique. The SI results from the conventional accelerometer-based method were compared with those from the proposed technique and showed good agreement, which confirms the validity and applicability of the proposed vision-based SI technique for building structures. Furthermore, SI directly employing MCS measured displacements to FDD was performed and showed identical results to those of conventional SI method.
Wang, Dengjiang; Zhang, Weifang; Wang, Xiangyu; Sun, Bo
2016-01-01
This study presents a novel monitoring method for hole-edge corrosion damage in plate structures based on Lamb wave tomographic imaging techniques. An experimental procedure with a cross-hole layout using 16 piezoelectric transducers (PZTs) was designed. The A0 mode of the Lamb wave was selected, which is sensitive to thickness-loss damage. The iterative algebraic reconstruction technique (ART) method was used to locate and quantify the corrosion damage at the edge of the hole. Hydrofluoric acid with a concentration of 20% was used to corrode the specimen artificially. To estimate the effectiveness of the proposed method, the real corrosion damage was compared with the predicted corrosion damage based on the tomographic method. The results show that the Lamb-wave-based tomographic method can be used to monitor the hole-edge corrosion damage accurately. PMID:28774041
Efficient grid-based techniques for density functional theory
NASA Astrophysics Data System (ADS)
Rodriguez-Hernandez, Juan Ignacio
Understanding the chemical and physical properties of molecules and materials at a fundamental level often requires quantum-mechanical models for these substance's electronic structure. This type of many body quantum mechanics calculation is computationally demanding, hindering its application to substances with more than a few hundreds atoms. The supreme goal of many researches in quantum chemistry---and the topic of this dissertation---is to develop more efficient computational algorithms for electronic structure calculations. In particular, this dissertation develops two new numerical integration techniques for computing molecular and atomic properties within conventional Kohn-Sham-Density Functional Theory (KS-DFT) of molecular electronic structure. The first of these grid-based techniques is based on the transformed sparse grid construction. In this construction, a sparse grid is generated in the unit cube and then mapped to real space according to the pro-molecular density using the conditional distribution transformation. The transformed sparse grid was implemented in program deMon2k, where it is used as the numerical integrator for the exchange-correlation energy and potential in the KS-DFT procedure. We tested our grid by computing ground state energies, equilibrium geometries, and atomization energies. The accuracy on these test calculations shows that our grid is more efficient than some previous integration methods: our grids use fewer points to obtain the same accuracy. The transformed sparse grids were also tested for integrating, interpolating and differentiating in different dimensions (n = 1,2,3,6). The second technique is a grid-based method for computing atomic properties within QTAIM. It was also implemented in deMon2k. The performance of the method was tested by computing QTAIM atomic energies, charges, dipole moments, and quadrupole moments. For medium accuracy, our method is the fastest one we know of.
MEMS-Based Power Generation Techniques for Implantable Biosensing Applications
Lueke, Jonathan; Moussa, Walied A.
2011-01-01
Implantable biosensing is attractive for both medical monitoring and diagnostic applications. It is possible to monitor phenomena such as physical loads on joints or implants, vital signs, or osseointegration in vivo and in real time. Microelectromechanical (MEMS)-based generation techniques can allow for the autonomous operation of implantable biosensors by generating electrical power to replace or supplement existing battery-based power systems. By supplementing existing battery-based power systems for implantable biosensors, the operational lifetime of the sensor is increased. In addition, the potential for a greater amount of available power allows additional components to be added to the biosensing module, such as computational and wireless and components, improving functionality and performance of the biosensor. Photovoltaic, thermovoltaic, micro fuel cell, electrostatic, electromagnetic, and piezoelectric based generation schemes are evaluated in this paper for applicability for implantable biosensing. MEMS-based generation techniques that harvest ambient energy, such as vibration, are much better suited for implantable biosensing applications than fuel-based approaches, producing up to milliwatts of electrical power. High power density MEMS-based approaches, such as piezoelectric and electromagnetic schemes, allow for supplemental and replacement power schemes for biosensing applications to improve device capabilities and performance. In addition, this may allow for the biosensor to be further miniaturized, reducing the need for relatively large batteries with respect to device size. This would cause the implanted biosensor to be less invasive, increasing the quality of care received by the patient. PMID:22319362
MEMS-based power generation techniques for implantable biosensing applications.
Lueke, Jonathan; Moussa, Walied A
2011-01-01
Implantable biosensing is attractive for both medical monitoring and diagnostic applications. It is possible to monitor phenomena such as physical loads on joints or implants, vital signs, or osseointegration in vivo and in real time. Microelectromechanical (MEMS)-based generation techniques can allow for the autonomous operation of implantable biosensors by generating electrical power to replace or supplement existing battery-based power systems. By supplementing existing battery-based power systems for implantable biosensors, the operational lifetime of the sensor is increased. In addition, the potential for a greater amount of available power allows additional components to be added to the biosensing module, such as computational and wireless and components, improving functionality and performance of the biosensor. Photovoltaic, thermovoltaic, micro fuel cell, electrostatic, electromagnetic, and piezoelectric based generation schemes are evaluated in this paper for applicability for implantable biosensing. MEMS-based generation techniques that harvest ambient energy, such as vibration, are much better suited for implantable biosensing applications than fuel-based approaches, producing up to milliwatts of electrical power. High power density MEMS-based approaches, such as piezoelectric and electromagnetic schemes, allow for supplemental and replacement power schemes for biosensing applications to improve device capabilities and performance. In addition, this may allow for the biosensor to be further miniaturized, reducing the need for relatively large batteries with respect to device size. This would cause the implanted biosensor to be less invasive, increasing the quality of care received by the patient.
Power strain imaging based on vibro-elastography techniques
NASA Astrophysics Data System (ADS)
Wen, Xu; Salcudean, S. E.
2007-03-01
This paper describes a new ultrasound elastography technique, power strain imaging, based on vibro-elastography (VE) techniques. With this method, tissue is compressed by a vibrating actuator driven by low-pass or band-pass filtered white noise, typically in the 0-20 Hz range. Tissue displacements at different spatial locations are estimated by correlation-based approaches on the raw ultrasound radio frequency signals and recorded in time sequences. The power spectra of these time sequences are computed by Fourier spectral analysis techniques. As the average of the power spectrum is proportional to the squared amplitude of the tissue motion, the square root of the average power over the range of excitation frequencies is used as a measure of the tissue displacement. Then tissue strain is determined by the least squares estimation of the gradient of the displacement field. The computation of the power spectra of the time sequences can be implemented efficiently by using Welch's periodogram method with moving windows or with accumulative windows with a forgetting factor. Compared to the transfer function estimation originally used in VE, the computation of cross spectral densities is not needed, which saves both the memory and computational times. Phantom experiments demonstrate that the proposed method produces stable and operator-independent strain images with high signal-to-noise ratio in real time. This approach has been also tested on a few patient data of the prostate region, and the results are encouraging.
Chen, Yu-Cheng; Tsai, Perng-Jy; Mou, Jin-Luh; Kuo, Yu-Chieh; Wang, Shih-Min; Young, Li-Hao; Wang, Ya-Fen
2012-09-01
In this study, the cost-benefit analysis technique was developed and incorporated into the Taguchi experimental design to determine the optimal operation combination for the purpose of providing a technique solution for controlling both emissions of PCDD/Fs and PAHs, and increasing both the sinter productivity (SP) and sinter strength (SS) simultaneously. Four operating parameters, including the water content, suction pressure, bed height, and type of hearth layer, were selected and all experimental campaigns were conducted on a pilot-scale sinter pot to simulate various sintering operating conditions of a real-scale sinter plant. The resultant optimal combination could reduce the total carcinogenic emissions arising from both emissions of PCDD/Fs and PAHs by 49.8%, and increase the sinter benefit associated with the increase in both SP and SS by 10.1%, as in comparison with the operation condition currently used in the real plant. The ANOVA results indicate that the suction pressure was the most dominant parameter in determining the optimal operation combination. The above result was theoretically plausible since the higher suction pressure provided more oxygen contents leading to the decrease in both PCDD/F and PAH emissions. But it should be noted that the results obtained from the present study were based on pilot scale experiments, conducting confirmation tests in a real scale plant are still necessary in the future. Copyright © 2012 Elsevier Ltd. All rights reserved.
A pressure-based force and torque prediction technique for the study of fish-like swimming
Dabiri, John O.; Lauder, George V.
2017-01-01
Many outstanding questions about the evolution and function of fish morphology are linked to swimming dynamics, and a detailed knowledge of time-varying forces and torques along the animal’s body is a key component in answering many of these questions. Yet, quantifying these forces and torques experimentally represents a major challenge that to date prevents a full understanding of fish-like swimming. Here, we develop a method for obtaining these force and torque data non-invasively using standard 2D digital particle image velocimetry in conjunction with a pressure field algorithm. We use a mechanical flapping foil apparatus to model fish-like swimming and measure forces and torques directly with a load cell, and compare these measured values to those estimated simultaneously using our pressure-based approach. We demonstrate that, when out-of-plane flows are relatively small compared to the planar flow, and when pressure effects sufficiently dominate shear effects, this technique is able to accurately reproduce the shape, magnitude, and timing of locomotor forces and torques experienced by a fish-like swimmer. We conclude by exploring of the limits of this approach and its feasibility in the study of freely-swimming fishes. PMID:29216264
Combination Base64 Algorithm and EOF Technique for Steganography
NASA Astrophysics Data System (ADS)
Rahim, Robbi; Nurdiyanto, Heri; Hidayat, Rahmat; Saleh Ahmar, Ansari; Siregar, Dodi; Putera Utama Siahaan, Andysah; Faisal, Ilham; Rahman, Sayuti; Suita, Diana; Zamsuri, Ahmad; Abdullah, Dahlan; Napitupulu, Darmawan; Ikhsan Setiawan, Muhammad; Sriadhi, S.
2018-04-01
The steganography process combines mathematics and computer science. Steganography consists of a set of methods and techniques to embed the data into another media so that the contents are unreadable to anyone who does not have the authority to read these data. The main objective of the use of base64 method is to convert any file in order to achieve privacy. This paper discusses a steganography and encoding method using base64, which is a set of encoding schemes that convert the same binary data to the form of a series of ASCII code. Also, the EoF technique is used to embed encoding text performed by Base64. As an example, for the mechanisms a file is used to represent the texts, and by using the two methods together will increase the security level for protecting the data, this research aims to secure many types of files in a particular media with a good security and not to damage the stored files and coverage media that used.
NASA Astrophysics Data System (ADS)
Chockalingam, Letchumanan
2005-01-01
The data of Gunung Ledang region of Malaysia acquired through LANDSAT are considered to map certain hydrogeolocial features. To map these significant features, image-processing tools such as contrast enhancement, edge detection techniques are employed. The advantages of these techniques over the other methods are evaluated from the point of their validity in properly isolating features of hydrogeolocial interest are discussed. As these techniques take the advantage of spectral aspects of the images, these techniques have several limitations to meet the objectives. To discuss these limitations, a morphological transformation, which generally considers the structural aspects rather than spectral aspects from the image, are applied to provide comparisons between the results derived from spectral based and the structural based filtering techniques.
Hypergraph Based Feature Selection Technique for Medical Diagnosis.
Somu, Nivethitha; Raman, M R Gauthama; Kirthivasan, Kannan; Sriram, V S Shankar
2016-11-01
The impact of internet and information systems across various domains have resulted in substantial generation of multidimensional datasets. The use of data mining and knowledge discovery techniques to extract the original information contained in the multidimensional datasets play a significant role in the exploitation of complete benefit provided by them. The presence of large number of features in the high dimensional datasets incurs high computational cost in terms of computing power and time. Hence, feature selection technique has been commonly used to build robust machine learning models to select a subset of relevant features which projects the maximal information content of the original dataset. In this paper, a novel Rough Set based K - Helly feature selection technique (RSKHT) which hybridize Rough Set Theory (RST) and K - Helly property of hypergraph representation had been designed to identify the optimal feature subset or reduct for medical diagnostic applications. Experiments carried out using the medical datasets from the UCI repository proves the dominance of the RSKHT over other feature selection techniques with respect to the reduct size, classification accuracy and time complexity. The performance of the RSKHT had been validated using WEKA tool, which shows that RSKHT had been computationally attractive and flexible over massive datasets.
Sheppard, P S; Stevenson, J M; Graham, R B
2016-05-01
The objective of the present study was to determine if there is a sex-based difference in lifting technique across increasing-load conditions. Eleven male and 14 female participants (n = 25) with no previous history of low back disorder participated in the study. Participants completed freestyle, symmetric lifts of a box with handles from the floor to a table positioned at 50% of their height for five trials under three load conditions (10%, 20%, and 30% of their individual maximum isometric back strength). Joint kinematic data for the ankle, knee, hip, and lumbar and thoracic spine were collected using a two-camera Optotrak motion capture system. Joint angles were calculated using a three-dimensional Euler rotation sequence. Principal component analysis (PCA) and single component reconstruction were applied to assess differences in lifting technique across the entire waveforms. Thirty-two PCs were retained from the five joints and three axes in accordance with the 90% trace criterion. Repeated-measures ANOVA with a mixed design revealed no significant effect of sex for any of the PCs. This is contrary to previous research that used discrete points on the lifting curve to analyze sex-based differences, but agrees with more recent research using more complex analysis techniques. There was a significant effect of load on lifting technique for five PCs of the lower limb (PC1 of ankle flexion, knee flexion, and knee adduction, as well as PC2 and PC3 of hip flexion) (p < 0.005). However, there was no significant effect of load on the thoracic and lumbar spine. It was concluded that when load is standardized to individual back strength characteristics, males and females adopted a similar lifting technique. In addition, as load increased male and female participants changed their lifting technique in a similar manner. Copyright © 2016. Published by Elsevier Ltd.
Hu, Fei; Cheng, Yayun; Gui, Liangqi; Wu, Liang; Zhang, Xinyi; Peng, Xiaohui; Su, Jinlong
2016-11-01
The polarization properties of thermal millimeter-wave emission capture inherent information of objects, e.g., material composition, shape, and surface features. In this paper, a polarization-based material-classification technique using passive millimeter-wave polarimetric imagery is presented. Linear polarization ratio (LPR) is created to be a new feature discriminator that is sensitive to material type and to remove the reflected ambient radiation effect. The LPR characteristics of several common natural and artificial materials are investigated by theoretical and experimental analysis. Based on a priori information about LPR characteristics, the optimal range of incident angle and the classification criterion are discussed. Simulation and measurement results indicate that the presented classification technique is effective for distinguishing between metals and dielectrics. This technique suggests possible applications for outdoor metal target detection in open scenes.
The key technique study of a kind of personal navigation oriented LBS system
NASA Astrophysics Data System (ADS)
Yan, Lei; Zheng, Jianghua; Zhang, Xin; Peng, Chunhua; He, Lina
2005-11-01
With the integration of GIS, IT technology and wireless communication techniques, LBS is fast developing and caused wide concern. Personal navigation is the critical application of LBS. It has higher requirement of data quality, positioning accuracy and multi-model services. The study discusses the key techniques of a personal navigation oriented LBS system. As an example for service platform of China Unicom, NAVISTAR especially emphasizes the importance of spatial data organization. Based-on CDMA1X network, it adopts gpsOne\\MS-Assisted dynamic positioning technique, and puts forward a data organization solution to realize multi-scale representation.
The Triangle Technique: a new evidence-based educational tool for pediatric medication calculations.
Sredl, Darlene
2006-01-01
Many nursing student verbalize an aversion to mathematical concepts and experience math anxiety whenever a mathematical problem is confronted. Since nurses confront mathematical problems on a daily basis, they must learn to feel comfortable with their ability to perform these calculations correctly. The Triangle Technique, a new educational tool available to nurse educators, incorporates evidence-based concepts within a graphic model using visual, auditory, and kinesthetic learning styles to demonstrate pediatric medication calculations of normal therapeutic ranges. The theoretical framework for the technique is presented, as is a pilot study examining the efficacy of the educational tool. Statistically significant results obtained by Pearson's product-moment correlation indicate that students are better able to calculate accurate pediatric therapeutic dosage ranges after participation in the educational intervention of learning the Triangle Technique.
NASA Astrophysics Data System (ADS)
Soldner, Dominic; Brands, Benjamin; Zabihyan, Reza; Steinmann, Paul; Mergheim, Julia
2017-10-01
Computing the macroscopic material response of a continuum body commonly involves the formulation of a phenomenological constitutive model. However, the response is mainly influenced by the heterogeneous microstructure. Computational homogenisation can be used to determine the constitutive behaviour on the macro-scale by solving a boundary value problem at the micro-scale for every so-called macroscopic material point within a nested solution scheme. Hence, this procedure requires the repeated solution of similar microscopic boundary value problems. To reduce the computational cost, model order reduction techniques can be applied. An important aspect thereby is the robustness of the obtained reduced model. Within this study reduced-order modelling (ROM) for the geometrically nonlinear case using hyperelastic materials is applied for the boundary value problem on the micro-scale. This involves the Proper Orthogonal Decomposition (POD) for the primary unknown and hyper-reduction methods for the arising nonlinearity. Therein three methods for hyper-reduction, differing in how the nonlinearity is approximated and the subsequent projection, are compared in terms of accuracy and robustness. Introducing interpolation or Gappy-POD based approximations may not preserve the symmetry of the system tangent, rendering the widely used Galerkin projection sub-optimal. Hence, a different projection related to a Gauss-Newton scheme (Gauss-Newton with Approximated Tensors- GNAT) is favoured to obtain an optimal projection and a robust reduced model.
Wood lens design philosophy based on a binary additive manufacturing technique
NASA Astrophysics Data System (ADS)
Marasco, Peter L.; Bailey, Christopher
2016-04-01
Using additive manufacturing techniques in optical engineering to construct a gradient index (GRIN) optic may overcome a number of limitations of GRIN technology. Such techniques are maturing quickly, yielding additional design degrees of freedom for the engineer. How best to employ these degrees of freedom is not completely clear at this time. This paper describes a preliminary design philosophy, including assumptions, pertaining to a particular printing technique for GRIN optics. It includes an analysis based on simulation and initial component measurement.
Vision based techniques for rotorcraft low altitude flight
NASA Technical Reports Server (NTRS)
Sridhar, Banavar; Suorsa, Ray; Smith, Philip
1991-01-01
An overview of research in obstacle detection at NASA Ames Research Center is presented. The research applies techniques from computer vision to automation of rotorcraft navigation. The development of a methodology for detecting the range to obstacles based on the maximum utilization of passive sensors is emphasized. The development of a flight and image data base for verification of vision-based algorithms, and a passive ranging methodology tailored to the needs of helicopter flight are discussed. Preliminary results indicate that it is possible to obtain adequate range estimates except at regions close to the FOE. Closer to the FOE, the error in range increases since the magnitude of the disparity gets smaller, resulting in a low SNR.
Olyaeemanesh, Alireza; Bavandpour, Elahe; Mobinizadeh, Mohammadreza; Ashrafinia, Mansoor; Bavandpour, Maryam; Nouhi, Mojtaba
2017-01-01
Background: Caesarean section (C-section) is the most common surgery among women worldwide, and the global rate of this surgical procedure has been continuously rising. Hence, it is significantly crucial to develop and apply highly effective and safe caesarean section techniques. In this review study, we aimed at assessing the safety and effectiveness of the Joel-Cohen-based technique and comparing the results with the transverse Pfannenstiel incision for C-section. Methods: In this study, various reliable databases such as the PubMed Central, COCHRANE, DARE, and Ovid MEDLINE were targeted. Reviews, systematic reviews, and randomized clinical trial studies comparing the Joel-Cohen-based technique and the transverse Pfannenstiel incision were selected based on the inclusion criteria. Selected studies were checked by 2 independent reviewers based on the inclusion criteria, and the quality of these studies was assessed. Then, their data were extracted and analyzed. Results: Five randomized clinical trial studies met the inclusion criteria. According to the exiting evidence, statistical results of the Joel-Cohen-based technique showed that this technique is more effective compared to the transverse Pfannenstiel incision. Metaanalysis results of the 3 outcomes were as follow: operation time (5 trials, 764 women; WMD -9.78; 95% CI:-14.49-5.07 minutes, p<0.001), blood loss (3 trials, 309 women; WMD -53.23ml; 95% –CI: 90.20-16.26 ml, p= 0.004), and post-operative hospital stay (3 trials, 453 women; WMD -.69 day; 95% CI: 1.4-0.03 day, p<0.001). Statistical results revealed a significant difference between the 2 techniques. Conclusion: According to the literature, despite having a number of side effects, the Joel-Cohen-based technique is generally more effective than the Pfannenstiel incision technique. In addition, it was recommended that the Joel-Cohen-based technique be used as a replacement for the Pfannenstiel incision technique according to the surgeons
Olyaeemanesh, Alireza; Bavandpour, Elahe; Mobinizadeh, Mohammadreza; Ashrafinia, Mansoor; Bavandpour, Maryam; Nouhi, Mojtaba
2017-01-01
Background: Caesarean section (C-section) is the most common surgery among women worldwide, and the global rate of this surgical procedure has been continuously rising. Hence, it is significantly crucial to develop and apply highly effective and safe caesarean section techniques. In this review study, we aimed at assessing the safety and effectiveness of the Joel-Cohen-based technique and comparing the results with the transverse Pfannenstiel incision for C-section. Methods: In this study, various reliable databases such as the PubMed Central, COCHRANE, DARE, and Ovid MEDLINE were targeted. Reviews, systematic reviews, and randomized clinical trial studies comparing the Joel-Cohen-based technique and the transverse Pfannenstiel incision were selected based on the inclusion criteria. Selected studies were checked by 2 independent reviewers based on the inclusion criteria, and the quality of these studies was assessed. Then, their data were extracted and analyzed. Results: Five randomized clinical trial studies met the inclusion criteria. According to the exiting evidence, statistical results of the Joel-Cohen-based technique showed that this technique is more effective compared to the transverse Pfannenstiel incision. Metaanalysis results of the 3 outcomes were as follow: operation time (5 trials, 764 women; WMD -9.78; 95% CI:-14.49-5.07 minutes, p<0.001), blood loss (3 trials, 309 women; WMD -53.23ml; 95% -CI: 90.20-16.26 ml, p= 0.004), and post-operative hospital stay (3 trials, 453 women; WMD -.69 day; 95% CI: 1.4-0.03 day, p<0.001). Statistical results revealed a significant difference between the 2 techniques. Conclusion: According to the literature, despite having a number of side effects, the Joel-Cohen-based technique is generally more effective than the Pfannenstiel incision technique. In addition, it was recommended that the Joel-Cohen-based technique be used as a replacement for the Pfannenstiel incision technique according to the surgeons' preferences
Applications of Advanced, Waveform Based AE Techniques for Testing Composite Materials
NASA Technical Reports Server (NTRS)
Prosser, William H.
1996-01-01
Advanced, waveform based acoustic emission (AE) techniques have been previously used to evaluate damage progression in laboratory tests of composite coupons. In these tests, broad band, high fidelity acoustic sensors were used to detect signals which were then digitized and stored for analysis. Analysis techniques were based on plate mode wave propagation characteristics. This approach, more recently referred to as Modal AE, provides an enhanced capability to discriminate and eliminate noise signals from those generated by damage mechanisms. This technique also allows much more precise source location than conventional, threshold crossing arrival time determination techniques. To apply Modal AE concepts to the interpretation of AE on larger composite structures, the effects of wave propagation over larger distances and through structural complexities must be well characterized and understood. In this research, measurements were made of the attenuation of the extensional and flexural plate mode components of broad band simulated AE signals in large composite panels. As these materials have applications in a cryogenic environment, the effects of cryogenic insulation on the attenuation of plate mode AE signals were also documented.
Mobility based key management technique for multicast security in mobile ad hoc networks.
Madhusudhanan, B; Chitra, S; Rajan, C
2015-01-01
In MANET multicasting, forward and backward secrecy result in increased packet drop rate owing to mobility. Frequent rekeying causes large message overhead which increases energy consumption and end-to-end delay. Particularly, the prevailing group key management techniques cause frequent mobility and disconnections. So there is a need to design a multicast key management technique to overcome these problems. In this paper, we propose the mobility based key management technique for multicast security in MANET. Initially, the nodes are categorized according to their stability index which is estimated based on the link availability and mobility. A multicast tree is constructed such that for every weak node, there is a strong parent node. A session key-based encryption technique is utilized to transmit a multicast data. The rekeying process is performed periodically by the initiator node. The rekeying interval is fixed depending on the node category so that this technique greatly minimizes the rekeying overhead. By simulation results, we show that our proposed approach reduces the packet drop rate and improves the data confidentiality.
Dash, Rajendra Narayan; Mohammed, Habibuddin; Humaira, Touseef
2016-01-01
We studied the application of Taguchi orthogonal array (TOA) design during the development of an isocratic stability indicating HPLC method for glimepiride as per TOA design; twenty-seven experiments were conducted by varying six chromatographic factors. Percentage of organic phase was the most significant (p < 0.001) on retention time, while buffer pH had the most significant (p < 0.001) effect on tailing factor and theoretical plates. TOA design has shortcoming, which identifies the only linear effect, while ignoring the quadratic and interaction effects. Hence, a response surface model for each response was created including the linear, quadratic and interaction terms. The developed models for each response found to be well predictive bearing an acceptable adjusted correlation coefficient (0.9152 for retention time, 0.8985 for tailing factor and 0.8679 for theoretical plates). The models were found to be significant (p < 0.001) having a high F value for each response (15.76 for retention time, 13.12 for tailing factor and 9.99 for theoretical plates). The optimal chromatographic condition uses acetonitrile - potassium dihydrogen phosphate (pH 4.0; 30 mM) (50:50, v/v) as the mobile phase. The temperature, flow rate and injection volume were selected as 35 ± 2 °C, 1.0 mL min(-1) and 20 μL respectively. The method was validated as per ICH guidelines and was found to be specific for analyzing glimepiride from a novel supersaturatable self-nanoemulsifying formulation.
NASA Astrophysics Data System (ADS)
El-Moslamy, Shahira H.; Elkady, Marwa F.; Rezk, Ahmed H.; Abdel-Fattah, Yasser R.
2017-03-01
Development of reliable and low-cost requirement for large-scale eco-friendly biogenic synthesis of metallic nanoparticles is an important step for industrial applications of bionanotechnology. In the present study, the mycosynthesis of spherical nano-Ag (12.7 ± 0.8 nm) from extracellular filtrate of local endophytic T. harzianum SYA.F4 strain which have interested mixed bioactive metabolites (alkaloids, flavonoids, tannins, phenols, nitrate reductase (320 nmol/hr/ml), carbohydrate (25 μg/μl) and total protein concentration (2.5 g/l) was reported. Industrial mycosynthesis of nano-Ag can be induced with different characters depending on the fungal cultivation and physical conditions. Taguchi design was applied to improve the physicochemical conditions for nano-Ag production, and the optimum conditions which increased its mass weight 3 times larger than a basal condition were as follows: AgNO3 (0.01 M), diluted reductant (10 v/v, pH 5) and incubated at 30 °C, 200 rpm for 24 hr. Kinetic conversion rates in submerged batch cultivation in 7 L stirred tank bioreactor on using semi-defined cultivation medium was as follows: the maximum biomass production (Xmax) and maximum nano-Ag mass weight (Pmax) calculated (60.5 g/l and 78.4 g/l respectively). The best nano-Ag concentration that formed large inhibition zones was 100 μg/ml which showed against A.alternate (43 mm) followed by Helminthosporium sp. (35 mm), Botrytis sp. (32 mm) and P. arenaria (28 mm).
Application of Petri net based analysis techniques to signal transduction pathways
Sackmann, Andrea; Heiner, Monika; Koch, Ina
2006-01-01
Background Signal transduction pathways are usually modelled using classical quantitative methods, which are based on ordinary differential equations (ODEs). However, some difficulties are inherent in this approach. On the one hand, the kinetic parameters involved are often unknown and have to be estimated. With increasing size and complexity of signal transduction pathways, the estimation of missing kinetic data is not possible. On the other hand, ODEs based models do not support any explicit insights into possible (signal-) flows within the network. Moreover, a huge amount of qualitative data is available due to high-throughput techniques. In order to get information on the systems behaviour, qualitative analysis techniques have been developed. Applications of the known qualitative analysis methods concern mainly metabolic networks. Petri net theory provides a variety of established analysis techniques, which are also applicable to signal transduction models. In this context special properties have to be considered and new dedicated techniques have to be designed. Methods We apply Petri net theory to model and analyse signal transduction pathways first qualitatively before continuing with quantitative analyses. This paper demonstrates how to build systematically a discrete model, which reflects provably the qualitative biological behaviour without any knowledge of kinetic parameters. The mating pheromone response pathway in Saccharomyces cerevisiae serves as case study. Results We propose an approach for model validation of signal transduction pathways based on the network structure only. For this purpose, we introduce the new notion of feasible t-invariants, which represent minimal self-contained subnets being active under a given input situation. Each of these subnets stands for a signal flow in the system. We define maximal common transition sets (MCT-sets), which can be used for t-invariant examination and net decomposition into smallest biologically
He, Wanzhong; Kivork, Christine; Machinani, Suman; Morphew, Mary K.; Gail, Anna M.; Tesar, Devin B.; Tiangco, Noreen E.; McIntosh, J. Richard; Bjorkman, Pamela J.
2007-01-01
We have developed methods to locate individual ligands that can be used for electron microscopy studies of dynamic events during endocytosis and subsequent intracellular trafficking. The methods are based on enlargement of 1.4 nm Nanogold attached to an endocytosed ligand. Nanogold, a small label that does not induce misdirection of ligand-receptor complexes, is ideal for labeling ligands endocytosed by live cells, but is too small to be routinely located in cells by electron microscopy. Traditional pre-embedding enhancement protocols to enlarge Nanogold are not compatible with high pressure freezing/freeze substitution fixation (HPF/FSF), the most accurate method to preserve ultrastructure and dynamic events during trafficking. We have developed an improved enhancement procedure for chemically-fixed samples that reduced autonucleation, and a new pre-embedding gold-enlarging technique for HPF/FSF samples that preserved contrast and ultrastructure and can be used for high-resolution tomography. We evaluated our methods using labeled Fc as a ligand for the neonatal Fc receptor. Attachment of Nanogold to Fc did not interfere with receptor binding or uptake, and gold-labeled Fc could be specifically enlarged to allow identification in 2D projections and in tomograms. These methods should be broadly applicable to many endocytosis and transcytosis studies. PMID:17723309
Fu, Yu; Pedrini, Giancarlo
2014-01-01
In recent years, optical interferometry-based techniques have been widely used to perform noncontact measurement of dynamic deformation in different industrial areas. In these applications, various physical quantities need to be measured in any instant and the Nyquist sampling theorem has to be satisfied along the time axis on each measurement point. Two types of techniques were developed for such measurements: one is based on high-speed cameras and the other uses a single photodetector. The limitation of the measurement range along the time axis in camera-based technology is mainly due to the low capturing rate, while the photodetector-based technology can only do the measurement on a single point. In this paper, several aspects of these two technologies are discussed. For the camera-based interferometry, the discussion includes the introduction of the carrier, the processing of the recorded images, the phase extraction algorithms in various domains, and how to increase the temporal measurement range by using multiwavelength techniques. For the detector-based interferometry, the discussion mainly focuses on the single-point and multipoint laser Doppler vibrometers and their applications for measurement under extreme conditions. The results show the effort done by researchers for the improvement of the measurement capabilities using interferometry-based techniques to cover the requirements needed for the industrial applications. PMID:24963503
Geographically correlated errors observed from a laser-based short-arc technique
NASA Astrophysics Data System (ADS)
Bonnefond, P.; Exertier, P.; Barlier, F.
1999-07-01
The laser-based short-arc technique has been developed in order to avoid local errors which affect the dynamical orbit computation, such as those due to mismodeling in the geopotential. It is based on a geometric method and consists in fitting short arcs (about 4000 km), issued from a global orbit, with satellite laser ranging tracking measurements from a ground station network. Ninety-two TOPEX/Poseidon (T/P) cycles of laser-based short-arc orbits have then been compared to JGM-2 and JGM-3 T/P orbits computed by the Precise Orbit Determination (POD) teams (Service d'Orbitographie Doris/Centre National d'Etudes Spatiales and Goddard Space Flight Center/NASA) over two areas: (1) the Mediterranean area and (2) a part of the Pacific (including California and Hawaii) called hereafter the U.S. area. Geographically correlated orbit errors in these areas are clearly evidenced: for example, -2.6 cm and +0.7 cm for the Mediterranean and U.S. areas, respectively, relative to JGM-3 orbits. However, geographically correlated errors (GCE) which are commonly linked to errors in the gravity model, can also be due to systematic errors in the reference frame and/or to biases in the tracking measurements. The short-arc technique being very sensitive to such error sources, our analysis however demonstrates that the induced geographical systematic effects are at the level of 1-2 cm on the radial orbit component. Results are also compared with those obtained with the GPS-based reduced dynamic technique. The time-dependent part of GCE has also been studied. Over 6 years of T/P data, coherent signals in the radial component of T/P Precise Orbit Ephemeris (POE) are clearly evidenced with a time period of about 6 months. In addition, impact of time varying-error sources coming from the reference frame and the tracking data accuracy has been analyzed, showing a possible linear trend of about 0.5-1 mm/yr in the radial component of T/P POE.
Three-Dimensional Inverse Transport Solver Based on Compressive Sensing Technique
NASA Astrophysics Data System (ADS)
Cheng, Yuxiong; Wu, Hongchun; Cao, Liangzhi; Zheng, Youqi
2013-09-01
According to the direct exposure measurements from flash radiographic image, a compressive sensing-based method for three-dimensional inverse transport problem is presented. The linear absorption coefficients and interface locations of objects are reconstructed directly at the same time. It is always very expensive to obtain enough measurements. With limited measurements, compressive sensing sparse reconstruction technique orthogonal matching pursuit is applied to obtain the sparse coefficients by solving an optimization problem. A three-dimensional inverse transport solver is developed based on a compressive sensing-based technique. There are three features in this solver: (1) AutoCAD is employed as a geometry preprocessor due to its powerful capacity in graphic. (2) The forward projection matrix rather than Gauss matrix is constructed by the visualization tool generator. (3) Fourier transform and Daubechies wavelet transform are adopted to convert an underdetermined system to a well-posed system in the algorithm. Simulations are performed and numerical results in pseudo-sine absorption problem, two-cube problem and two-cylinder problem when using compressive sensing-based solver agree well with the reference value.
Study of synthesis techniques for insensitive aircraft control systems
NASA Technical Reports Server (NTRS)
Harvey, C. A.; Pope, R. E.
1977-01-01
Insensitive flight control system design criteria was defined in terms of maximizing performance (handling qualities, RMS gust response, transient response, stability margins) over a defined parameter range. Wing load alleviation for the C-5A was chosen as a design problem. The C-5A model was a 79-state, two-control structure with uncertainties assumed to exist in dynamic pressure, structural damping and frequency, and the stability derivative, M sub w. Five new techniques (mismatch estimation, uncertainty weighting, finite dimensional inverse, maximum difficulty, dual Lyapunov) were developed. Six existing techniques (additive noise, minimax, multiplant, sensitivity vector augmentation, state dependent noise, residualization) and the mismatch estimation and uncertainty weighting techniques were synthesized and evaluated on the design example. Evaluation and comparison of these six techniques indicated that the minimax and the uncertainty weighting techniques were superior to the other six, and of these two, uncertainty weighting has lower computational requirements. Techniques based on the three remaining new concepts appear promising and are recommended for further research.
Tomographic techniques for the study of exceptionally preserved fossils
Sutton, Mark D
2008-01-01
Three-dimensional fossils, especially those preserving soft-part anatomy, are a rich source of palaeontological information; they can, however, be difficult to work with. Imaging of serial planes through an object (tomography) allows study of both the inside and outside of three-dimensional fossils. Tomography may be performed using physical grinding or sawing coupled with photography, through optical techniques of serial focusing, or using a variety of scanning technologies such as neutron tomography, magnetic resonance imaging and most usefully X-ray computed tomography. This latter technique is applicable at a variety of scales, and when combined with a synchrotron X-ray source can produce very high-quality data that may be augmented by phase-contrast information to enhance contrast. Tomographic data can be visualized in several ways, the most effective of which is the production of isosurface-based ‘virtual fossils’ that can be manipulated and dissected interactively. PMID:18426749
Epoxy matrix composites filled with micro-sized LD sludge: wear characterization and analysis
NASA Astrophysics Data System (ADS)
Purohit, Abhilash; Satapathy, Alok
2016-02-01
Owing to the very high cost of conventional filler materials in polymer composites, exploring the possibility of using low cost minerals and industrial wastes for this purpose has become the need of the hour. In view of this, the present work includes the development and the wear performance evaluation of a new class of composites consisting of epoxy and microsized LD sludge. LD sludge or the Linz-Donawitz Sludge (LDS) are the fine solid particles recovered after wet cleaning of the gas emerging from LD convertors during steel making. Epoxy composites filled with different proportions (0, 5, 10, 15 and 20 wt %) of LDS are fabricated by conventional hand lay-up technique. Dry sliding wear trials are performed on the composite specimens under different test conditions as per ASTM G 99 following a design of experiment approach based on Taguchi's orthogonal arrays. The Taguchi approach leads to the recognition of most powerful variables that predominantly control the wear rate. This parametric analysis reveals that LDS content and sliding velocity affects the specific wear rate more significantly than normal load and sliding distance. Furthermore with increase in LDS content specific wear rate of the composite decreases for a constant sliding velocity. The sliding wear behavior of these composites under an extended range of test conditions is predicted by a model based on the artificial neural network (ANN).
Improvement in QEPAS system utilizing a second harmonic based wavelength calibration technique
NASA Astrophysics Data System (ADS)
Zhang, Qinduan; Chang, Jun; Wang, Fupeng; Wang, Zongliang; Xie, Yulei; Gong, Weihua
2018-05-01
A simple laser wavelength calibration technique, based on second harmonic signal, is demonstrated in this paper to improve the performance of quartz enhanced photoacoustic spectroscopy (QEPAS) gas sensing system, e.g. improving the signal to noise ratio (SNR), detection limit and long-term stability. Constant current, corresponding to the gas absorption line, combining f/2 frequency sinusoidal signal are used to drive the laser (constant driving mode), a software based real-time wavelength calibration technique is developed to eliminate the wavelength drift due to ambient fluctuations. Compared to conventional wavelength modulation spectroscopy (WMS), this method allows lower filtering bandwidth and averaging algorithm applied to QEPAS system, improving SNR and detection limit. In addition, the real-time wavelength calibration technique guarantees the laser output is modulated steadily at gas absorption line. Water vapor is chosen as an objective gas to evaluate its performance compared to constant driving mode and conventional WMS system. The water vapor sensor was designed insensitive to the incoherent external acoustic noise by the numerical averaging technique. As a result, the SNR increases 12.87 times in wavelength calibration technique based system compared to conventional WMS system. The new system achieved a better linear response (R2 = 0 . 9995) in concentration range from 300 to 2000 ppmv, and achieved a minimum detection limit (MDL) of 630 ppbv.
Diagnostic techniques in deflagration and detonation studies.
Proud, William G; Williamson, David M; Field, John E; Walley, Stephen M
2015-12-01
Advances in experimental, high-speed techniques can be used to explore the processes occurring within energetic materials. This review describes techniques used to study a wide range of processes: hot-spot formation, ignition thresholds, deflagration, sensitivity and finally the detonation process. As this is a wide field the focus will be on small-scale experiments and quantitative studies. It is important that such studies are linked to predictive models, which inform the experimental design process. The stimuli range includes, thermal ignition, drop-weight, Hopkinson Bar and Plate Impact studies. Studies made with inert simulants are also included as these are important in differentiating between reactive response and purely mechanical behaviour.
NASA Astrophysics Data System (ADS)
Kleshnin, Mikhail; Orlova, Anna; Kirillin, Mikhail; Golubiatnikov, German; Turchin, Ilya
2017-07-01
A new approach to optical measuring blood oxygen saturation was developed and implemented. This technique is based on an original three-stage algorithm for reconstructing the relative concentration of biological chromophores (hemoglobin, water, lipids) from the measured spectra of diffusely scattered light at different distances from the probing radiation source. The numerical experiments and approbation of the proposed technique on a biological phantom have shown the high reconstruction accuracy and the possibility of correct calculation of hemoglobin oxygenation in the presence of additive noise and calibration errors. The obtained results of animal studies have agreed with the previously published results of other research groups and demonstrated the possibility to apply the developed technique to monitor oxygen saturation in tumor tissue.
Lens-free shadow image based high-throughput continuous cell monitoring technique.
Jin, Geonsoo; Yoo, In-Hwa; Pack, Seung Pil; Yang, Ji-Woon; Ha, Un-Hwan; Paek, Se-Hwan; Seo, Sungkyu
2012-01-01
A high-throughput continuous cell monitoring technique which does not require any labeling reagents or destruction of the specimen is demonstrated. More than 6000 human alveolar epithelial A549 cells are monitored for up to 72 h simultaneously and continuously with a single digital image within a cost and space effective lens-free shadow imaging platform. In an experiment performed within a custom built incubator integrated with the lens-free shadow imaging platform, the cell nucleus division process could be successfully characterized by calculating the signal-to-noise ratios (SNRs) and the shadow diameters (SDs) of the cell shadow patterns. The versatile nature of this platform also enabled a single cell viability test followed by live cell counting. This study firstly shows that the lens-free shadow imaging technique can provide a continuous cell monitoring without any staining/labeling reagent and destruction of the specimen. This high-throughput continuous cell monitoring technique based on lens-free shadow imaging may be widely utilized as a compact, low-cost, and high-throughput cell monitoring tool in the fields of drug and food screening or cell proliferation and viability testing. Copyright © 2012 Elsevier B.V. All rights reserved.
A novel mesh processing based technique for 3D plant analysis
2012-01-01
Background In recent years, imaging based, automated, non-invasive, and non-destructive high-throughput plant phenotyping platforms have become popular tools for plant biology, underpinning the field of plant phenomics. Such platforms acquire and record large amounts of raw data that must be accurately and robustly calibrated, reconstructed, and analysed, requiring the development of sophisticated image understanding and quantification algorithms. The raw data can be processed in different ways, and the past few years have seen the emergence of two main approaches: 2D image processing and 3D mesh processing algorithms. Direct image quantification methods (usually 2D) dominate the current literature due to comparative simplicity. However, 3D mesh analysis provides the tremendous potential to accurately estimate specific morphological features cross-sectionally and monitor them over-time. Result In this paper, we present a novel 3D mesh based technique developed for temporal high-throughput plant phenomics and perform initial tests for the analysis of Gossypium hirsutum vegetative growth. Based on plant meshes previously reconstructed from multi-view images, the methodology involves several stages, including morphological mesh segmentation, phenotypic parameters estimation, and plant organs tracking over time. The initial study focuses on presenting and validating the accuracy of the methodology on dicotyledons such as cotton but we believe the approach will be more broadly applicable. This study involved applying our technique to a set of six Gossypium hirsutum (cotton) plants studied over four time-points. Manual measurements, performed for each plant at every time-point, were used to assess the accuracy of our pipeline and quantify the error on the morphological parameters estimated. Conclusion By directly comparing our automated mesh based quantitative data with manual measurements of individual stem height, leaf width and leaf length, we obtained the mean
van Til, Janine; Groothuis-Oudshoorn, Catharina; Lieferink, Marijke; Dolan, James; Goetghebeur, Mireille
2014-01-01
There is an increased interest in the use of multi-criteria decision analysis (MCDA) to support regulatory and reimbursement decision making. The EVIDEM framework was developed to provide pragmatic multi-criteria decision support in health care, to estimate the value of healthcare interventions, and to aid in priority-setting. The objectives of this study were to test 1) the influence of different weighting techniques on the overall outcome of an MCDA exercise, 2) the discriminative power in weighting different criteria of such techniques, and 3) whether different techniques result in similar weights in weighting the criteria set proposed by the EVIDEM framework. A sample of 60 Dutch and Canadian students participated in the study. Each student used an online survey to provide weights for 14 criteria with two different techniques: a five-point rating scale and one of the following techniques selected randomly: ranking, point allocation, pairwise comparison and best worst scaling. The results of this study indicate that there is no effect of differences in weights on value estimates at the group level. On an individual level, considerable differences in criteria weights and rank order occur as a result of the weight elicitation method used, and the ability of different techniques to discriminate in criteria importance. Of the five techniques tested, the pair-wise comparison of criteria has the highest ability to discriminate in weights when fourteen criteria are compared. When weights are intended to support group decisions, the choice of elicitation technique has negligible impact on criteria weights and the overall value of an innovation. However, when weights are used to support individual decisions, the choice of elicitation technique influences outcome and studies that use dissimilar techniques cannot be easily compared. Weight elicitation through pairwise comparison of criteria is preferred when taking into account its superior ability to discriminate between
ANALYSIS OF SAMPLING TECHNIQUES FOR IMBALANCED DATA: AN N=648 ADNI STUDY
Dubey, Rashmi; Zhou, Jiayu; Wang, Yalin; Thompson, Paul M.; Ye, Jieping
2013-01-01
Many neuroimaging applications deal with imbalanced imaging data. For example, in Alzheimer’s Disease Neuroimaging Initiative (ADNI) dataset, the mild cognitive impairment (MCI) cases eligible for the study are nearly two times the Alzheimer’s disease (AD) patients for structural magnetic resonance imaging (MRI) modality and six times the control cases for proteomics modality. Constructing an accurate classifier from imbalanced data is a challenging task. Traditional classifiers that aim to maximize the overall prediction accuracy tend to classify all data into the majority class. In this paper, we study an ensemble system of feature selection and data sampling for the class imbalance problem. We systematically analyze various sampling techniques by examining the efficacy of different rates and types of undersampling, oversampling, and a combination of over and under sampling approaches. We thoroughly examine six widely used feature selection algorithms to identify significant biomarkers and thereby reduce the complexity of the data. The efficacy of the ensemble techniques is evaluated using two different classifiers including Random Forest and Support Vector Machines based on classification accuracy, area under the receiver operating characteristic curve (AUC), sensitivity, and specificity measures. Our extensive experimental results show that for various problem settings in ADNI, (1). a balanced training set obtained with K-Medoids technique based undersampling gives the best overall performance among different data sampling techniques and no sampling approach; and (2). sparse logistic regression with stability selection achieves competitive performance among various feature selection algorithms. Comprehensive experiments with various settings show that our proposed ensemble model of multiple undersampled datasets yields stable and promising results. PMID:24176869
Towards a balanced software team formation based on Belbin team role using fuzzy technique
NASA Astrophysics Data System (ADS)
Omar, Mazni; Hasan, Bikhtiyar; Ahmad, Mazida; Yasin, Azman; Baharom, Fauziah; Mohd, Haslina; Darus, Norida Muhd
2016-08-01
In software engineering (SE), team roles play significant impact in determining the project success. To ensure the optimal outcome of the project the team is working on, it is essential to ensure that the team members are assigned to the right role with the right characteristics. One of the prevalent team roles is Belbin team role. A successful team must have a balance of team roles. Thus, this study demonstrates steps taken to determine balance of software team formation based on Belbin team role using fuzzy technique. Fuzzy technique was chosen because it allows analyzing of imprecise data and classifying selected criteria. In this study, two roles in Belbin team role, which are Shaper (Sh) and Plant (Pl) were chosen to assign the specific role in software team. Results show that the technique is able to be used for determining the balance of team roles. Future works will focus on the validation of the proposed method by using empirical data in industrial setting.
N-isopropylacrylamide-based fine-dispersed thermosensitive ferrogels obtained via in-situ technique.
Korotych, O; Samchenko, Yu; Boldeskul, I; Ulberg, Z; Zholobak, N; Sukhodub, L
2013-03-01
Thermosensitive hydrogels with magnetic properties (ferrogels) are very promising for medical application, first of all, for the design of targeted delivery systems with controlled release of drugs and for magnetic hyperthermia and chemotherapy treatment of cancer. These magnetic hydrogels could be obtained using diverse techniques: ex- and in-situ syntheses. The present work is devoted to the study of magnetite (Fe(3)O(4)) formation inside the nanoreactors of (co)polymeric hydrogels. Polymeric templates (hydrogel films and fine-dispersed hydrogels) used for obtaining ferrogels were based on acrylic monomers: thermosensitive N-isopropylacrylamide, and hydrophilic acrylamide. Covalent cross-linking was accomplished using bifunctional monomer N,N'-methylenebisacrylamide. Influence of hydrophilic-lipophilic balance of polymeric templates and concentration of iron cations on the magnetite formation were investigated along with the development of ferrogel preparation technique. Cytotoxicity, physical and chemical properties of obtained magnetic hydrogels have been studied in this work. Copyright © 2012 Elsevier B.V. All rights reserved.
Validation techniques for fault emulation of SRAM-based FPGAs
Quinn, Heather; Wirthlin, Michael
2015-08-07
A variety of fault emulation systems have been created to study the effect of single-event effects (SEEs) in static random access memory (SRAM) based field-programmable gate arrays (FPGAs). These systems are useful for augmenting radiation-hardness assurance (RHA) methodologies for verifying the effectiveness for mitigation techniques; understanding error signatures and failure modes in FPGAs; and failure rate estimation. For radiation effects researchers, it is important that these systems properly emulate how SEEs manifest in FPGAs. If the fault emulation systems does not mimic the radiation environment, the system will generate erroneous data and incorrect predictions of behavior of the FPGA inmore » a radiation environment. Validation determines whether the emulated faults are reasonable analogs to the radiation-induced faults. In this study we present methods for validating fault emulation systems and provide several examples of validated FPGA fault emulation systems.« less
Optical technique to study the impact of heavy rain on aircraft performance
NASA Technical Reports Server (NTRS)
Hess, C. F.; Li, F.
1985-01-01
A laser based technique was investigated and shown to have the potential to obtain measurements of the size and velocity of water droplets used in a wind tunnel to simulate rain. A theoretical model was developed which included some simple effects due to droplet nonsphericity. Parametric studies included the variation of collection distance (up to 4 m), angle of collection, effect of beam interference by the spray, and droplet shape. Accurate measurements were obtained under extremely high liquid water content and spray interference. The technique finds applications in the characterization of two phase flows where the size and velocity of particles are needed.
Compression technique for large statistical data bases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Eggers, S.J.; Olken, F.; Shoshani, A.
1981-03-01
The compression of large statistical databases is explored and are proposed for organizing the compressed data, such that the time required to access the data is logarithmic. The techniques exploit special characteristics of statistical databases, namely, variation in the space required for the natural encoding of integer attributes, a prevalence of a few repeating values or constants, and the clustering of both data of the same length and constants in long, separate series. The techniques are variations of run-length encoding, in which modified run-lengths for the series are extracted from the data stream and stored in a header, which ismore » used to form the base level of a B-tree index into the database. The run-lengths are cumulative, and therefore the access time of the data is logarithmic in the size of the header. The details of the compression scheme and its implementation are discussed, several special cases are presented, and an analysis is given of the relative performance of the various versions.« less
Jabeen, Safia; Mehmood, Zahid; Mahmood, Toqeer; Saba, Tanzila; Rehman, Amjad; Mahmood, Muhammad Tariq
2018-01-01
For the last three decades, content-based image retrieval (CBIR) has been an active research area, representing a viable solution for retrieving similar images from an image repository. In this article, we propose a novel CBIR technique based on the visual words fusion of speeded-up robust features (SURF) and fast retina keypoint (FREAK) feature descriptors. SURF is a sparse descriptor whereas FREAK is a dense descriptor. Moreover, SURF is a scale and rotation-invariant descriptor that performs better in the case of repeatability, distinctiveness, and robustness. It is robust to noise, detection errors, geometric, and photometric deformations. It also performs better at low illumination within an image as compared to the FREAK descriptor. In contrast, FREAK is a retina-inspired speedy descriptor that performs better for classification-based problems as compared to the SURF descriptor. Experimental results show that the proposed technique based on the visual words fusion of SURF-FREAK descriptors combines the features of both descriptors and resolves the aforementioned issues. The qualitative and quantitative analysis performed on three image collections, namely Corel-1000, Corel-1500, and Caltech-256, shows that proposed technique based on visual words fusion significantly improved the performance of the CBIR as compared to the feature fusion of both descriptors and state-of-the-art image retrieval techniques.
Savage, Zoliakha; Munjal, Ramesh
2015-10-01
A 47-year-old male suffered a traumatic transtibial amputation; initial limb use was unsuccessful and the short tibial length was thought to be the cause of failure. The patient underwent gradual tibia lengthening using the Ilizarov technique and utilised a weight-bearing prosthesis to expedite bone growth and repair. The patient is now able to use his prosthesis successfully without aids. This case study demonstrates that combining a scientifically based surgical technique with a tailored rehabilitation approach had an improved outcome for the patient. This study reports one case where tibial lengthening using the Ilizarov technique combined with a collaborative team approach has enabled a patient to return to successful prosthetic limb use. © The International Society for Prosthetics and Orthotics 2014.
NASA Astrophysics Data System (ADS)
Singh, Jaswinder; Chauhan, Amit
2017-12-01
This study investigates the mechanical behavior of aluminum 2024 matrix composites reinforced with silicon carbide and red mud particles. The hybrid reinforcements were successfully incorporated into the alloy matrix using the stir casting process. An orthogonal array based on Taguchi's technique was used to acquire experimental data for mechanical properties (hardness and impact energy) of the composites. The analysis of variance (ANOVA) and response surface methodology (RSM) techniques were used to evaluate the influence of test parameters (reinforcement ratio, particle size and ageing time). The morphological analysis of the surfaces (fractured during impact tests) was conducted to identify the failure mechanism. Finally, a confirmation experiment was performed to check the adequacy of the developed model. The results indicate that the ageing time is the most effective parameter as far as the hardness of the hybrid composites is concerned. It has also been revealed that red mud wt.% has maximum influence on the impact energy characteristics of the hybrid composites. The study concludes that Al2024/SiC/red mud hybrid composites possess superior mechanical performance in comparison to pure alloy under optimized conditions.
The influence of polishing techniques on pre-polymerized CAD\\CAM acrylic resin denture bases
Alammari, Manal Rahma
2017-01-01
Background Lately, computer-aided design and computer-aided manufacturing (CAD/CAM) has broadly been successfully employed in dentistry. The CAD/CAM systems have recently become commercially available for fabrication of complete dentures, and are considered as an alternative technique to conventionally processed acrylic resin bases. However, they have not yet been fully investigated. Objective The purpose of this study was to inspect the effects of mechanical polishing and chemical polishing on the surface roughness (Ra) and contact angle (wettability) of heat-cured, auto-cured and CAD/CAM denture base acrylic resins. Methods This study was conducted at the Advanced Dental Research Laboratory Center of King Abdulaziz University from March to June 2017. Three denture base materials were selected: heat cure poly-methylmethacrylate resin, thermoplastic (polyamide resin) and (CAD\\CAM) denture base resin. Sixty specimens were prepared and divided into three groups, twenty in each. Each group was divided according to the polishing techniques into (Mech P) and (Chem P), ten specimens in each; surface roughness and wettability were investigated. Data were analyzed by SPSS version 22, using one-way ANOVA and Pearson coefficient. Results One-way analysis of variance (ANOVA) and post hoc tests were used for comparing the surface roughness values between three groups which revealed a statistical significant difference between them (p1<0.001). Heat-cured denture base material of (Group I) in both methods, showed the highest mean surface roughness value (2.44±0.07, 2.72±0.09, Mech P and Chem P respectively); while CAD\\CAM denture base material (group III) showed the least mean values (1.08±0.23, 1.39±0.31, Mech P and Chem P respectively). CAD/CAM showed the least contact angle in both polishing methods, which were statistically significant at 5% level (p=0.034 and p<0.001). Conclusion Mechanical polishing produced lower surface roughness of CAD\\CAM denture base resin with
The influence of polishing techniques on pre-polymerized CAD\\CAM acrylic resin denture bases.
Alammari, Manal Rahma
2017-10-01
Lately, computer-aided design and computer-aided manufacturing (CAD/CAM) has broadly been successfully employed in dentistry. The CAD/CAM systems have recently become commercially available for fabrication of complete dentures, and are considered as an alternative technique to conventionally processed acrylic resin bases. However, they have not yet been fully investigated. The purpose of this study was to inspect the effects of mechanical polishing and chemical polishing on the surface roughness (Ra) and contact angle (wettability) of heat-cured, auto-cured and CAD/CAM denture base acrylic resins. This study was conducted at the Advanced Dental Research Laboratory Center of King Abdulaziz University from March to June 2017. Three denture base materials were selected: heat cure poly-methylmethacrylate resin, thermoplastic (polyamide resin) and (CAD\\CAM) denture base resin. Sixty specimens were prepared and divided into three groups, twenty in each. Each group was divided according to the polishing techniques into (Mech P) and (Chem P), ten specimens in each; surface roughness and wettability were investigated. Data were analyzed by SPSS version 22, using one-way ANOVA and Pearson coefficient. One-way analysis of variance (ANOVA) and post hoc tests were used for comparing the surface roughness values between three groups which revealed a statistical significant difference between them (p 1 <0.001). Heat-cured denture base material of (Group I) in both methods, showed the highest mean surface roughness value (2.44±0.07, 2.72±0.09, Mech P and Chem P respectively); while CAD\\CAM denture base material (group III) showed the least mean values (1.08±0.23, 1.39±0.31, Mech P and Chem P respectively). CAD/CAM showed the least contact angle in both polishing methods, which were statistically significant at 5% level (p=0.034 and p<0.001). Mechanical polishing produced lower surface roughness of CAD\\CAM denture base resin with superior smooth surface compared to chemical
A new simple technique for improving the random properties of chaos-based cryptosystems
NASA Astrophysics Data System (ADS)
Garcia-Bosque, M.; Pérez-Resa, A.; Sánchez-Azqueta, C.; Celma, S.
2018-03-01
A new technique for improving the security of chaos-based stream ciphers has been proposed and tested experimentally. This technique manages to improve the randomness properties of the generated keystream by preventing the system to fall into short period cycles due to digitation. In order to test this technique, a stream cipher based on a Skew Tent Map algorithm has been implemented on a Virtex 7 FPGA. The randomness of the keystream generated by this system has been compared to the randomness of the keystream generated by the same system with the proposed randomness-enhancement technique. By subjecting both keystreams to the National Institute of Standards and Technology (NIST) tests, we have proved that our method can considerably improve the randomness of the generated keystreams. In order to incorporate our randomness-enhancement technique, only 41 extra slices have been needed, proving that, apart from effective, this method is also efficient in terms of area and hardware resources.
Comparison of two structured illumination techniques based on different 3D illumination patterns
NASA Astrophysics Data System (ADS)
Shabani, H.; Patwary, N.; Doblas, A.; Saavedra, G.; Preza, C.
2017-02-01
Manipulating the excitation pattern in optical microscopy has led to several super-resolution techniques. Among different patterns, the lateral sinusoidal excitation was used for the first demonstration of structured illumination microscopy (SIM), which provides the fastest SIM acquisition system (based on the number of raw images required) compared to the multi-spot illumination approach. Moreover, 3D patterns that include lateral and axial variations in the illumination have attracted more attention recently as they address resolution enhancement in three dimensions. A threewave (3W) interference technique based on coherent illumination has already been shown to provide super-resolution and optical sectioning in 3D-SIM. In this paper, we investigate a novel tunable technique that creates a 3D pattern from a set of multiple incoherently illuminated parallel slits that act as light sources for a Fresnel biprism. This setup is able to modulate the illumination pattern in the object space both axially and laterally with adjustable modulation frequencies. The 3D forward model for the new system is developed here to consider the effect of the axial modulation due to the 3D patterned illumination. The performance of 3D-SIM based on 3W interference and the tunable system are investigated in simulation and compared based on two different criteria. First, restored images obtained for both 3D-SIM systems using a generalized Wiener filter are compared to determine the effect of the illumination pattern on the reconstruction. Second, the effective frequency response of both systems is studied to determine the axial and lateral resolution enhancement that is obtained in each case.
NASA Astrophysics Data System (ADS)
Suresha, B. L.; Sumantha, H. S.; Salman, K. Mohammed; Pramod, N. G.; Abhiram, J.
2018-04-01
The ionization potential is usually found to be less in acid and more in base. The experiment proves that the ionization potential increases on dilution of acid to base and reduces from base to acid. The potential can be tailored according to the desired properties based on our choice of acid or base. The experimental study establishes a direct relationship between pH and electric potential. This work provides theoretical insights on the need for a basic media of pH 10 in chemical thin film growth techniques called Chemical Bath Deposition Techniques.
Engineering Design Education Program for Graduate School
NASA Astrophysics Data System (ADS)
Ohbuchi, Yoshifumi; Iida, Haruhiko
The new educational methods of engineering design have attempted to improve mechanical engineering education for graduate students in a way of the collaboration in education of engineer and designer. The education program is based on the lecture and practical exercises concerning the product design, and has engineering themes and design process themes, i.e. project management, QFD, TRIZ, robust design (Taguchi method) , ergonomics, usability, marketing, conception etc. At final exercise, all students were able to design new product related to their own research theme by applying learned knowledge and techniques. By the method of engineering design education, we have confirmed that graduate students are able to experience technological and creative interest.
SEMG signal compression based on two-dimensional techniques.
de Melo, Wheidima Carneiro; de Lima Filho, Eddie Batista; da Silva Júnior, Waldir Sabino
2016-04-18
Recently, two-dimensional techniques have been successfully employed for compressing surface electromyographic (SEMG) records as images, through the use of image and video encoders. Such schemes usually provide specific compressors, which are tuned for SEMG data, or employ preprocessing techniques, before the two-dimensional encoding procedure, in order to provide a suitable data organization, whose correlations can be better exploited by off-the-shelf encoders. Besides preprocessing input matrices, one may also depart from those approaches and employ an adaptive framework, which is able to directly tackle SEMG signals reassembled as images. This paper proposes a new two-dimensional approach for SEMG signal compression, which is based on a recurrent pattern matching algorithm called multidimensional multiscale parser (MMP). The mentioned encoder was modified, in order to efficiently work with SEMG signals and exploit their inherent redundancies. Moreover, a new preprocessing technique, named as segmentation by similarity (SbS), which has the potential to enhance the exploitation of intra- and intersegment correlations, is introduced, the percentage difference sorting (PDS) algorithm is employed, with different image compressors, and results with the high efficiency video coding (HEVC), H.264/AVC, and JPEG2000 encoders are presented. Experiments were carried out with real isometric and dynamic records, acquired in laboratory. Dynamic signals compressed with H.264/AVC and HEVC, when combined with preprocessing techniques, resulted in good percent root-mean-square difference [Formula: see text] compression factor figures, for low and high compression factors, respectively. Besides, regarding isometric signals, the modified two-dimensional MMP algorithm outperformed state-of-the-art schemes, for low compression factors, the combination between SbS and HEVC proved to be competitive, for high compression factors, and JPEG2000, combined with PDS, provided good performance
On the Use of Statistics in Design and the Implications for Deterministic Computer Experiments
NASA Technical Reports Server (NTRS)
Simpson, Timothy W.; Peplinski, Jesse; Koch, Patrick N.; Allen, Janet K.
1997-01-01
Perhaps the most prevalent use of statistics in engineering design is through Taguchi's parameter and robust design -- using orthogonal arrays to compute signal-to-noise ratios in a process of design improvement. In our view, however, there is an equally exciting use of statistics in design that could become just as prevalent: it is the concept of metamodeling whereby statistical models are built to approximate detailed computer analysis codes. Although computers continue to get faster, analysis codes always seem to keep pace so that their computational time remains non-trivial. Through metamodeling, approximations of these codes are built that are orders of magnitude cheaper to run. These metamodels can then be linked to optimization routines for fast analysis, or they can serve as a bridge for integrating analysis codes across different domains. In this paper we first review metamodeling techniques that encompass design of experiments, response surface methodology, Taguchi methods, neural networks, inductive learning, and kriging. We discuss their existing applications in engineering design and then address the dangers of applying traditional statistical techniques to approximate deterministic computer analysis codes. We conclude with recommendations for the appropriate use of metamodeling techniques in given situations and how common pitfalls can be avoided.
Optimization of a chemical identification algorithm
NASA Astrophysics Data System (ADS)
Chyba, Thomas H.; Fisk, Brian; Gunning, Christin; Farley, Kevin; Polizzi, Amber; Baughman, David; Simpson, Steven; Slamani, Mohamed-Adel; Almassy, Robert; Da Re, Ryan; Li, Eunice; MacDonald, Steve; Slamani, Ahmed; Mitchell, Scott A.; Pendell-Jones, Jay; Reed, Timothy L.; Emge, Darren
2010-04-01
A procedure to evaluate and optimize the performance of a chemical identification algorithm is presented. The Joint Contaminated Surface Detector (JCSD) employs Raman spectroscopy to detect and identify surface chemical contamination. JCSD measurements of chemical warfare agents, simulants, toxic industrial chemicals, interferents and bare surface backgrounds were made in the laboratory and under realistic field conditions. A test data suite, developed from these measurements, is used to benchmark algorithm performance throughout the improvement process. In any one measurement, one of many possible targets can be present along with interferents and surfaces. The detection results are expressed as a 2-category classification problem so that Receiver Operating Characteristic (ROC) techniques can be applied. The limitations of applying this framework to chemical detection problems are discussed along with means to mitigate them. Algorithmic performance is optimized globally using robust Design of Experiments and Taguchi techniques. These methods require figures of merit to trade off between false alarms and detection probability. Several figures of merit, including the Matthews Correlation Coefficient and the Taguchi Signal-to-Noise Ratio are compared. Following the optimization of global parameters which govern the algorithm behavior across all target chemicals, ROC techniques are employed to optimize chemical-specific parameters to further improve performance.
Analysis of sampling techniques for imbalanced data: An n = 648 ADNI study.
Dubey, Rashmi; Zhou, Jiayu; Wang, Yalin; Thompson, Paul M; Ye, Jieping
2014-02-15
Many neuroimaging applications deal with imbalanced imaging data. For example, in Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset, the mild cognitive impairment (MCI) cases eligible for the study are nearly two times the Alzheimer's disease (AD) patients for structural magnetic resonance imaging (MRI) modality and six times the control cases for proteomics modality. Constructing an accurate classifier from imbalanced data is a challenging task. Traditional classifiers that aim to maximize the overall prediction accuracy tend to classify all data into the majority class. In this paper, we study an ensemble system of feature selection and data sampling for the class imbalance problem. We systematically analyze various sampling techniques by examining the efficacy of different rates and types of undersampling, oversampling, and a combination of over and undersampling approaches. We thoroughly examine six widely used feature selection algorithms to identify significant biomarkers and thereby reduce the complexity of the data. The efficacy of the ensemble techniques is evaluated using two different classifiers including Random Forest and Support Vector Machines based on classification accuracy, area under the receiver operating characteristic curve (AUC), sensitivity, and specificity measures. Our extensive experimental results show that for various problem settings in ADNI, (1) a balanced training set obtained with K-Medoids technique based undersampling gives the best overall performance among different data sampling techniques and no sampling approach; and (2) sparse logistic regression with stability selection achieves competitive performance among various feature selection algorithms. Comprehensive experiments with various settings show that our proposed ensemble model of multiple undersampled datasets yields stable and promising results. © 2013 Elsevier Inc. All rights reserved.
The Geometric Organizer: A Study Technique.
ERIC Educational Resources Information Center
Derr, Alice M.; Peters, Chris L.
1986-01-01
The geometric organizer, a multisensory technique using visual mnemonic devices that key information to color-coded geometric shapes, can help learning disabled students read, organize, and study information in content subject textbooks. (CL)
Reliable quantification techniques for carbon nanotubes (CNTs) are limited. In this study, a new procedure was developed for quantifying multi-walled carbon nanotubes (MWNTs) in earthworms (Eisenia fetida) based on freeze drying and microwave-induced heating. Specifically, earthw...
Chen, Yu-Cheng; Tsai, Perng-Jy; Mou, Jin-Luh
2008-07-15
This study is the first one using the Taguchi experimental design to identify the optimal operating condition for reducing polychlorinated dibenzo-p-dioxins and dibenzofurans (PCDD/ Fs) formations during the iron ore sintering process. Four operating parameters, including the water content (Wc; range = 6.0-7.0 wt %), suction pressure (Ps; range = 1000-1400 mmH2O), bed height (Hb; range = 500-600 mm), and type of hearth layer (including sinter, hematite, and limonite), were selected for conducting experiments in a pilot scale sinter pot to simulate various sintering operating conditions of a real-scale sinter plant We found that the resultant optimal combination (Wc = 6.5 wt%, Hb = 500 mm, Ps = 1000 mmH2O, and hearth layer = hematite) could decrease the emission factor of total PCDD/Fs (total EF(PCDD/Fs)) up to 62.8% by reference to the current operating condition of the real-scale sinter plant (Wc = 6.5 wt %, Hb = 550 mm, Ps = 1200 mmH2O, and hearth layer = sinter). Through the ANOVA analysis, we found that Wc was the most significant parameter in determining total EF(PCDD/Fs (accounting for 74.7% of the total contribution of the four selected parameters). The resultant optimal combination could also enhance slightly in both sinter productivity and sinter strength (30.3 t/m2/day and 72.4%, respectively) by reference to those obtained from the reference operating condition (29.9 t/m (2)/day and 72.2%, respectively). The above results further ensure the applicability of the obtained optimal combination for the real-scale sinter production without interfering its sinter productivity and sinter strength.
EL-Moslamy, Shahira H.; Elkady, Marwa F.; Rezk, Ahmed H.; Abdel-Fattah, Yasser R.
2017-01-01
Development of reliable and low-cost requirement for large-scale eco-friendly biogenic synthesis of metallic nanoparticles is an important step for industrial applications of bionanotechnology. In the present study, the mycosynthesis of spherical nano-Ag (12.7 ± 0.8 nm) from extracellular filtrate of local endophytic T. harzianum SYA.F4 strain which have interested mixed bioactive metabolites (alkaloids, flavonoids, tannins, phenols, nitrate reductase (320 nmol/hr/ml), carbohydrate (25 μg/μl) and total protein concentration (2.5 g/l) was reported. Industrial mycosynthesis of nano-Ag can be induced with different characters depending on the fungal cultivation and physical conditions. Taguchi design was applied to improve the physicochemical conditions for nano-Ag production, and the optimum conditions which increased its mass weight 3 times larger than a basal condition were as follows: AgNO3 (0.01 M), diluted reductant (10 v/v, pH 5) and incubated at 30 °C, 200 rpm for 24 hr. Kinetic conversion rates in submerged batch cultivation in 7 L stirred tank bioreactor on using semi-defined cultivation medium was as follows: the maximum biomass production (Xmax) and maximum nano-Ag mass weight (Pmax) calculated (60.5 g/l and 78.4 g/l respectively). The best nano-Ag concentration that formed large inhibition zones was 100 μg/ml which showed against A.alternate (43 mm) followed by Helminthosporium sp. (35 mm), Botrytis sp. (32 mm) and P. arenaria (28 mm). PMID:28349997
Jabeen, Safia; Mehmood, Zahid; Mahmood, Toqeer; Saba, Tanzila; Rehman, Amjad; Mahmood, Muhammad Tariq
2018-01-01
For the last three decades, content-based image retrieval (CBIR) has been an active research area, representing a viable solution for retrieving similar images from an image repository. In this article, we propose a novel CBIR technique based on the visual words fusion of speeded-up robust features (SURF) and fast retina keypoint (FREAK) feature descriptors. SURF is a sparse descriptor whereas FREAK is a dense descriptor. Moreover, SURF is a scale and rotation-invariant descriptor that performs better in the case of repeatability, distinctiveness, and robustness. It is robust to noise, detection errors, geometric, and photometric deformations. It also performs better at low illumination within an image as compared to the FREAK descriptor. In contrast, FREAK is a retina-inspired speedy descriptor that performs better for classification-based problems as compared to the SURF descriptor. Experimental results show that the proposed technique based on the visual words fusion of SURF-FREAK descriptors combines the features of both descriptors and resolves the aforementioned issues. The qualitative and quantitative analysis performed on three image collections, namely Corel-1000, Corel-1500, and Caltech-256, shows that proposed technique based on visual words fusion significantly improved the performance of the CBIR as compared to the feature fusion of both descriptors and state-of-the-art image retrieval techniques. PMID:29694429
Weight optimization of an aerobrake structural concept for a lunar transfer vehicle
NASA Technical Reports Server (NTRS)
Bush, Lance B.; Unal, Resit; Rowell, Lawrence F.; Rehder, John J.
1992-01-01
An aerobrake structural concept for a lunar transfer vehicle was weight optimized through the use of the Taguchi design method, finite element analyses, and element sizing routines. Six design parameters were chosen to represent the aerobrake structural configuration. The design parameters included honeycomb core thickness, diameter-depth ratio, shape, material, number of concentric ring frames, and number of radial frames. Each parameter was assigned three levels. The aerobrake structural configuration with the minimum weight was 44 percent less than the average weight of all the remaining satisfactory experimental configurations. In addition, the results of this study have served to bolster the advocacy of the Taguchi method for aerospace vehicle design. Both reduced analysis time and an optimized design demonstrated the applicability of the Taguchi method to aerospace vehicle design.
Photogrammetric Techniques for Paleoanthropological Objects Preserving and Studying
NASA Astrophysics Data System (ADS)
Knyaz, V. A.; Leybova, N. A.; Galeev, R.; Novikov, M.; Gaboutchian, A. V.
2018-05-01
Paleo-anthropological research has its specificity closely related with studied objects. Their complicated shape arises from anatomical features of human skull and other skeletal bones. The degree of preservation is associated with the fragility of palaeo-anthropological material which usually has high historical and scientific value. The circumstances mentioned above enhance the relevance of photogrammetry implementation in anthropological studies. Thus, such combination of scientific methodologies with up-to-date technology creates a potential for improvement of various stages of palaeo-anthropological studies. This can be referred to accurate documenting of anthropological material and creation of databases accessible for wide range of users, predominantly research scientists and students; preservation of highly valuable samples and possibility of sharing information as 3D images or printed copies, improving co-operation of scientists world-wide; potential for replication of contact anthropometric studies on 3D images or printed copies providing for development of new biometric methods, and etc. This paper presents an approach based on photogrammetric techniques and non-contact measurements, providing technological and methodological development of paleo-anthropological studies, including data capturing, processing and representing.
Application of seismic-refraction techniques to hydrologic studies
Haeni, F.P.
1986-01-01
During the past 30 years, seismic-refraction methods have been used extensively in petroleum, mineral, and engineering investigations, and to some extent for hydrologic applications. Recent advances in equipment, sound sources, and computer interpretation techniques make seismic refraction a highly effective and economical means of obtaining subsurface data in hydrologic studies. Aquifers that can be defined by one or more high seismic-velocity surfaces, such as (1) alluvial or glacial deposits in consolidated rock valleys, (2) limestone or sandstone underlain by metamorphic or igneous rock, or (3) saturated unconsolidated deposits overlain by unsaturated unconsolidated deposits,are ideally suited for applying seismic-refraction methods. These methods allow the economical collection of subsurface data, provide the basis for more efficient collection of data by test drilling or aquifer tests, and result in improved hydrologic studies.This manual briefly reviews the basics of seismic-refraction theory and principles. It emphasizes the use of this technique in hydrologic investigations and describes the planning, equipment, field procedures, and intrepretation techniques needed for this type of study.Examples of the use of seismic-refraction techniques in a wide variety of hydrologic studies are presented.
Application of seismic-refraction techniques to hydrologic studies
Haeni, F.P.
1988-01-01
During the past 30 years, seismic-refraction methods have been used extensively in petroleum, mineral, and engineering investigations and to some extent for hydrologic applications. Recent advances in equipment, sound sources, and computer interpretation techniques make seismic refraction a highly effective and economical means of obtaining subsurface data in hydrologic studies. Aquifers that can be defined by one or more high-seismic-velocity surface, such as (1) alluvial or glacial deposits in consolidated rock valleys, (2) limestone or sandstone underlain by metamorphic or igneous rock, or (3) saturated unconsolidated deposits overlain by unsaturated unconsolidated deposits, are ideally suited for seismic-refraction methods. These methods allow economical collection of subsurface data, provide the basis for more efficient collection of data by test drilling or aquifer tests, and result in improved hydrologic studies. This manual briefly reviews the basics of seismic-refraction theory and principles. It emphasizes the use of these techniques in hydrologic investigations and describes the planning, equipment, field procedures, and interpretation techniques needed for this type of study. Further-more, examples of the use of seismic-refraction techniques in a wide variety of hydrologic studies are presented.
Gray, B.A.; Zori, Roberto T.; McGuire, P.M.; Bonde, R.K.
2002-01-01
Detailed chromosome studies were conducted for the Florida manatee (Trichechus manatus latirostris) utilizing primary chromosome banding techniques (G- and Q-banding). Digital microscopic imaging methods were employed and a standard G-banded karyotype was constructed for both sexes. Based on chromosome banding patterns and measurements obtained in these studies, a standard karyotype and ideogram are proposed. Characterization of additional cytogenetic features of this species by supplemental chromosome banding techniques, C-banding (constitutive heterochromatin), Ag-NOR staining (nucleolar organizer regions), and DA/DAPI staining, was also performed. These studies provide detailed cytogenetic data for T. manatus latirostris, which could enhance future genetic mapping projects and interspecific and intraspecific genomic comparisons by techniques such as zoo-FISH.
Feng, Jingwen; Feng, Tong; Yang, Chengwen; Wang, Wei; Sa, Yu; Feng, Yuanming
2018-06-01
This study was to explore the feasibility of prediction and classification of cells in different stages of apoptosis with a stain-free method based on diffraction images and supervised machine learning. Apoptosis was induced in human chronic myelogenous leukemia K562 cells by cis-platinum (DDP). A newly developed technique of polarization diffraction imaging flow cytometry (p-DIFC) was performed to acquire diffraction images of the cells in three different statuses (viable, early apoptotic and late apoptotic/necrotic) after cell separation through fluorescence activated cell sorting with Annexin V-PE and SYTOX® Green double staining. The texture features of the diffraction images were extracted with in-house software based on the Gray-level co-occurrence matrix algorithm to generate datasets for cell classification with supervised machine learning method. Therefore, this new method has been verified in hydrogen peroxide induced apoptosis model of HL-60. Results show that accuracy of higher than 90% was achieved respectively in independent test datasets from each cell type based on logistic regression with ridge estimators, which indicated that p-DIFC system has a great potential in predicting and classifying cells in different stages of apoptosis.
Nano Petri dishes: a new polystyrene platform for studying cell-nanoengineered surface interactions
NASA Astrophysics Data System (ADS)
Cha, Kyoung Je; Na, Moon-Hee; Kim, Hyung Woo; Kim, Dong Sung
2014-05-01
In this study, we fabricated and fully characterized a new type of polystyrene (PS) cell-culture platform containing nanoengineered surfaces (NES), referred to as a nano Petri dish, which can be used at the transition stage of basic cell-NES interaction studies for clinical applications. Nano-injection molding in this study was used for the mass production of the nano Petri dish having nanopore arrays. The effects of processing parameters of the injection molding on the replication quality of the nanopore arrays were quantitatively evaluated by means of design of experiments based on the Taguchi method. This allowed efficient and reliable cell culture studies by providing large numbers of the same dishes, in addition to removing the fixation step of the NES plates inside the cell-culture container. Physical, chemical and mechanical properties of the NES, as well as cell behavior including attachment and proliferation of human osteosarcoma MG-63 cells on the NES, were then characterized, with and without the oxygen plasma surface treatment.
Novel technique: a pupillometer-based objective chromatic perimetry
NASA Astrophysics Data System (ADS)
Rotenstreich, Ygal; Skaat, Alon; Sher, Ifat; Kolker, Andru; Rosenfeld, Elkana; Melamed, Shlomo; Belkin, Michael
2014-02-01
Evaluation of visual field (VF) is important for clinical diagnosis and patient monitoring. The current VF methods are subjective and require patient cooperation. Here we developed a novel objective perimetry technique based on the pupil response (PR) to multifocal chromatic stimuli in normal subjects and in patients with glaucoma and retinitis pigmentosa (RP). A computerized infrared video pupillometer was used to record PR to short- and long-wavelength stimuli (peak 485 nm and 620 nm, respectively) at light intensities of 15-100 cd-s/m2 at thirteen different points of the VF. The RP study included 30 eyes of 16 patients and 20 eyes of 12 healthy participants. The glaucoma study included 22 eyes of 11 patients and 38 eyes of 19 healthy participants. Significantly reduced PR was observed in RP patients in response to short-wavelength stimuli at 40 cd-s/m2 in nearly all perimetric locations (P <0.05). By contrast, RP patients demonstrated nearly normal PR to long-wavelength in majority of perimetric locations. The glaucoma group showed significantly reduced PR to long- and short-wavelength stimuli at high intensity in all perimetric locations (P <0.05). The PR of glaucoma patients was significantly lower than normal in response to short-wavelength stimuli at low intensity mostly in central and 20° locations (p<0.05). This study demonstrates the feasibility of using pupillometer-based chromatic perimetry for objectively assessing VF defects and retinal function and optic nerve damage in patients with retinal dystrophies and glaucoma. Furthermore, this method may be used to distinguish between the damaged cells underlying the VF defect.
NASA Astrophysics Data System (ADS)
Guidang, Excel Philip B.; Llanda, Christopher John R.; Palaoag, Thelma D.
2018-03-01
Face Detection Technique as a strategy in controlling a multimedia instructional material was implemented in this study. Specifically, it achieved the following objectives: 1) developed a face detection application that controls an embedded mother-tongue-based instructional material for face-recognition configuration using Python; 2) determined the perceptions of the students using the Mutt Susan’s student app review rubric. The study concludes that face detection technique is effective in controlling an electronic instructional material. It can be used to change the method of interaction of the student with an instructional material. 90% of the students perceived the application to be a great app and 10% rated the application to be good.
Optimization of Profile and Material of Abrasive Water Jet Nozzle
NASA Astrophysics Data System (ADS)
Anand Bala Selwin, K. P.; Ramachandran, S.
2017-05-01
The objective of this work is to study the behaviour of the abrasive water jet nozzle with different profiles and materials. Taguchi-Grey relational analysis optimization technique is used to optimize the value with different material and different profiles. Initially the 3D models of the nozzle are modelled with different profiles by changing the tapered inlet angle of the nozzle. The different profile models are analysed with different materials and the results are optimized. The optimized results would give the better result taking wear and machining behaviour of the nozzle.
Validation techniques of agent based modelling for geospatial simulations
NASA Astrophysics Data System (ADS)
Darvishi, M.; Ahmadi, G.
2014-10-01
One of the most interesting aspects of modelling and simulation study is to describe the real world phenomena that have specific properties; especially those that are in large scales and have dynamic and complex behaviours. Studying these phenomena in the laboratory is costly and in most cases it is impossible. Therefore, Miniaturization of world phenomena in the framework of a model in order to simulate the real phenomena is a reasonable and scientific approach to understand the world. Agent-based modelling and simulation (ABMS) is a new modelling method comprising of multiple interacting agent. They have been used in the different areas; for instance, geographic information system (GIS), biology, economics, social science and computer science. The emergence of ABM toolkits in GIS software libraries (e.g. ESRI's ArcGIS, OpenMap, GeoTools, etc) for geospatial modelling is an indication of the growing interest of users to use of special capabilities of ABMS. Since ABMS is inherently similar to human cognition, therefore it could be built easily and applicable to wide range applications than a traditional simulation. But a key challenge about ABMS is difficulty in their validation and verification. Because of frequent emergence patterns, strong dynamics in the system and the complex nature of ABMS, it is hard to validate and verify ABMS by conventional validation methods. Therefore, attempt to find appropriate validation techniques for ABM seems to be necessary. In this paper, after reviewing on Principles and Concepts of ABM for and its applications, the validation techniques and challenges of ABM validation are discussed.
The Extended-Image Tracking Technique Based on the Maximum Likelihood Estimation
NASA Technical Reports Server (NTRS)
Tsou, Haiping; Yan, Tsun-Yee
2000-01-01
This paper describes an extended-image tracking technique based on the maximum likelihood estimation. The target image is assume to have a known profile covering more than one element of a focal plane detector array. It is assumed that the relative position between the imager and the target is changing with time and the received target image has each of its pixels disturbed by an independent additive white Gaussian noise. When a rotation-invariant movement between imager and target is considered, the maximum likelihood based image tracking technique described in this paper is a closed-loop structure capable of providing iterative update of the movement estimate by calculating the loop feedback signals from a weighted correlation between the currently received target image and the previously estimated reference image in the transform domain. The movement estimate is then used to direct the imager to closely follow the moving target. This image tracking technique has many potential applications, including free-space optical communications and astronomy where accurate and stabilized optical pointing is essential.
van der Ploeg, Tjeerd; Austin, Peter C; Steyerberg, Ewout W
2014-12-22
Modern modelling techniques may potentially provide more accurate predictions of binary outcomes than classical techniques. We aimed to study the predictive performance of different modelling techniques in relation to the effective sample size ("data hungriness"). We performed simulation studies based on three clinical cohorts: 1282 patients with head and neck cancer (with 46.9% 5 year survival), 1731 patients with traumatic brain injury (22.3% 6 month mortality) and 3181 patients with minor head injury (7.6% with CT scan abnormalities). We compared three relatively modern modelling techniques: support vector machines (SVM), neural nets (NN), and random forests (RF) and two classical techniques: logistic regression (LR) and classification and regression trees (CART). We created three large artificial databases with 20 fold, 10 fold and 6 fold replication of subjects, where we generated dichotomous outcomes according to different underlying models. We applied each modelling technique to increasingly larger development parts (100 repetitions). The area under the ROC-curve (AUC) indicated the performance of each model in the development part and in an independent validation part. Data hungriness was defined by plateauing of AUC and small optimism (difference between the mean apparent AUC and the mean validated AUC <0.01). We found that a stable AUC was reached by LR at approximately 20 to 50 events per variable, followed by CART, SVM, NN and RF models. Optimism decreased with increasing sample sizes and the same ranking of techniques. The RF, SVM and NN models showed instability and a high optimism even with >200 events per variable. Modern modelling techniques such as SVM, NN and RF may need over 10 times as many events per variable to achieve a stable AUC and a small optimism than classical modelling techniques such as LR. This implies that such modern techniques should only be used in medical prediction problems if very large data sets are available.
Yang, Rongbing; Nam, Kihoon; Kim, Sung Wan; Turkson, James; Zou, Ye; Zuo, Yi Y; Haware, Rahul V; Chougule, Mahavir B
2017-01-03
Desired characteristics of nanocarriers are crucial to explore its therapeutic potential. This investigation aimed to develop tunable bioresponsive newly synthesized unique arginine grafted poly(cystaminebis(acrylamide)-diaminohexane) [ABP] polymeric matrix based nanocarriers by using L9 Taguchi factorial design, desirability function, and multivariate method. The selected formulation and process parameters were ABP concentration, acetone concentration, the volume ratio of acetone to ABP solution, and drug concentration. The measured nanocarrier characteristics were particle size, polydispersity index, zeta potential, and percentage drug loading. Experimental validation of nanocarrier characteristics computed from initially developed predictive model showed nonsignificant differences (p > 0.05). The multivariate modeling based optimized cationic nanocarrier formulation of <100 nm loaded with hydrophilic acetaminophen was readapted for a hydrophobic etoposide loading without significant changes (p > 0.05) except for improved loading percentage. This is the first study focusing on ABP polymeric matrix based nanocarrier development. Nanocarrier particle size was stable in PBS 7.4 for 48 h. The increase of zeta potential at lower pH 6.4, compared to the physiological pH, showed possible endosomal escape capability. The glutathione triggered release at the physiological conditions indicated the competence of cytosolic targeting delivery of the loaded drug from bioresponsive nanocarriers. In conclusion, this unique systematic approach provides rational evaluation and prediction of a tunable bioresponsive ABP based matrix nanocarrier, which was built on selected limited number of smart experimentation.
Nanomaterials-Based Optical Techniques for the Detection of Acetylcholinesterase and Pesticides
Xia, Ning; Wang, Qinglong; Liu, Lin
2015-01-01
The large amount of pesticide residues in the environment is a threat to global health by inhibition of acetylcholinesterase (AChE). Biosensors for inhibition of AChE have been thus developed for the detection of pesticides. In line with the rapid development of nanotechnology, nanomaterials have attracted great attention and have been intensively studied in biological analysis due to their unique chemical, physical and size properties. The aim of this review is to provide insight into nanomaterial-based optical techniques for the determination of AChE and pesticides, including colorimetric and fluorescent assays and surface plasmon resonance. PMID:25558991
Biometric image enhancement using decision rule based image fusion techniques
NASA Astrophysics Data System (ADS)
Sagayee, G. Mary Amirtha; Arumugam, S.
2010-02-01
Introducing biometrics into information systems may result in considerable benefits. Most of the researchers confirmed that the finger print is widely used than the iris or face and more over it is the primary choice for most privacy concerned applications. For finger prints applications, choosing proper sensor is at risk. The proposed work deals about, how the image quality can be improved by introducing image fusion technique at sensor levels. The results of the images after introducing the decision rule based image fusion technique are evaluated and analyzed with its entropy levels and root mean square error.
Pseudorandom Noise Code-Based Technique for Cloud and Aerosol Discrimination Applications
NASA Technical Reports Server (NTRS)
Campbell, Joel F.; Prasad, Narasimha S.; Flood, Michael A.; Harrison, Fenton Wallace
2011-01-01
NASA Langley Research Center is working on a continuous wave (CW) laser based remote sensing scheme for the detection of CO2 and O2 from space based platforms suitable for ACTIVE SENSING OF CO2 EMISSIONS OVER NIGHTS, DAYS, AND SEASONS (ASCENDS) mission. ASCENDS is a future space-based mission to determine the global distribution of sources and sinks of atmospheric carbon dioxide (CO2). A unique, multi-frequency, intensity modulated CW (IMCW) laser absorption spectrometer (LAS) operating at 1.57 micron for CO2 sensing has been developed. Effective aerosol and cloud discrimination techniques are being investigated in order to determine concentration values with accuracies less than 0.3%. In this paper, we discuss the demonstration of a PN code based technique for cloud and aerosol discrimination applications. The possibility of using maximum length (ML)-sequences for range and absorption measurements is investigated. A simple model for accomplishing this objective is formulated, Proof-of-concept experiments carried out using SONAR based LIDAR simulator that was built using simple audio hardware provided promising results for extension into optical wavelengths. Keywords: ASCENDS, CO2 sensing, O2 sensing, PN codes, CW lidar
Agent-based modeling: Methods and techniques for simulating human systems
Bonabeau, Eric
2002-01-01
Agent-based modeling is a powerful simulation modeling technique that has seen a number of applications in the last few years, including applications to real-world business problems. After the basic principles of agent-based simulation are briefly introduced, its four areas of application are discussed by using real-world applications: flow simulation, organizational simulation, market simulation, and diffusion simulation. For each category, one or several business applications are described and analyzed. PMID:12011407
In vivo studies of brain development by magnetic resonance techniques.
Inder, T E; Huppi, P S
2000-01-01
Understanding of the morphological development of the human brain has largely come from neuropathological studies obtained postmortem. Magnetic resonance (MR) techniques have recently allowed the provision of detailed structural, metabolic, and functional information in vivo on the human brain. These techniques have been utilized in studies from premature infants to adults and have provided invaluable data on the sequence of normal human brain development. This article will focus on MR techniques including conventional structural MR imaging techniques, quantitative morphometric MR techniques, diffusion weighted MR techniques, and MR spectroscopy. In order to understand the potential applications and limitations of MR techniques, relevant physical and biological principles for each of the MR techniques are first reviewed. This is followed by a review of the understanding of the sequence of normal brain development utilizing these techniques. MRDD Research Reviews 6:59-67, 2000. Copyright 2000 Wiley-Liss, Inc.
Dorsey, Shannon; Kerns, Suzanne E U; Lucid, Leah; Pullmann, Michael D; Harrison, Julie P; Berliner, Lucy; Thompson, Kelly; Deblinger, Esther
2018-01-24
Workplace-based clinical supervision as an implementation strategy to support evidence-based treatment (EBT) in public mental health has received limited research attention. A commonly provided infrastructure support, it may offer a relatively cost-neutral implementation strategy for organizations. However, research has not objectively examined workplace-based supervision of EBT and specifically how it might differ from EBT supervision provided in efficacy and effectiveness trials. Data come from a descriptive study of supervision in the context of a state-funded EBT implementation effort. Verbal interactions from audio recordings of 438 supervision sessions between 28 supervisors and 70 clinicians from 17 public mental health organizations (in 23 offices) were objectively coded for presence and intensity coverage of 29 supervision strategies (16 content and 13 technique items), duration, and temporal focus. Random effects mixed models estimated proportion of variance in content and techniques attributable to the supervisor and clinician levels. Interrater reliability among coders was excellent. EBT cases averaged 12.4 min of supervision per session. Intensity of coverage for EBT content varied, with some discussed frequently at medium or high intensity (exposure) and others infrequently discussed or discussed only at low intensity (behavior management; assigning/reviewing client homework). Other than fidelity assessment, supervision techniques common in treatment trials (e.g., reviewing actual practice, behavioral rehearsal) were used rarely or primarily at low intensity. In general, EBT content clustered more at the clinician level; different techniques clustered at either the clinician or supervisor level. Workplace-based clinical supervision may be a feasible implementation strategy for supporting EBT implementation, yet it differs from supervision in treatment trials. Time allotted per case is limited, compressing time for EBT coverage. Techniques that
Objective Assessment of Patient Inhaler User Technique Using an Audio-Based Classification Approach.
Taylor, Terence E; Zigel, Yaniv; Egan, Clarice; Hughes, Fintan; Costello, Richard W; Reilly, Richard B
2018-02-01
Many patients make critical user technique errors when using pressurised metered dose inhalers (pMDIs) which reduce the clinical efficacy of respiratory medication. Such critical errors include poor actuation coordination (poor timing of medication release during inhalation) and inhaling too fast (peak inspiratory flow rate over 90 L/min). Here, we present a novel audio-based method that objectively assesses patient pMDI user technique. The Inhaler Compliance Assessment device was employed to record inhaler audio signals from 62 respiratory patients as they used a pMDI with an In-Check Flo-Tone device attached to the inhaler mouthpiece. Using a quadratic discriminant analysis approach, the audio-based method generated a total frame-by-frame accuracy of 88.2% in classifying sound events (actuation, inhalation and exhalation). The audio-based method estimated the peak inspiratory flow rate and volume of inhalations with an accuracy of 88.2% and 83.94% respectively. It was detected that 89% of patients made at least one critical user technique error even after tuition from an expert clinical reviewer. This method provides a more clinically accurate assessment of patient inhaler user technique than standard checklist methods.
NASA Astrophysics Data System (ADS)
Colon-Pagan, Ian; Kuo, Ying-Hwa
2008-10-01
In this study, we compare precipitable water vapor (PWV) values from ground-based GPS water vapor sensing and COSMIC radio occultation (RO) measurements over the Caribbean Sea, Gulf of Mexico, and United States regions as well as global analyses from NCEP and ECMWF models. The results show good overall agreement; however, the PWV values estimated by ground-based GPS receivers tend to have a slight dry bias for low PWV values and a slight wet bias for higher PWV values, when compared with GPS RO measurements and global analyses. An application of a student T-test indicates that there is a significant difference between both ground- and space-based GPS measured datasets. The dry bias associated with space-based GPS is attributed to the missing low altitude data, where the concentration of water vapor is large. The close agreements between space-based and global analyses are due to the fact that these global analyses assimilate space-based GPS RO data from COSMIC, and the retrieval of water vapor profiles from space-based technique requires the use of global analyses as the first guess. This work is supported by UCAR SOARS and a grant from the National Oceanic and Atmospheric Administration, Educational Partnership Program under the cooperative agreement NA06OAR4810187.
Narrative Form and Yam Lau's "Room": The Encounter in Arts Based Research
ERIC Educational Resources Information Center
Boulton-Funke, Adrienne
2014-01-01
In this paper, I consider "the encounter" (O'Sullivan, 2006) and conceptualizations of subjectivity and identity proposed by post qualitative scholars (Jackson & Mazzei, 2012; Lather & St. Pierre, 2013; Lenz Taguchi, 2012; MacLure, 2013; St. Pierre, 2010) and contemporary art theory (O'Sullivan, 2006; 2012) to attend to…
Ionospheric Plasma Drift Analysis Technique Based On Ray Tracing
NASA Astrophysics Data System (ADS)
Ari, Gizem; Toker, Cenk
2016-07-01
Ionospheric drift measurements provide important information about the variability in the ionosphere, which can be used to quantify ionospheric disturbances caused by natural phenomena such as solar, geomagnetic, gravitational and seismic activities. One of the prominent ways for drift measurement depends on instrumentation based measurements, e.g. using an ionosonde. The drift estimation of an ionosonde depends on measuring the Doppler shift on the received signal, where the main cause of Doppler shift is the change in the length of the propagation path of the signal between the transmitter and the receiver. Unfortunately, ionosondes are expensive devices and their installation and maintenance require special care. Furthermore, the ionosonde network over the world or even Europe is not dense enough to obtain a global or continental drift map. In order to overcome the difficulties related to an ionosonde, we propose a technique to perform ionospheric drift estimation based on ray tracing. First, a two dimensional TEC map is constructed by using the IONOLAB-MAP tool which spatially interpolates the VTEC estimates obtained from the EUREF CORS network. Next, a three dimensional electron density profile is generated by inputting the TEC estimates to the IRI-2015 model. Eventually, a close-to-real situation electron density profile is obtained in which ray tracing can be performed. These profiles can be constructed periodically with a period of as low as 30 seconds. By processing two consequent snapshots together and calculating the propagation paths, we estimate the drift measurements over any coordinate of concern. We test our technique by comparing the results to the drift measurements taken at the DPS ionosonde at Pruhonice, Czech Republic. This study is supported by TUBITAK 115E915 and Joint TUBITAK 114E092 and AS CR14/001 projects.
Izadifar, Zahra; Belev, George; Izadifar, Mohammad; Izadifar, Zohreh; Chapman, Dean
2014-12-07
Observing cavitation bubbles deep within tissue is very difficult. The development of a method for probing cavitation, irrespective of its location in tissues, would improve the efficiency and application of ultrasound in the clinic. A synchrotron x-ray imaging technique, which is capable of detecting cavitation bubbles induced in water by a sonochemistry system, is reported here; this could possibly be extended to the study of therapeutic ultrasound in tissues. The two different x-ray imaging techniques of Analyzer Based Imaging (ABI) and phase contrast imaging (PCI) were examined in order to detect ultrasound induced cavitation bubbles. Cavitation was not observed by PCI, however it was detectable with ABI. Acoustic cavitation was imaged at six different acoustic power levels and six different locations through the acoustic beam in water at a fixed power level. The results indicate the potential utility of this technique for cavitation studies in tissues, but it is time consuming. This may be improved by optimizing the imaging method.
Acoustic Biometric System Based on Preprocessing Techniques and Linear Support Vector Machines
del Val, Lara; Izquierdo-Fuente, Alberto; Villacorta, Juan J.; Raboso, Mariano
2015-01-01
Drawing on the results of an acoustic biometric system based on a MSE classifier, a new biometric system has been implemented. This new system preprocesses acoustic images, extracts several parameters and finally classifies them, based on Support Vector Machine (SVM). The preprocessing techniques used are spatial filtering, segmentation—based on a Gaussian Mixture Model (GMM) to separate the person from the background, masking—to reduce the dimensions of images—and binarization—to reduce the size of each image. An analysis of classification error and a study of the sensitivity of the error versus the computational burden of each implemented algorithm are presented. This allows the selection of the most relevant algorithms, according to the benefits required by the system. A significant improvement of the biometric system has been achieved by reducing the classification error, the computational burden and the storage requirements. PMID:26091392
Acoustic Biometric System Based on Preprocessing Techniques and Linear Support Vector Machines.
del Val, Lara; Izquierdo-Fuente, Alberto; Villacorta, Juan J; Raboso, Mariano
2015-06-17
Drawing on the results of an acoustic biometric system based on a MSE classifier, a new biometric system has been implemented. This new system preprocesses acoustic images, extracts several parameters and finally classifies them, based on Support Vector Machine (SVM). The preprocessing techniques used are spatial filtering, segmentation-based on a Gaussian Mixture Model (GMM) to separate the person from the background, masking-to reduce the dimensions of images-and binarization-to reduce the size of each image. An analysis of classification error and a study of the sensitivity of the error versus the computational burden of each implemented algorithm are presented. This allows the selection of the most relevant algorithms, according to the benefits required by the system. A significant improvement of the biometric system has been achieved by reducing the classification error, the computational burden and the storage requirements.
Evaluation of a rule-based compositing technique for Landsat-5 TM and Landsat-7 ETM+ images
NASA Astrophysics Data System (ADS)
Lück, W.; van Niekerk, A.
2016-05-01
Image compositing is a multi-objective optimization process. Its goal is to produce a seamless cloud and artefact-free artificial image. This is achieved by aggregating image observations and by replacing poor and cloudy data with good observations from imagery acquired within the timeframe of interest. This compositing process aims to minimise the visual artefacts which could result from different radiometric properties, caused by atmospheric conditions, phenologic patterns and land cover changes. It has the following requirements: (1) image compositing must be cloud free, which requires the detection of clouds and shadows, and (2) the image composite must be seamless, minimizing artefacts and visible across inter image seams. This study proposes a new rule-based compositing technique (RBC) that combines the strengths of several existing methods. A quantitative and qualitative evaluation is made of the RBC technique by comparing it to the maximum NDVI (MaxNDVI), minimum red (MinRed) and maximum ratio (MaxRatio) compositing techniques. A total of 174 Landsat TM and ETM+ images, covering three study sites and three different timeframes for each site, are used in the evaluation. A new set of quantitative/qualitative evaluation techniques for compositing quality measurement was developed and showed that the RBC technique outperformed all other techniques, with MaxRatio, MaxNDVI, and MinRed techniques in order of performance from best to worst.
NASA Technical Reports Server (NTRS)
1976-01-01
Data base management techniques and applicable equipment are described. Recommendations which will assist potential NASA data users in selecting and using appropriate data base management tools and techniques are presented. Classes of currently available data processing equipment ranging from basic terminals to large minicomputer systems were surveyed as they apply to the needs of potential SEASAT data users. Cost and capabilities projections for this equipment through 1985 were presented. A test of a typical data base management system was described, as well as the results of this test and recommendations to assist potential users in determining when such a system is appropriate for their needs. The representative system tested was UNIVAC's DMS 1100.
Artificial Intelligence based technique for BTS placement
NASA Astrophysics Data System (ADS)
Alenoghena, C. O.; Emagbetere, J. O.; Aibinu, A. M.
2013-12-01
The increase of the base transceiver station (BTS) in most urban areas can be traced to the drive by network providers to meet demand for coverage and capacity. In traditional network planning, the final decision of BTS placement is taken by a team of radio planners, this decision is not fool proof against regulatory requirements. In this paper, an intelligent based algorithm for optimal BTS site placement has been proposed. The proposed technique takes into consideration neighbour and regulation considerations objectively while determining cell site. The application will lead to a quantitatively unbiased evaluated decision making process in BTS placement. An experimental data of a 2km by 3km territory was simulated for testing the new algorithm, results obtained show a 100% performance of the neighbour constrained algorithm in BTS placement optimization. Results on the application of GA with neighbourhood constraint indicate that the choices of location can be unbiased and optimization of facility placement for network design can be carried out.
FIELD STUDIES OF GEOMEMBRANE INSTALLATION TECHNIQUES
Fourteen construction sites where geomembranes were being installed were visited to observe subgrade preparation and liner installation techniques. These sites were visited during a study conducted for the U.S. EPA, Solid and Hazardous Waste Research Division. The sites included ...
Slowness based CCP stacking technique in suppressing crustal multiples
NASA Astrophysics Data System (ADS)
Guan, Z.; Niu, F.
2016-12-01
Common-conversion-point (CCP) stacking of receiver function is a widely used technique to image velocity discontinuities in the mantle, such as the lithosphere-asthenosphere boundary (LAB) in the upper mantle, the 410-km and the 660-km discontinuities in the mantle transition zone. In a layered medium, a teleseismic record can be considered as the summation of the direct arrival and a series of conversions and reflections at boundaries below the station. Receiver functions are an attempt to approximate a Green's function associated with structure beneath the receiver by deconvolving one component of a teleseismic signal from another to remove source signals from seismograms. The CCP technique assumes that receiver functions composed solely of P to S conversions at velocity boundaries, whose depths can be mapped out through their arrival times. The multiple reflections at shallow boundaries with large velocity contrasts, such as the base of unconsolidated sediments and the Moho, can pose significant challenges to the accuracy of CCP imaging. In principle, the P to S conversions and multiples originated from deep and shallow boundaries arrive at a seismic station with incident angles that are, respectively, smaller and larger than that of the direct P wave. Therefore the corresponding slowness can be used to isolate the conversions from multiples, allowing for minimizing multiple-induced artifacts. We developed a refined CCP stacking method that uses relative slowness as a weighting factor to suppress the multiples. We performed extensive numerical tests with synthetic data to seek the best weighting scheme and to verify the robustness of the images. We applied the refined technique to the NECESSArray data, and found that the complicated low velocity structures in the depth range of 200-400 km shown in the CCP images of previous studies are mostly artifacts resulted from crustal multiples.
A novel jet-based nano-hydroxyapatite patterning technique for osteoblast guidance
Li, Xiang; Koller, Garrit; Huang, Jie; Di Silvio, Lucy; Renton, Tara; Esat, Minoo; Bonfield, William; Edirisinghe, Mohan
2010-01-01
Surface topography is well known to play a crucial role in influencing cellular responses to an implant material and is therefore important in bone tissue regeneration. A novel jet-based patterning technique, template-assisted electrohydrodynamic atomization spraying, was recently devised to control precisely the surface structure as well as its dimensions. In the present study, a detailed investigation of this patterning process was carried out. A range of nano-hydroxyapatite (nHA) line-shaped patterns <20 µm in width were successfully deposited on a commercially pure Ti surface by controlling the flow of an nHA suspension in an electric field. In vitro studies showed that the nHA patterns generated are capable of regulating the human osteoblast cell attachment and orientation. PMID:19493897
Optimizing inhalation technique using web-based videos in obstructive lung diseases.
Müller, Tobias; Müller, Annegret; Hübel, Christian; Knipel, Verena; Windisch, Wolfram; Cornelissen, Christian Gabriel; Dreher, Michael
2017-08-01
Inhaled agents are widely used in the treatment of chronic airway diseases. Correct technique is required to ensure appropriate drug deposition, but poor technique is common. This study investigated whether inhalation technique could be improved by patient training using short videos from the German Airway League. Outpatients from a university hospital respiratory clinic who had incorrect inhalation technique were asked to demonstrate this again immediately after viewing the training videos, and after 4-8 weeks' follow-up. Inhalation technique was rated by a study nurse using specific checklists. One hundred and twelve patients with obstructive lung disease treated with inhaled bronchodilators or corticosteroids were included. More than half (51.8%) had at least one mistake in inhalation technique at baseline. Of these, most (88%) understood the training videos, 76% demonstrated correct device use immediately after training, and 72% were still able to demonstrate correct inhalation technique at follow-up (p = 0.0008 for trend). In addition, the number of mistakes decreased significantly after video training (by 1.82 [95% confidence interval 1.39-2.25]; p < 0.0001 vs. baseline). German Airway League inhalation technique training videos were easy to understand and effectively improved inhalation technique in patients with airway diseases. Copyright © 2017 Elsevier Ltd. All rights reserved.
Evaluation of Clipping Based Iterative PAPR Reduction Techniques for FBMC Systems
Kollár, Zsolt
2014-01-01
This paper investigates filter bankmulticarrier (FBMC), a multicarrier modulation technique exhibiting an extremely low adjacent channel leakage ratio (ACLR) compared to conventional orthogonal frequency division multiplexing (OFDM) technique. The low ACLR of the transmitted FBMC signal makes it especially favorable in cognitive radio applications, where strict requirements are posed on out-of-band radiation. Large dynamic range resulting in high peak-to-average power ratio (PAPR) is characteristic of all sorts of multicarrier signals. The advantageous spectral properties of the high-PAPR FBMC signal are significantly degraded if nonlinearities are present in the transceiver chain. Spectral regrowth may appear, causing harmful interference in the neighboring frequency bands. This paper presents novel clipping based PAPR reduction techniques, evaluated and compared by simulations and measurements, with an emphasis on spectral aspects. The paper gives an overall comparison of PAPR reduction techniques, focusing on the reduction of the dynamic range of FBMC signals without increasing out-of-band radiation. An overview is presented on transmitter oriented techniques employing baseband clipping, which can maintain the system performance with a desired bit error rate (BER). PMID:24558338
Overdenture retained by teeth using a definitive denture base technique: a case report.
Nascimento, D F F; dos Santos, J F F; Marchini, L
2010-09-01
This paper presents a technique involving the use of a definitive denture base to make overdentures. Cores with ball attachments were cemented over remaining lower teeth. Impressions of the edentulous maxilla and mandible were taken to obtain a definitive acrylic resin base. The definitive base of the mandible was perforated at the location of ball attachments and its female components were fixed to the base using acrylic resin directly in the patient's mouth. Wax rims were then made, jaw relationships recorded, teeth mounted and tried in, and the dentures were cured. This technique allowed for easy fixing of female components and better retention during the recording of jaw relationships, and can also be used in the construction of implant retained dentures.
Yatsushiro, Satoshi; Sunohara, Saeko; Hayashi, Naokazu; Hirayama, Akihiro; Matsumae, Mitsunori; Atsumi, Hideki; Kuroda, Kagayaki
2018-04-10
A correlation mapping technique delineating delay time and maximum correlation for characterizing pulsatile cerebrospinal fluid (CSF) propagation was proposed. After proofing its technical concept, this technique was applied to healthy volunteers and idiopathic normal pressure hydrocephalus (iNPH) patients. A time-resolved three dimensional-phase contrast (3D-PC) sampled the cardiac-driven CSF velocity at 32 temporal points per cardiac period at each spatial location using retrospective cardiac gating. The proposed technique visualized distributions of propagation delay and correlation coefficient of the PC-based CSF velocity waveform with reference to a waveform at a particular point in the CSF space. The delay time was obtained as the amount of time-shift, giving the maximum correlation for the velocity waveform at an arbitrary location with that at the reference location. The validity and accuracy of the technique were confirmed in a flow phantom equipped with a cardiovascular pump. The technique was then applied to evaluate the intracranial CSF motions in young, healthy (N = 13), and elderly, healthy (N = 13) volunteers and iNPH patients (N = 13). The phantom study demonstrated that root mean square error of the delay time was 2.27%, which was less than the temporal resolution of PC measurement used in this study (3.13% of a cardiac cycle). The human studies showed a significant difference (P < 0.01) in the mean correlation coefficient between the young, healthy group and the other two groups. A significant difference (P < 0.05) was also recognized in standard deviation of the correlation coefficients in intracranial CSF space among all groups. The result suggests that the CSF space compliance of iNPH patients was lower than that of healthy volunteers. The correlation mapping technique allowed us to visualize pulsatile CSF velocity wave propagations as still images. The technique may help to classify diseases related to CSF dynamics, such as iNPH.
de Oliveira, Paula Duarte; Menezes, Ana Maria Baptista; Bertoldi, Andréa Dâmaso; Wehrmeister, Fernando César; Macedo, Silvia Elaine Cardozo
2014-01-01
OBJECTIVE: To identify incorrect inhaler techniques employed by patients with respiratory diseases in southern Brazil and to profile the individuals who make such errors. METHODS: This was a population-based, cross-sectional study involving subjects ≥ 10 years of age using metered dose inhalers (MDIs) or dry powder inhalers (DPIs) in 1,722 households in the city of Pelotas, Brazil. RESULTS: We included 110 subjects, who collectively used 94 MDIs and 49 DPIs. The most common errors in the use of MDIs and DPIs were not exhaling prior to inhalation (66% and 47%, respectively), not performing a breath-hold after inhalation (29% and 25%), and not shaking the MDI prior to use (21%). Individuals ≥ 60 years of age more often made such errors. Among the demonstrations of the use of MDIs and DPIs, at least one error was made in 72% and 51%, respectively. Overall, there were errors made in all steps in 11% of the demonstrations, whereas there were no errors made in 13%.Among the individuals who made at least one error, the proportion of those with a low level of education was significantly greater than was that of those with a higher level of education, for MDIs (85% vs. 60%; p = 0.018) and for DPIs (81% vs. 35%; p = 0.010). CONCLUSIONS: In this sample, the most common errors in the use of inhalers were not exhaling prior to inhalation, not performing a breath-hold after inhalation, and not shaking the MDI prior to use. Special attention should be given to education regarding inhaler techniques for patients of lower socioeconomic status and with less formal education, as well as for those of advanced age, because those populations are at a greater risk of committing errors in their use of inhalers. PMID:25410839
NASA Astrophysics Data System (ADS)
Dostal, P.; Krasula, L.; Klima, M.
2012-06-01
Various image processing techniques in multimedia technology are optimized using visual attention feature of the human visual system. Spatial non-uniformity causes that different locations in an image are of different importance in terms of perception of the image. In other words, the perceived image quality depends mainly on the quality of important locations known as regions of interest. The performance of such techniques is measured by subjective evaluation or objective image quality criteria. Many state-of-the-art objective metrics are based on HVS properties; SSIM, MS-SSIM based on image structural information, VIF based on the information that human brain can ideally gain from the reference image or FSIM utilizing the low-level features to assign the different importance to each location in the image. But still none of these objective metrics utilize the analysis of regions of interest. We solve the question if these objective metrics can be used for effective evaluation of images reconstructed by processing techniques based on ROI analysis utilizing high-level features. In this paper authors show that the state-of-the-art objective metrics do not correlate well with subjective evaluation while the demosaicing based on ROI analysis is used for reconstruction. The ROI were computed from "ground truth" visual attention data. The algorithm combining two known demosaicing techniques on the basis of ROI location is proposed to reconstruct the ROI in fine quality while the rest of image is reconstructed with low quality. The color image reconstructed by this ROI approach was compared with selected demosaicing techniques by objective criteria and subjective testing. The qualitative comparison of the objective and subjective results indicates that the state-of-the-art objective metrics are still not suitable for evaluation image processing techniques based on ROI analysis and new criteria is demanded.
Development of a Blood-Based Biomarker Panel for Indeterminate Lung Nodules
2016-09-01
Taguchi CONTRACTING ORGANIZATION: University of Texas MD Anderson Cancer Center Houston, TX 77030 REPORT DATE: September 2016 TYPE OF REPORT...PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) 8. PERFORMING ORGANIZATION REPORT NUMBER University of Texas MD Anderson Cancer Center Rene...protein, autoantibody and microRNA) to distinguish malignant lung nodules from benign lung nodules. We have so far assayed 20 protein markers, 122
Generalized image contrast enhancement technique based on Heinemann contrast discrimination model
NASA Astrophysics Data System (ADS)
Liu, Hong; Nodine, Calvin F.
1994-03-01
This paper presents a generalized image contrast enhancement technique which equalizes perceived brightness based on the Heinemann contrast discrimination model. This is a modified algorithm which presents an improvement over the previous study by Mokrane in its mathematically proven existence of a unique solution and in its easily tunable parameterization. The model uses a log-log representation of contrast luminosity between targets and the surround in a fixed luminosity background setting. The algorithm consists of two nonlinear gray-scale mapping functions which have seven parameters, two of which are adjustable Heinemann constants. Another parameter is the background gray level. The remaining four parameters are nonlinear functions of gray scale distribution of the image, and can be uniquely determined once the previous three are given. Tests have been carried out to examine the effectiveness of the algorithm for increasing the overall contrast of images. It can be demonstrated that the generalized algorithm provides better contrast enhancement than histogram equalization. In fact, the histogram equalization technique is a special case of the proposed mapping.
Development of evaluation technique of GMAW welding quality based on statistical analysis
NASA Astrophysics Data System (ADS)
Feng, Shengqiang; Terasaki, Hidenri; Komizo, Yuichi; Hu, Shengsun; Chen, Donggao; Ma, Zhihua
2014-11-01
Nondestructive techniques for appraising gas metal arc welding(GMAW) faults plays a very important role in on-line quality controllability and prediction of the GMAW process. On-line welding quality controllability and prediction have several disadvantages such as high cost, low efficiency, complication and greatly being affected by the environment. An enhanced, efficient evaluation technique for evaluating welding faults based on Mahalanobis distance(MD) and normal distribution is presented. In addition, a new piece of equipment, designated the weld quality tester(WQT), is developed based on the proposed evaluation technique. MD is superior to other multidimensional distances such as Euclidean distance because the covariance matrix used for calculating MD takes into account correlations in the data and scaling. The values of MD obtained from welding current and arc voltage are assumed to follow a normal distribution. The normal distribution has two parameters: the mean µ and standard deviation σ of the data. In the proposed evaluation technique used by the WQT, values of MD located in the range from zero to µ+3 σ are regarded as "good". Two experiments which involve changing the flow of shielding gas and smearing paint on the surface of the substrate are conducted in order to verify the sensitivity of the proposed evaluation technique and the feasibility of using WQT. The experimental results demonstrate the usefulness of the WQT for evaluating welding quality. The proposed technique can be applied to implement the on-line welding quality controllability and prediction, which is of great importance to design some novel equipment for weld quality detection.
An additional study and implementation of tone calibrated technique of modulation
NASA Technical Reports Server (NTRS)
Rafferty, W.; Bechtel, L. K.; Lay, N. E.
1985-01-01
The Tone Calibrated Technique (TCT) was shown to be theoretically free from an error floor, and is only limited, in practice, by implementation constraints. The concept of the TCT transmission scheme along with a baseband implementation of a suitable demodulator is introduced. Two techniques for the generation of the TCT signal are considered: a Manchester source encoding scheme (MTCT) and a subcarrier based technique (STCT). The results are summarized for the TCT link computer simulation. The hardware implementation of the MTCT system is addressed and the digital signal processing design considerations involved in satisfying the modulator/demodulator requirements are outlined. The program findings are discussed and future direction are suggested based on conclusions made regarding the suitability of the TCT system for the transmission channel presently under consideration.
Cox, Martine Elizabeth; Small, Hannah Julie; Boyes, Allison W; O'Brien, Lorna; Rose, Shiho Karina; Baker, Amanda L; Henskens, Frans A; Kirkwood, Hannah Naomi; Roach, Della M
2017-01-01
, six publications met the eligibility criteria and were included in a narrative synthesis. All six studies addressed the effect of smileys or emoticons on participant responses, message interpretation, or social presence of the writer. None of these studies specifically targeted chronic conditions. It was found that emoticons were more effective in influencing the emotional impact of a message than no cue and that smileys and emoticons were able to convey a limited amount of emotion. No studies addressed other techniques for conveying emotion in written communication. No studies addressed the effects of any techniques on the nature of the interaction (eg, therapeutic alliance), patient perceptions of the interaction (eg, participant satisfaction), or psychological outcomes (depression, anxiety, or distress). Conclusions There is a need for greater empirical attention to the effects of the various proposed techniques for conveying emotion in Web-based typed communications to inform health service providers regarding best-practice communication skills in this setting. PMID:29066426
Multi-technique characterisation of commercial alizarin-based lakes
NASA Astrophysics Data System (ADS)
Pronti, Lucilla; Mazzitelli, Jean-Baptiste; Bracciale, Maria Paola; Massini Rosati, Lorenzo; Vieillescazes, Cathy; Santarelli, Maria Laura; Felici, Anna Candida
2018-07-01
The characterization of ancient and modern alizarin-based lakes is a largely studied topic in the literature. Analytical data on contemporary alizarin-based lakes, however, are still poor, though of primary importance, since these lakes might be indeed present in contemporary and fake paintings as well as in retouchings. In this work we systematically investigate the chemical composition and the optical features of fifteen alizarin-based lakes, by a multi-analytical technique approach combining spectroscopic methods (i.e. Energy Dispersive X-ray Fluorescence Spectroscopy, EDXRF; Attenuated Total Reflectance Fourier-Transform Infrared Spectroscopy, ATR-FTIR; X-ray Powder Diffraction, XRD; UV induced fluorescence and reflectance spectroscopies) and chromatography (i.e. High-performance Liquid Chromatography coupled with a Photodiode Array Detector, HPLC-PDA). Most of the samples contain typical compounds from the natural roots of madder, as occurring in ancient and modern lakes, but in two samples (23600-Kremer-Pigmente and alizarin crimson-Zecchi) any anthraquinonic structures were identified, thus leading to hypothesize the presence of synthetic dyes. The detection of lucidin primeveroside and ruberythrique acid in some lakes suggest the use of Rubia tinctorum. One sample (23610-Kremer-Pigmente) presents alizarin as the sole compound, thereby revealing to be a synthetic dye. Moreover, gibbsite, alunite and kaolinite were found to be used as substrates and/or mordants. Visible absorption spectra of the anthraquinonic lakes show two main absorption bands at about 494-511 nm and 537-564 nm, along with a shoulder at about 473-479 nm in presence of high amounts of purpurin. Finally, from the results obtained by UV induced fluorescence spectroscopy it is possible to figure out that, although it is commonly assumed that the madder lake presents an orange-pink fluorescence, the inorganic compounds, added to the recipe, could induce a quenching phenomenon or an inhibition
Monsoon Forecasting based on Imbalanced Classification Techniques
NASA Astrophysics Data System (ADS)
Ribera, Pedro; Troncoso, Alicia; Asencio-Cortes, Gualberto; Vega, Inmaculada; Gallego, David
2017-04-01
Monsoonal systems are quasiperiodic processes of the climatic system that control seasonal precipitation over different regions of the world. The Western North Pacific Summer Monsoon (WNPSM) is one of those monsoons and it is known to have a great impact both over the global climate and over the total precipitation of very densely populated areas. The interannual variability of the WNPSM along the last 50-60 years has been related to different climatic indices such as El Niño, El Niño Modoki, the Indian Ocean Dipole or the Pacific Decadal Oscillation. Recently, a new and longer series characterizing the monthly evolution of the WNPSM, the WNP Directional Index (WNPDI), has been developed, extending its previous length from about 50 years to more than 100 years (1900-2007). Imbalanced classification techniques have been applied to the WNPDI in order to check the capability of traditional climate indices to capture and forecast the evolution of the WNPSM. The problem of forecasting has been transformed into a binary classification problem, in which the positive class represents the occurrence of an extreme monsoon event. Given that the number of extreme monsoons is much lower than the number of non-extreme monsoons, the resultant classification problem is highly imbalanced. The complete dataset is composed of 1296 instances, where only 71 (5.47%) samples correspond to extreme monsoons. Twenty predictor variables based on the cited climatic indices have been proposed, and namely, models based on trees, black box models such as neural networks, support vector machines and nearest neighbors, and finally ensemble-based techniques as random forests have been used in order to forecast the occurrence of extreme monsoons. It can be concluded that the methodology proposed here reports promising results according to the quality parameters evaluated and predicts extreme monsoons for a temporal horizon of a month with a high accuracy. From a climatological point of view
New calibration technique for KCD-based megavoltage imaging
NASA Astrophysics Data System (ADS)
Samant, Sanjiv S.; Zheng, Wei; DiBianca, Frank A.; Zeman, Herbert D.; Laughter, Joseph S.
1999-05-01
In megavoltage imaging, current commercial electronic portal imaging devices (EPIDs), despite having the advantage of immediate digital imaging over film, suffer from poor image contrast and spatial resolution. The feasibility of using a kinestatic charge detector (KCD) as an EPID to provide superior image contrast and spatial resolution for portal imaging has already been demonstrated in a previous paper. The KCD system had the additional advantage of requiring an extremely low dose per acquired image, allowing for superior imaging to be reconstructed form a single linac pulse per image pixel. The KCD based images utilized a dose of two orders of magnitude less that for EPIDs and film. Compared with the current commercial EPIDs and film, the prototype KCD system exhibited promising image qualities, despite being handicapped by the use of a relatively simple image calibration technique, and the performance limits of medical linacs on the maximum linac pulse frequency and energy flux per pulse delivered. This image calibration technique fixed relative image pixel values based on a linear interpolation of extrema provided by an air-water calibration, and accounted only for channel-to-channel variations. The counterpart of this for area detectors is the standard flat fielding method. A comprehensive calibration protocol has been developed. The new technique additionally corrects for geometric distortions due to variations in the scan velocity, and timing artifacts caused by mis-synchronization between the linear accelerator and the data acquisition system (DAS). The role of variations in energy flux (2 - 3%) on imaging is demonstrated to be not significant for the images considered. The methodology is presented, and the results are discussed for simulated images. It also allows for significant improvements in the signal-to- noise ratio (SNR) by increasing the dose using multiple images without having to increase the linac pulse frequency or energy flux per pulse. The
Mohapatra, Sonali; Dandapat, Snigdha Jyotsna; Thatoi, Hrudayanath
2017-02-01
Acid as well as ultrasono-assisted acid pretreatment of lignocellulosic biomass of two Pennisetum sp.; Denanath grass (DG) and Hybrid Napier grass (HNG) have been investigated for enhanced delignification and maximum exposure of cellulose for production of bioethanol. Screening of pretreatment with different acids such as H 2 SO 4 , HCl, H 3 PO 4 and H 2 NO 3 were optimized for different temperature, soaking time and acid concentrations using Taguchi orthogonal array and the data obtained were statistically validated using artificial neural networking. HCl was found to be the most effective acid for pretreatment of both the Pennisetum sp. The optimized conditions of HCl pretreatment were acid concentration of 1% and 1.5%, soaking time 130 and 50 min and temperature 121 °C and 110 °C which yielded maximum delignification of 33.0% and 33.8% for DG and HNG respectively. Further ultrosono-assisted HCl pretreatment with a power supply of 100 W, temperature of 353 K, and duty cycle of 70% has resulted in significantly higher delignification of 80.4% and 82.1% for both DG and HNG respectively than that of acid pretreatment. Investigation using SEM, FTIR and autofloresence microscopy for both acid and ultrasono-assisted acid pretreatment lignocellulosic biomass revealed conformational changes of pretreated lignocellulosic biomass with decreased lignin content and increased exposure of cellulose, with greater effectiveness in case of ultrasono assisted acid pretreatment condition. Copyright © 2016. Published by Elsevier Ltd.
The development of additive manufacturing technique for nickel-base alloys: A review
NASA Astrophysics Data System (ADS)
Zadi-Maad, Ahmad; Basuki, Arif
2018-04-01
Nickel-base alloys are an attractive alloy due to its excellent mechanical properties, a high resistance to creep deformation, corrosion, and oxidation. However, it is a hard task to control performance when casting or forging for this material. In recent years, additive manufacturing (AM) process has been implemented to replace the conventional directional solidification process for the production of nickel-base alloys. Due to its potentially lower cost and flexibility manufacturing process, AM is considered as a substitute technique for the existing. This paper provides a comprehensive review of the previous work related to the AM techniques for Ni-base alloys while highlighting current challenges and methods to solving them. The properties of conventionally manufactured Ni-base alloys are also compared with the AM fabricated alloys. The mechanical properties obtained from tension, hardness and fatigue test are included, along with discussions of the effect of post-treatment process. Recommendations for further work are also provided.
NASA Astrophysics Data System (ADS)
Asghari, E.; Ashassi-Sorkhabi, H.; Ahangari, M.; Bagheri, R.
2016-04-01
Factors such as inhibitor concentration, solution hydrodynamics, and temperature influence the performance of corrosion inhibitor mixtures. The simultaneous studying of the impact of different factors is a time- and cost-consuming process. The use of experimental design methods can be useful in minimizing the number of experiments and finding local optimized conditions for factors under the investigation. In the present work, the inhibition performance of a three-component inhibitor mixture against corrosion of St37 steel rotating disk electrode, RDE, was studied. The mixture was composed of citric acid, lanthanum(III) nitrate, and tetrabutylammonium perchlorate. In order to decrease the number of experiments, the L16 Taguchi orthogonal array was used. The "control factors" were the concentration of each component and the rotation rate of RDE and the "response factor" was the inhibition efficiency. The scanning electron microscopy and energy dispersive x-ray spectroscopy techniques verified the formation of islands of adsorbed citrate complexes with lanthanum ions and insoluble lanthanum(III) hydroxide. From the Taguchi analysis results the mixture of 0.50 mM lanthanum(III) nitrate, 0.50 mM citric acid, and 2.0 mM tetrabutylammonium perchlorate under the electrode rotation rate of 1000 rpm was found as optimum conditions.
Photoacoustic Techniques for Trace Gas Sensing Based on Semiconductor Laser Sources
Elia, Angela; Lugarà, Pietro Mario; Di Franco, Cinzia; Spagnolo, Vincenzo
2009-01-01
The paper provides an overview on the use of photoacoustic sensors based on semiconductor laser sources for the detection of trace gases. We review the results obtained using standard, differential and quartz enhanced photoacoustic techniques. PMID:22303143
A preclustering-based ensemble learning technique for acute appendicitis diagnoses.
Lee, Yen-Hsien; Hu, Paul Jen-Hwa; Cheng, Tsang-Hsiang; Huang, Te-Chia; Chuang, Wei-Yao
2013-06-01
Acute appendicitis is a common medical condition, whose effective, timely diagnosis can be difficult. A missed diagnosis not only puts the patient in danger but also requires additional resources for corrective treatments. An acute appendicitis diagnosis constitutes a classification problem, for which a further fundamental challenge pertains to the skewed outcome class distribution of instances in the training sample. A preclustering-based ensemble learning (PEL) technique aims to address the associated imbalanced sample learning problems and thereby support the timely, accurate diagnosis of acute appendicitis. The proposed PEL technique employs undersampling to reduce the number of majority-class instances in a training sample, uses preclustering to group similar majority-class instances into multiple groups, and selects from each group representative instances to create more balanced samples. The PEL technique thereby reduces potential information loss from random undersampling. It also takes advantage of ensemble learning to improve performance. We empirically evaluate this proposed technique with 574 clinical cases obtained from a comprehensive tertiary hospital in southern Taiwan, using several prevalent techniques and a salient scoring system as benchmarks. The comparative results show that PEL is more effective and less biased than any benchmarks. The proposed PEL technique seems more sensitive to identifying positive acute appendicitis than the commonly used Alvarado scoring system and exhibits higher specificity in identifying negative acute appendicitis. In addition, the sensitivity and specificity values of PEL appear higher than those of the investigated benchmarks that follow the resampling approach. Our analysis suggests PEL benefits from the more representative majority-class instances in the training sample. According to our overall evaluation results, PEL records the best overall performance, and its area under the curve measure reaches 0.619. The
Multiple Uses of a Word Study Technique
ERIC Educational Resources Information Center
Joseph, Laurice M.; Orlins, Andrew
2005-01-01
This paper presents two case studies that illustrate the multiple uses of word sorts, a word study phonics technique. Case study children were Sara, a second grader, who had difficulty with reading basic words and John, a third grader, who had difficulty with spelling basic words. Multiple baseline designs were employed to study the effects of…
Initial planetary base construction techniques and machine implementation
NASA Technical Reports Server (NTRS)
Crockford, William W.
1987-01-01
Conceptual designs of (1) initial planetary base structures, and (2) an unmanned machine to perform the construction of these structures using materials local to the planet are presented. Rock melting is suggested as a possible technique to be used by the machine in fabricating roads, platforms, and interlocking bricks. Identification of problem areas in machine design and materials processing is accomplished. The feasibility of the designs is contingent upon favorable results of an analysis of the engineering behavior of the product materials. The analysis requires knowledge of several parameters for solution of the constitutive equations of the theory of elasticity. An initial collection of these parameters is presented which helps to define research needed to perform a realistic feasibility study. A qualitative approach to estimating power and mass lift requirements for the proposed machine is used which employs specifications of currently available equipment. An initial, unmanned mission scenario is discussed with emphasis on identifying uncompleted tasks and suggesting design considerations for vehicles and primitive structures which use the products of the machine processing.
Antimisting kerosene: Base fuel effects, blending and quality control techniques
NASA Technical Reports Server (NTRS)
Yavrouian, A. H.; Ernest, J.; Sarohia, V.
1984-01-01
The problems associated with blending of the AMK additive with Jet A, and the base fuel effects on AMK properties are addressed. The results from the evaluation of some of the quality control techniques for AMK are presented. The principal conclusions of this investigation are: significant compositional differences for base fuel (Jet A) within the ASTM specification DI655; higher aromatic content of the base fuel was found to be beneficial for the polymer dissolution at ambient (20 C) temperature; using static mixer technology, the antimisting additive (FM-9) is in-line blended with Jet A, producing AMK which has adequate fire-protection properties 15 to 20 minutes after blending; degradability of freshly blended and equilibrated AMK indicated that maximum degradability is reached after adequate fire protection is obtained; the results of AMK degradability as measured by filter ratio, confirmed previous RAE data that power requirements to decade freshly blended AMK are significantly higher than equilibrated AMK; blending of the additive by using FM-9 concentrate in Jet A produces equilibrated AMK almost instantly; nephelometry offers a simple continuous monitoring capability and is used as a real time quality control device for AMK; and trajectory (jet thurst) and pressure drop tests are useful laboratory techniques for evaluating AMK quality.
NASA Astrophysics Data System (ADS)
Zaripov, D. I.; Renfu, Li
2018-05-01
The implementation of high-efficiency digital image correlation methods based on a zero-normalized cross-correlation (ZNCC) procedure for high-speed, time-resolved measurements using a high-resolution digital camera is associated with big data processing and is often time consuming. In order to speed-up ZNCC computation, a high-speed technique based on a parallel projection correlation procedure is proposed. The proposed technique involves the use of interrogation window projections instead of its two-dimensional field of luminous intensity. This simplification allows acceleration of ZNCC computation up to 28.8 times compared to ZNCC calculated directly, depending on the size of interrogation window and region of interest. The results of three synthetic test cases, such as a one-dimensional uniform flow, a linear shear flow and a turbulent boundary-layer flow, are discussed in terms of accuracy. In the latter case, the proposed technique is implemented together with an iterative window-deformation technique. On the basis of the results of the present work, the proposed technique is recommended to be used for initial velocity field calculation, with further correction using more accurate techniques.
A Comparative Study between Universal Eclectic Septoplasty Technique and Cottle
Amaral Neto, Odim Ferreira do; Mizoguchi, Flavio Massao; Freitas, Renato da Silva; Maniglia, João Jairney; Maniglia, Fábio Fabrício; Maniglia, Ricardo Fabrício
2017-01-01
Introduction Since the last century surgical correction of nasal septum deviation has been improved. The Universal Eclectic Technique was recently reported and there are still few studies dedicated to address this surgical approach. Objective The objective of this study is to compare the results of septal deviation correction achieved using the Universal Eclectic Technique (UET) with those obtained through Cottle's Technique. Methods This is a prospective study with two consecutive case series totaling 90 patients (40 women and 50 men), aged between 18 and 55 years. We divided patients into two groups according to the surgical approach. Fifty-three patients underwent septoplasty through Universal Eclectic Technique (UET) and thirty-seven patients were submitted to classical Cottle's septoplasty technique. All patients have answered the Nasal Obstruction Symptom Evaluation Scale (NOSE) questionnaire to assess pre and postoperative nasal obstruction. Results Statistical analysis showed a significantly shorter operating time for the UET group. Nasal edema assessment performed seven days after the surgery showed a prevalence of mild edema in UET group and moderate edema in Cottle's technique group. In regard to complication rates, UET presented a single case of septal hematoma while in Cottle's technique group we observed: 02 cases of severe edemas, 01 case of incapacitating headache, and 01 complaint of nasal pain. Conclusion The Universal Eclectic Technique (UET) has proven to be a safe and effective surgical technique with faster symptomatic improvement, low complication rates, and reduced surgical time when compared with classical Cottle's technique. PMID:28680499
Theorists and Techniques: Connecting Education Theories to Lamaze Teaching Techniques
Podgurski, Mary Jo
2016-01-01
ABSTRACT Should childbirth educators connect education theory to technique? Is there more to learning about theorists than memorizing facts for an assessment? Are childbirth educators uniquely poised to glean wisdom from theorists and enhance their classes with interactive techniques inspiring participant knowledge and empowerment? Yes, yes, and yes. This article will explore how an awareness of education theory can enhance retention of material through interactive learning techniques. Lamaze International childbirth classes already prepare participants for the childbearing year by using positive group dynamics; theory will empower childbirth educators to address education through well-studied avenues. Childbirth educators can provide evidence-based learning techniques in their classes and create true behavioral change. PMID:26848246
A relative quantitative assessment of myocardial perfusion by first-pass technique: animal study
NASA Astrophysics Data System (ADS)
Chen, Jun; Zhang, Zhang; Yu, Xuefang; Zhou, Kenneth J.
2015-03-01
The purpose of this study is to quantitatively assess the myocardial perfusion by first-pass technique in swine model. Numerous techniques based on the analysis of Computed Tomography (CT) Hounsfield Unit (HU) density have emerged. Although these methods proposed to be able to assess haemodynamically significant coronary artery stenosis, their limitations are noticed. There are still needs to develop some new techniques. Experiments were performed upon five (5) closed-chest swine. Balloon catheters were placed into the coronary artery to simulate different degrees of luminal stenosis. Myocardial Blood Flow (MBF) was measured using color microsphere technique. Fractional Flow Reserve (FFR) was measured using pressure wire. CT examinations were performed twice during First-pass phase under adenosine-stress condition. CT HU Density (HUDCT) and CT HU Density Ratio (HUDRCT) were calculated using the acquired CT images. Our study presents that HUDRCT shows a good (y=0.07245+0.09963x, r2=0.898) correlation with MBF and FFR. In receiver operating characteristic (ROC) curve analyses, HUDRCT provides excellent diagnostic performance for the detection of significant ischemia during adenosine-stress as defined by FFR indicated by the value of Area Under the Curve (AUC) of 0.927. HUDRCT has the potential to be developed as a useful indicator of quantitative assessment of myocardial perfusion.
Content based image retrieval using local binary pattern operator and data mining techniques.
Vatamanu, Oana Astrid; Frandeş, Mirela; Lungeanu, Diana; Mihalaş, Gheorghe-Ioan
2015-01-01
Content based image retrieval (CBIR) concerns the retrieval of similar images from image databases, using feature vectors extracted from images. These feature vectors globally define the visual content present in an image, defined by e.g., texture, colour, shape, and spatial relations between vectors. Herein, we propose the definition of feature vectors using the Local Binary Pattern (LBP) operator. A study was performed in order to determine the optimum LBP variant for the general definition of image feature vectors. The chosen LBP variant is then subsequently used to build an ultrasound image database, and a database with images obtained from Wireless Capsule Endoscopy. The image indexing process is optimized using data clustering techniques for images belonging to the same class. Finally, the proposed indexing method is compared to the classical indexing technique, which is nowadays widely used.
Efficient low-bit-rate adaptive mesh-based motion compensation technique
NASA Astrophysics Data System (ADS)
Mahmoud, Hanan A.; Bayoumi, Magdy A.
2001-08-01
This paper proposes a two-stage global motion estimation method using a novel quadtree block-based motion estimation technique and an active mesh model. In the first stage, motion parameters are estimated by fitting block-based motion vectors computed using a new efficient quadtree technique, that divides a frame into equilateral triangle blocks using the quad-tree structure. Arbitrary partition shapes are achieved by allowing 4-to-1, 3-to-1 and 2-1 merge/combine of sibling blocks having the same motion vector . In the second stage, the mesh is constructed using an adaptive triangulation procedure that places more triangles over areas with high motion content, these areas are estimated during the first stage. finally the motion compensation is achieved by using a novel algorithm that is carried by both the encoder and the decoder to determine the optimal triangulation of the resultant partitions followed by affine mapping at the encoder. Computer simulation results show that the proposed method gives better performance that the conventional ones in terms of the peak signal-to-noise ration (PSNR) and the compression ratio (CR).
Preliminary studies on the planetary entry to Jupiter by aerocapture technique
NASA Astrophysics Data System (ADS)
Aso, Shigeru; Yasaka, Tetsuo; Hirayama, Hiroshi; Poetro, Ridanto Eko; Hatta, Shinji
2006-10-01
Preliminary studies on the planetary entry to Jupiter by aerocapture technique are studied in order to complete technological challenges to deliver scientific probe with low cost and smaller mass of the spacecraft to Jupiter. Jupiter aerocapture corridor determination based on maximum deceleration limit of 5g (lower corridor) and aerocapture capability (upper corridor) at Jupiter are carefully considered and calculated. The results show about 1700 m/s of saving velocity due to aerocapture could be possible in some cases for the spacecraft to be captured by Jovian gravitational field. However, the results also show that Jovian aerocapture is not available in some cases. Hence, careful selection is needed to realize Jovian aerocapture. Also the numerical simulation of aerodynamic heating to the spacecraft has been conducted. DSMC method is used for the simulation of flow fields around the spacecraft. The transient changes of drag due to Jovian atmosphere and total heat loads to the spacecraft are obtained. The results show that the estimated heat loads could be within allowable amount heat load when some ablation heat shield technique is applied.
Preliminary studies on the planetary entry to Jupiter by aerocapture technique
NASA Astrophysics Data System (ADS)
Aso, Shigeru; Yasaka, Tetsuo; Hirayama, Hiroshi; Eko Poetro, Ridanto; Hatta, Shinji
2003-11-01
Preliminary studies on the planetary entry to Jupiter by aerocapture technique are studied in order to complete technological challenges to deliver scientific probe with low cost and smaller mass of the spacecraft to Jupiter. Jupiter aerocapture corridor determination based on maximum deceleration limit of 5g (lower corridor) and aerocapture capability (upper corridor) at Jupiter are carefully considered and calculated. The results show about 1700 m/s of saving velocity due to aerocapture could be possible in some cases for the spacecraft to be captured by Jovian gravitational field. However, the results also show that Jovian aerocapture is not available in some cases. Hence, careful selection is needed to realise Jovian aerocapture. Also the numerical simulation of aerodynamic heating to the spacecraft has been conducted. DSMC method is used for the simulation of flow fields around the spacecraft. The transient changes of drag due to Jovian atmosphere and total heat loads to the spacecraft are obtained. The results show the estimated heat loads could be within allowable amount heat load when some ablation heat shield technique is applied.
Comparative study of nail sampling techniques in onychomycosis.
Shemer, Avner; Davidovici, Batya; Grunwald, Marcelo H; Trau, Henri; Amichai, Boaz
2009-07-01
Onychomycosis is a common problem. Obtaining accurate laboratory test results before treatment is important in clinical practice. The purpose of this study was to compare results of curettage and drilling techniques of nail sampling in the diagnosis of onychomycosis, and to establish the best technique and location of sampling. We evaluated 60 patients suffering from distal and lateral subungual onychomycosis and lateral subungual onychomycosis using curettage and vertical and horizontal drilling sampling techniques from three different sites of the infected nail. KOH examination and fungal culture were used for detection and identification of fungal infection. At each sample site, the horizontal drilling technique has a better culture sensitivity than curettage. Trichophyton rubrum was by far the most common pathogen detected by both techniques from all sampling sites. The drilling technique was found to be statistically better than curettage at each site of sampling, furthermore vertical drilling from the proximal part of the affected nail was found to be the best procedure for nail sampling. With each technique we found that the culture sensitivity improved as the location of the sample was more proximal. More types of pathogens were detected in samples taken by both methods from proximal parts of the affected nails.
NASA Astrophysics Data System (ADS)
Raj, Anil; Wins, K. Leo Dev; Varadarajan, A. S.
2016-09-01
Surface roughness is one of the important parameters, which not only affects the service life of a component but also serves as a good index of machinability. Near Dry Machining, methods (NDM) are considered as sustainable alternative for workshops trying to bring down their dependence on cutting fluids and the hazards associated with their indiscriminate usage. The present work presents a comparison of the surface roughness and chip characteristics during hard turning of AISI H13 tool work steel using hard metal inserts under two popular NDM techniques namely the minimal fluid application and the Minimum Quantity Lubrication technique(MQL) using an experiment designed based on Taguchi's techniques. The statistical method of analysis of variance (ANOVA) was used to determine the relative significance of input parameters consisting of cutting speed, feed and depth of cut on the attainable surface finish and the chip characteristics. It was observed that the performance during minimal fluid application was better than that during MQL application.
Lakbub, Jude C; Shipman, Joshua T; Desaire, Heather
2018-04-01
Disulfide bonds are important structural moieties of proteins: they ensure proper folding, provide stability, and ensure proper function. With the increasing use of proteins for biotherapeutics, particularly monoclonal antibodies, which are highly disulfide bonded, it is now important to confirm the correct disulfide bond connectivity and to verify the presence, or absence, of disulfide bond variants in the protein therapeutics. These studies help to ensure safety and efficacy. Hence, disulfide bonds are among the critical quality attributes of proteins that have to be monitored closely during the development of biotherapeutics. However, disulfide bond analysis is challenging because of the complexity of the biomolecules. Mass spectrometry (MS) has been the go-to analytical tool for the characterization of such complex biomolecules, and several methods have been reported to meet the challenging task of mapping disulfide bonds in proteins. In this review, we describe the relevant, recent MS-based techniques and provide important considerations needed for efficient disulfide bond analysis in proteins. The review focuses on methods for proper sample preparation, fragmentation techniques for disulfide bond analysis, recent disulfide bond mapping methods based on the fragmentation techniques, and automated algorithms designed for rapid analysis of disulfide bonds from liquid chromatography-MS/MS data. Researchers involved in method development for protein characterization can use the information herein to facilitate development of new MS-based methods for protein disulfide bond analysis. In addition, individuals characterizing biotherapeutics, especially by disulfide bond mapping in antibodies, can use this review to choose the best strategies for disulfide bond assignment of their biologic products. Graphical Abstract This review, describing characterization methods for disulfide bonds in proteins, focuses on three critical components: sample preparation, mass
Ekinci, Saniye; Çiftçi, Arbay Özden; Karnak, İbrahim; Şenocak, Mehmet Emin
2016-04-01
Hypospadias is a common congenital anomaly. Over 300 techniques have been described for repairing hypospadias. Eccentric circummeatal based flap with combined limited urethral mobilization technique (ECMB-LUM) is a simple procedure to repair distal hypospadias with minimal complication rate. This study presents results of this technique, highlighting surgical pitfalls to achieve the best result. Medical records of patients with distal hypospadias operated on using the same technique between 1998 and 2011 were reviewed retrospectively. Age at surgery, position of meatus preoperatively and postoperatively, duration of urethral catheterization and hospitalization, early and late complications, previous hypospadias repairs, and secondary surgical interventions were evaluated. In the surgical technique an eccentric circummeatal based flap is outlined. The proximal part of the flap is dissected from the underlying urethra and Buck's fascia. If the flap is not long enough, the distal urethra is mobilized a few millimeters (Figure). The eccentric flap is sutured to the tip of the glans. The glans wings are approximated in the midline. A urethral catheter of 6 Fr or 8 Fr is passed and left in the bulbous urethra or the urinary bladder. Diverged limbs of corpus spongiosum are approximated on the urethra, then, the glans and skin of the penile shaft are sutured. Of the 171 consecutive patients operated on using the ECMB-LUM technique; 115 had coronal, 47 had subcoronal, and nine had glanular meatus. The mean age at surgery was 4.5 (1-17) years. Patients were hospitalized for 2.2 ± 0.7 days. Mean duration of urethral catheterization was 2.3 ± 0.5 days. All but eight patients had ECBF-LUM as primary repair. There were no early complications such as bleeding, hematoma, and wound infection. All patients voided spontaneously after catheter removal. Late complications were meatal stenosis, urethrocutaneous fistula, meatal regression, and glandular dehiscence (Table). These
Kumar, M Praveen; Patil, Suneel G; Dheeraj, Bhandari; Reddy, Keshav; Goel, Dinker; Krishna, Gopi
2015-01-01
Background: The difficulty in obtaining an acceptable impression increases exponentially as the number of abutments increases. Accuracy of the impression material and the use of a suitable impression technique are of utmost importance in the fabrication of a fixed partial denture. This study compared the accuracy of the matrix impression system with conventional putty reline and multiple mix technique for individual dies by comparing the inter-abutment distance in the casts obtained from the impressions. Materials and Methods: Three groups, 10 impressions each with three impression techniques (matrix impression system, putty reline technique and multiple mix technique) were made of a master die. Typodont teeth were embedded in a maxillary frasaco model base. The left first premolar was removed to create a three-unit fixed partial denture situation and the left canine and second premolar were prepared conservatively, and hatch marks were made on the abutment teeth. The final casts obtained from the impressions were examined under a profile projector and the inter-abutment distance was calculated for all the casts and compared. Results: The results from this study showed that in the mesiodistal dimensions the percentage deviation from master model in Group I was 0.1 and 0.2, in Group II was 0.9 and 0.3, and Group III was 1.6 and 1.5, respectively. In the labio-palatal dimensions the percentage deviation from master model in Group I was 0.01 and 0.4, Group II was 1.9 and 1.3, and Group III was 2.2 and 2.0, respectively. In the cervico-incisal dimensions the percentage deviation from the master model in Group I was 1.1 and 0.2, Group II was 3.9 and 1.7, and Group III was 1.9 and 3.0, respectively. In the inter-abutment dimension of dies, percentage deviation from master model in Group I was 0.1, Group II was 0.6, and Group III was 1.0. Conclusion: The matrix impression system showed more accuracy of reproduction for individual dies when compared with putty reline
Kumar, M Praveen; Patil, Suneel G; Dheeraj, Bhandari; Reddy, Keshav; Goel, Dinker; Krishna, Gopi
2015-06-01
The difficulty in obtaining an acceptable impression increases exponentially as the number of abutments increases. Accuracy of the impression material and the use of a suitable impression technique are of utmost importance in the fabrication of a fixed partial denture. This study compared the accuracy of the matrix impression system with conventional putty reline and multiple mix technique for individual dies by comparing the inter-abutment distance in the casts obtained from the impressions. Three groups, 10 impressions each with three impression techniques (matrix impression system, putty reline technique and multiple mix technique) were made of a master die. Typodont teeth were embedded in a maxillary frasaco model base. The left first premolar was removed to create a three-unit fixed partial denture situation and the left canine and second premolar were prepared conservatively, and hatch marks were made on the abutment teeth. The final casts obtained from the impressions were examined under a profile projector and the inter-abutment distance was calculated for all the casts and compared. The results from this study showed that in the mesiodistal dimensions the percentage deviation from master model in Group I was 0.1 and 0.2, in Group II was 0.9 and 0.3, and Group III was 1.6 and 1.5, respectively. In the labio-palatal dimensions the percentage deviation from master model in Group I was 0.01 and 0.4, Group II was 1.9 and 1.3, and Group III was 2.2 and 2.0, respectively. In the cervico-incisal dimensions the percentage deviation from the master model in Group I was 1.1 and 0.2, Group II was 3.9 and 1.7, and Group III was 1.9 and 3.0, respectively. In the inter-abutment dimension of dies, percentage deviation from master model in Group I was 0.1, Group II was 0.6, and Group III was 1.0. The matrix impression system showed more accuracy of reproduction for individual dies when compared with putty reline technique and multiple mix technique in all the three
2017-11-01
ARL-TR-8225 ● NOV 2017 US Army Research Laboratory Methodology for Designing and Developing a New Ultra-Wideband Antenna Based...Research Laboratory Methodology for Designing and Developing a New Ultra-Wideband Antenna Based on Bio-Inspired Optimization Techniques by...SUBTITLE Methodology for Designing and Developing a New Ultra-Wideband Antenna Based on Bio-Inspired Optimization Techniques 5a. CONTRACT NUMBER
Comparison of carpal tunnel injection techniques: a cadaver study.
Ozturk, Kahraman; Esenyel, Cem Zeki; Sonmez, Mesut; Esenyel, Meltem; Kahraman, Sinan; Senel, Berna
2008-01-01
The purpose of the study was to evaluate the accuracy of injections into the carpal tunnel using three different portals in cadavers, and to define safe guidelines. In this study, 150 wrists of 75 cadavers (54 male, 21 female) were included. To compare three injection sites, 50 wrists of 25 cadavers were used for each technique; we used 23 gauge needles, and acrylic dye. The first injection technique: the needle was inserted 1cm proximal to the wrist crease and directed distally by roughly 45 in an ulnar direction through the flexor carpi radialis tendon. The second injection technique: the needle was inserted into the carpal tunnel from a point just ulnar to the palmaris longus tendon and 1cm proximal to the wrist crease. The third injection technique: the needle was inserted just distal to the distal skin crease of the wrist in line with the fourth ray. The first injection technique gave the highest accuracy rate, and this was also the safest injection site. Median nerve injuries caused by injection was seen mostly with the second technique. Although a steroid injection may provide symptomatic relief in patients with carpal tunnel syndrome, the median nerve and other structures in the carpal tunnel are at risk of injury. Because of that, the injection should be given using the correct technique by physicians skilled in carpal tunnel surgery.
Kim, Mi Jeong; Maeng, Sung Joon; Cho, Yong Soo
2015-01-01
In this paper, a distributed synchronization technique based on a bio-inspired algorithm is proposed for an orthogonal frequency division multiple access (OFDMA)-based wireless mesh network (WMN) with a time difference of arrival. The proposed time- and frequency-synchronization technique uses only the signals received from the neighbor nodes, by considering the effect of the propagation delay between the nodes. It achieves a fast synchronization with a relatively low computational complexity because it is operated in a distributed manner, not requiring any feedback channel for the compensation of the propagation delays. In addition, a self-organization scheme that can be effectively used to construct 1-hop neighbor nodes is proposed for an OFDMA-based WMN with a large number of nodes. The performance of the proposed technique is evaluated with regard to the convergence property and synchronization success probability using a computer simulation. PMID:26225974
Kim, Mi Jeong; Maeng, Sung Joon; Cho, Yong Soo
2015-07-28
In this paper, a distributed synchronization technique based on a bio-inspired algorithm is proposed for an orthogonal frequency division multiple access (OFDMA)-based wireless mesh network (WMN) with a time difference of arrival. The proposed time- and frequency-synchronization technique uses only the signals received from the neighbor nodes, by considering the effect of the propagation delay between the nodes. It achieves a fast synchronization with a relatively low computational complexity because it is operated in a distributed manner, not requiring any feedback channel for the compensation of the propagation delays. In addition, a self-organization scheme that can be effectively used to construct 1-hop neighbor nodes is proposed for an OFDMA-based WMN with a large number of nodes. The performance of the proposed technique is evaluated with regard to the convergence property and synchronization success probability using a computer simulation.
Cihandide, Ercan; Kayiran, Oguz; Aydin, Elif Eren; Uzunismail, Adnan
2016-06-01
Otoplasty techniques are generally divided into 2 categories as cartilage-cutting and cartilage-sparing. The cartilage-cutting techniques have been criticized because of their high risk of hematoma, skin necrosis, and ear deformity. As a result, suture-based cartilage-sparing methods like Mustardé and Furnas-type suture techniques have become increasingly popular. However, with these techniques postauricular suture extrusion may be seen and recurrence rates of up to 25% have been reported. In this study, cartilage-sparing otoplasty is redefined by introduction of the distally based perichondrio-adipo-dermal flap which is elevated from the postauricular region. Thirty-seven ears (17 bilateral and 3 unilateral) in 20 patients (14 females and 6 males) have been operated with the defined technique by the same surgeon. The distally based perichondrio-adipo-dermal flap is advanced posteriorly to correct the deformity, also acting as a strong postauricular support to prevent recurrence. In addition to the resultant natural-looking antihelical fold, the posterior advancement of the flap corrects both the otherwise wide conchoscaphal and conchomastoid angles. The operative technique is explained in detail with results and the literature is reviewed. There were no hematomas. After an average follow-up of 8.3 months (2-16 months), recurrence was seen in only 1 patient who requested no further surgery. No patients developed suture extrusion or granuloma. The authors introduce a simple and safe procedure to correct prominent ears with benefits including a resultant natural-looking antihelical fold and less tissue trauma. The distally based perichondrio-adipo-dermal flap seems to prevent suture extrusion and may also help to reduce recurrence rates. By forming neochondrogenesis which is stimulated by elevation of the perichondrium, this flap gives the promise of longer durability of the newly formed antihelical fold.
Imputatoin and Model-Based Updating Technique for Annual Forest Inventories
Ronald E. McRoberts
2001-01-01
The USDA Forest Service is developing an annual inventory system to establish the capability of producing annual estimates of timber volume and related variables. The inventory system features measurement of an annual sample of field plots with options for updating data for plots measured in previous years. One imputation and two model-based updating techniques are...
Urbanová, Petra; Hejna, Petr; Jurda, Mikoláš
2015-05-01
Three-dimensional surface technologies particularly close range photogrammetry and optical surface scanning have recently advanced into affordable, flexible and accurate techniques. Forensic postmortem investigation as performed on a daily basis, however, has not yet fully benefited from their potentials. In the present paper, we tested two approaches to 3D external body documentation - digital camera-based photogrammetry combined with commercial Agisoft PhotoScan(®) software and stereophotogrammetry-based Vectra H1(®), a portable handheld surface scanner. In order to conduct the study three human subjects were selected, a living person, a 25-year-old female, and two forensic cases admitted for postmortem examination at the Department of Forensic Medicine, Hradec Králové, Czech Republic (both 63-year-old males), one dead to traumatic, self-inflicted, injuries (suicide by hanging), the other diagnosed with the heart failure. All three cases were photographed in 360° manner with a Nikon 7000 digital camera and simultaneously documented with the handheld scanner. In addition to having recorded the pre-autopsy phase of the forensic cases, both techniques were employed in various stages of autopsy. The sets of collected digital images (approximately 100 per case) were further processed to generate point clouds and 3D meshes. Final 3D models (a pair per individual) were counted for numbers of points and polygons, then assessed visually and compared quantitatively using ICP alignment algorithm and a cloud point comparison technique based on closest point to point distances. Both techniques were proven to be easy to handle and equally laborious. While collecting the images at autopsy took around 20min, the post-processing was much more time-demanding and required up to 10h of computation time. Moreover, for the full-body scanning the post-processing of the handheld scanner required rather time-consuming manual image alignment. In all instances the applied approaches
Digital education and dynamic assessment of tongue diagnosis based on Mashup technique.
Tsai, Chin-Chuan; Lo, Yen-Cheng; Chiang, John Y; Sainbuyan, Natsagdorj
2017-01-24
To assess the digital education and dynamic assessment of tongue diagnosis based on Mashup technique (DEDATD) according to specifific user's answering pattern, and provide pertinent information tailored to user's specifific needs supplemented by the teaching materials constantly updated through the Mashup technique. Fifty-four undergraduate students were tested with DEDATD developed. The effificacy of the DEDATD was evaluated based on the pre- and post-test performance, with interleaving training sessions targeting on the weakness of the student under test. The t-test demonstrated that signifificant difference was reached in scores gained during pre- and post-test sessions, and positive correlation between scores gained and length of time spent on learning, while no signifificant differences between the gender and post-test score, and the years of students in school and the progress in score gained. DEDATD, coupled with Mashup technique, could provide updated materials fifiltered through diverse sources located across the network. The dynamic assessment could tailor each individual learner's needs to offer custom-made learning materials. DEDATD poses as a great improvement over the traditional teaching methods.
New MPLS network management techniques based on adaptive learning.
Anjali, Tricha; Scoglio, Caterina; de Oliveira, Jaudelice Cavalcante
2005-09-01
The combined use of the differentiated services (DiffServ) and multiprotocol label switching (MPLS) technologies is envisioned to provide guaranteed quality of service (QoS) for multimedia traffic in IP networks, while effectively using network resources. These networks need to be managed adaptively to cope with the changing network conditions and provide satisfactory QoS. An efficient strategy is to map the traffic from different DiffServ classes of service on separate label switched paths (LSPs), which leads to distinct layers of MPLS networks corresponding to each DiffServ class. In this paper, three aspects of the management of such a layered MPLS network are discussed. In particular, an optimal technique for the setup of LSPs, capacity allocation of the LSPs and LSP routing are presented. The presented techniques are based on measurement of the network state to adapt the network configuration to changing traffic conditions.
Interactions between heavy metals and photosynthetic materials studied by optical techniques.
Ventrella, Andrea; Catucci, Lucia; Piletska, Elena; Piletsky, Sergey; Agostiano, Angela
2009-11-01
In this work studies on rapid inhibitory interactions between heavy metals and photosynthetic materials at different organization levels were carried out by optical assay techniques, investigating the possibility of applications in the heavy metal detection field. Spinach chloroplasts, thylakoids and Photosystem II proteins were employed as biotools in combination with colorimetric assays based on dichlorophenol indophenole (DCIP) photoreduction and on fluorescence emission techniques. It was found that copper and mercury demonstrated a strong and rapid photosynthetic activity inhibition, that varied from proteins to membranes, while other metals like nickel, cobalt and manganese produced only slight inhibition effects on all tested photosynthetic materials. By emission measurements, only copper was found to rapidly influence the photosynthetic material signals. These findings give interesting information about the rapid effects of heavy metals on isolated photosynthetic samples, and are in addition to the literature data concerning the effects of growth in heavy metal enriched media.
Measurements of strain at plate boundaries using space based geodetic techniques
NASA Technical Reports Server (NTRS)
Robaudo, Stefano; Harrison, Christopher G. A.
1993-01-01
We have used the space based geodetic techniques of Satellite Laser Ranging (SLR) and VLBI to study strain along subduction and transform plate boundaries and have interpreted the results using a simple elastic dislocation model. Six stations located behind island arcs were analyzed as representative of subduction zones while 13 sites located on either side of the San Andreas fault were used for the transcurrent zones. The length deformation scale was then calculated for both tectonic margins by fitting the relative strain to an exponentially decreasing function of distance from the plate boundary. Results show that space-based data for the transcurrent boundary along the San Andreas fault help to define better the deformation length scale in the area while fitting nicely the elastic half-space earth model. For subduction type bonndaries the analysis indicates that there is no single scale length which uniquely describes the deformation. This is mainly due to the difference in subduction characteristics for the different areas.
Hopkins, D S; Phoenix, R D; Abrahamsen, T C
1997-09-01
A technique for the fabrication of light-activated maxillary record bases is described. The use of a segmental polymerization process provides improved palatal adaptation by minimizing the effects of polymerization shrinkage. Utilization of this technique results in record bases that are well adapted to the corresponding master casts.
Osteotomy in direct sinus lift. A comparative study of the rotary technique and ultrasound
Peñarrocha-Diago, Miguel; Sanchez-Recio, Cristina; Peñarrocha-Oltra, David; Romero-Millán, Javier
2012-01-01
Purpose: The present study investigates sinus membrane rupture in direct maxillary sinus lift with the rotary technique and with ultrasound, examining the survival of implants placed after sinus augmentation, and analyzing the bone gain obtained after the operation and 12 months after placement of the prosthetic restoration. Material and Methods: A retrospective study was made of 45 patients requiring maxillary sinus lift or augmentation for implant-prosthetic rehabilitation. Use was made of the hand piece and ostectomy drills for the rotary technique, and of specific tips for ultrasound. The implant success criteria were based on those developed by Buser. The bone gain obtained as a result of sinus lift was calculated from the postoperative panoramic X-rays. Results: A total of 57 direct elevations of the maxillary sinus were carried out: 32 with the rotary technique and 25 with ultrasound. Perforations of Schneider’s membrane with the rotary technique and ultrasound occurred in 7% and 1.7% of the cases, respectively, with membrane integrity being preserved in 91.2%. Of the 100 implants placed, 5 failed after one year of follow-up in the rotary technique group, while one implant failed in the ultrasound group. The rotary technique in turn afforded a bone gain of 5.9 mm, versus 6.7 mm with ultrasound. Conclusions: Perforations of the membrane sinusal in direct lift were more frequent with the rotary technique (7%) than with ultrasound (1.7%). Implant survival and bone gain were both greater when ultrasound was used. Key words:Bone sectioning, maxillary sinus augmentation, piezosurgery. PMID:22143735
Na, Wongi S; Baek, Jongdae
2018-04-24
The birth of smart materials such as piezoelectric (PZT) transducers has aided in revolutionizing the field of structural health monitoring (SHM) based on non-destructive testing (NDT) methods. While a relatively new NDT method known as the electromechanical (EMI) technique has been investigated for more than two decades, there are still various problems that must be solved before it is applied to real structures. The technique, which has a significant potential to contribute to the creation of one of the most effective SHM systems, involves the use of a single PZT for exciting and sensing of the host structure. In this paper, studies applied for the past decade related to the EMI technique have been reviewed to understand its trend. In addition, new concepts and ideas proposed by various authors are also surveyed, and the paper concludes with a discussion of the potential directions for future works.
Na, Wongi S.; Baek, Jongdae
2018-01-01
The birth of smart materials such as piezoelectric (PZT) transducers has aided in revolutionizing the field of structural health monitoring (SHM) based on non-destructive testing (NDT) methods. While a relatively new NDT method known as the electromechanical (EMI) technique has been investigated for more than two decades, there are still various problems that must be solved before it is applied to real structures. The technique, which has a significant potential to contribute to the creation of one of the most effective SHM systems, involves the use of a single PZT for exciting and sensing of the host structure. In this paper, studies applied for the past decade related to the EMI technique have been reviewed to understand its trend. In addition, new concepts and ideas proposed by various authors are also surveyed, and the paper concludes with a discussion of the potential directions for future works. PMID:29695067
The Survey of Vision-based 3D Modeling Techniques
NASA Astrophysics Data System (ADS)
Ruan, Mingzhe
2017-10-01
This paper reviews the vision-based localization and map construction methods from the perspectives of VSLAM, SFM, 3DMax and Unity3D. It focuses on the key technologies and the latest research progress on each aspect, analyzes the advantages and disadvantages of each method, illustrates their implementation process and system framework, and further discusses the way to promote the combination for their complementary strength. Finally, the future opportunity of the combination of the four techniques is expected.
Secure positioning technique based on the encrypted visible light map
NASA Astrophysics Data System (ADS)
Lee, Y. U.; Jung, G.
2017-01-01
For overcoming the performance degradation problems of the conventional visible light (VL) positioning system, which are due to the co-channel interference by adjacent light and the irregularity of the VL reception position in the three dimensional (3-D) VL channel, the secure positioning technique based on the two dimensional (2-D) encrypted VL map is proposed, implemented as the prototype for the specific embedded positioning system, and verified by performance tests in this paper. It is shown from the test results that the proposed technique achieves the performance enhancement over 21.7% value better than the conventional one in the real positioning environment, and the well known PN code is the optimal stream encryption key for the good VL positioning.
Two Student Self-Management Techniques Applied to Data-Based Program Modification.
ERIC Educational Resources Information Center
Wesson, Caren
Two student self-management techniques, student charting and student selection of instructional activities, were applied to ongoing data-based program modification. Forty-two elementary school resource room students were assigned randomly (within teacher) to one of three treatment conditions: Teacher Chart-Teacher Select Instructional Activities…
Whole arm manipulation planning based on feedback velocity fields and sampling-based techniques.
Talaei, B; Abdollahi, F; Talebi, H A; Omidi Karkani, E
2013-09-01
Changing the configuration of a cooperative whole arm manipulator is not easy while enclosing an object. This difficulty is mainly because of risk of jamming caused by kinematic constraints. To reduce this risk, this paper proposes a feedback manipulation planning algorithm that takes grasp kinematics into account. The idea is based on a vector field that imposes perturbation in object motion inducing directions when the movement is considerably along manipulator redundant directions. Obstacle avoidance problem is then considered by combining the algorithm with sampling-based techniques. As experimental results confirm, the proposed algorithm is effective in avoiding jamming as well as obstacles for a 6-DOF dual arm whole arm manipulator. Copyright © 2013 ISA. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Bi, Chuan-Xing; Hu, Ding-Yu; Zhang, Yong-Bin; Jing, Wen-Qian
2015-06-01
In previous studies, an equivalent source method (ESM)-based technique for recovering the free sound field in a noisy environment has been successfully applied to exterior problems. In order to evaluate its performance when applied to a more general noisy environment, that technique is used to identify active sources inside cavities where the sound field is composed of the field radiated by active sources and that reflected by walls. A patch approach with two semi-closed surfaces covering the target active sources is presented to perform the measurements, and the field that would be radiated by these target active sources into free space is extracted from the mixed field by using the proposed technique, which will be further used as the input of nearfield acoustic holography for source identification. Simulation and experimental results validate the effectiveness of the proposed technique for source identification in cavities, and show the feasibility of performing the measurements with a double layer planar array.