Sample records for qa plan boiling

  1. Guidance for Quality Assurance Project Plans, EPA QA/G-5

    EPA Pesticide Factsheets

    provides guidance to EPA employees and other organizations involved in developing Quality Assurance (QA) Project Plans that address the specifications listed in EPA Requirements for QA Project Plans (QA/R-5)

  2. WE-AB-201-00: Treatment Planning System Commissioning and QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    Treatment planning systems (TPS) are a cornerstone of modern radiation therapy. Errors in their commissioning or use can have a devastating impact on many patients. To support safe and high quality care, medical physicists must conduct efficient and proper commissioning, good clinical integration, and ongoing quality assurance (QA) of the TPS. AAPM Task Group 53 and related publications have served as seminal benchmarks for TPS commissioning and QA over the past two decades. Over the same time, continuing innovations have made the TPS even more complex and more central to the clinical process. Medical goals are now expressed in termsmore » of the dose and margins around organs and tissues that are delineated from multiple imaging modalities (CT, MR and PET); and even temporally resolved (i.e., 4D) imaging. This information is passed on to optimization algorithms to establish accelerator movements that are programmed directly for IMRT, VMAT and stereotactic treatments. These advances have made commissioning and QA of the TPS much more challenging. This education session reviews up-to-date experience and guidance on this subject; including the recently published AAPM Medical Physics Practice Guideline (MPPG) #5 “Commissioning and QA of Treatment Planning Dose Calculations: Megavoltage Photon and Electron Beams”. Treatment Planning System Commissioning and QA: Challenges and Opportunities (Greg Salomons) This session will provide some key background and review publications describing prominent incidents relating to TPS commissioning and QA. Traditional approaches have been hardware and feature oriented. They aim to establish a functional configuration and establish specifications for regular testing of features (like dose calculation) to assure stable operation and detect failures. With the advent of more complex systems, more patient-specific testing has also been adopted. A number of actual TPS defects will be presented along with heuristics for identifying

  3. WE-AB-201-03: TPS Commissioning and QA: Incorporating the Entire Planning Process

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mutic, S.

    Treatment planning systems (TPS) are a cornerstone of modern radiation therapy. Errors in their commissioning or use can have a devastating impact on many patients. To support safe and high quality care, medical physicists must conduct efficient and proper commissioning, good clinical integration, and ongoing quality assurance (QA) of the TPS. AAPM Task Group 53 and related publications have served as seminal benchmarks for TPS commissioning and QA over the past two decades. Over the same time, continuing innovations have made the TPS even more complex and more central to the clinical process. Medical goals are now expressed in termsmore » of the dose and margins around organs and tissues that are delineated from multiple imaging modalities (CT, MR and PET); and even temporally resolved (i.e., 4D) imaging. This information is passed on to optimization algorithms to establish accelerator movements that are programmed directly for IMRT, VMAT and stereotactic treatments. These advances have made commissioning and QA of the TPS much more challenging. This education session reviews up-to-date experience and guidance on this subject; including the recently published AAPM Medical Physics Practice Guideline (MPPG) #5 “Commissioning and QA of Treatment Planning Dose Calculations: Megavoltage Photon and Electron Beams”. Treatment Planning System Commissioning and QA: Challenges and Opportunities (Greg Salomons) This session will provide some key background and review publications describing prominent incidents relating to TPS commissioning and QA. Traditional approaches have been hardware and feature oriented. They aim to establish a functional configuration and establish specifications for regular testing of features (like dose calculation) to assure stable operation and detect failures. With the advent of more complex systems, more patient-specific testing has also been adopted. A number of actual TPS defects will be presented along with heuristics for identifying

  4. WE-AB-201-01: Treatment Planning System Commissioning and QA: Challenges and Opportunities

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Salomons, G.

    Treatment planning systems (TPS) are a cornerstone of modern radiation therapy. Errors in their commissioning or use can have a devastating impact on many patients. To support safe and high quality care, medical physicists must conduct efficient and proper commissioning, good clinical integration, and ongoing quality assurance (QA) of the TPS. AAPM Task Group 53 and related publications have served as seminal benchmarks for TPS commissioning and QA over the past two decades. Over the same time, continuing innovations have made the TPS even more complex and more central to the clinical process. Medical goals are now expressed in termsmore » of the dose and margins around organs and tissues that are delineated from multiple imaging modalities (CT, MR and PET); and even temporally resolved (i.e., 4D) imaging. This information is passed on to optimization algorithms to establish accelerator movements that are programmed directly for IMRT, VMAT and stereotactic treatments. These advances have made commissioning and QA of the TPS much more challenging. This education session reviews up-to-date experience and guidance on this subject; including the recently published AAPM Medical Physics Practice Guideline (MPPG) #5 “Commissioning and QA of Treatment Planning Dose Calculations: Megavoltage Photon and Electron Beams”. Treatment Planning System Commissioning and QA: Challenges and Opportunities (Greg Salomons) This session will provide some key background and review publications describing prominent incidents relating to TPS commissioning and QA. Traditional approaches have been hardware and feature oriented. They aim to establish a functional configuration and establish specifications for regular testing of features (like dose calculation) to assure stable operation and detect failures. With the advent of more complex systems, more patient-specific testing has also been adopted. A number of actual TPS defects will be presented along with heuristics for identifying

  5. QA program plan plutonium stabilization and handling project W-460

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    SCHULTZ, J.W.

    This Quality Assurance Program Plan (QAPP) identifies Project Quality Assurance (QA) program requirements for all parties participating in the design, procurement, demolition, construction, installation, inspection and testing for Project W-460.

  6. Test/QA Plan for Verification of Semi-Continuous Ambient Air Monitoring Systems - Second Round

    EPA Science Inventory

    Test/QA Plan for Verification of Semi-Continuous Ambient Air Monitoring Systems - Second Round. Changes reflect performance of second round of testing at new location and with various changes to personnel. Additional changes reflect general improvements to the Version 1 test/QA...

  7. SU-F-P-07: Applying Failure Modes and Effects Analysis to Treatment Planning System QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mathew, D; Alaei, P

    2016-06-15

    Purpose: A small-scale implementation of Failure Modes and Effects Analysis (FMEA) for treatment planning system QA by utilizing methodology of AAPM TG-100 report. Methods: FMEA requires numerical values for severity (S), occurrence (O) and detectability (D) of each mode of failure. The product of these three values gives a risk priority number (RPN). We have implemented FMEA for the treatment planning system (TPS) QA for two clinics which use Pinnacle and Eclipse TPS. Quantitative monthly QA data dating back to 4 years for Pinnacle and 1 year for Eclipse have been used to determine values for severity (deviations from predeterminedmore » doses at points or volumes), and occurrence of such deviations. The TPS QA protocol includes a phantom containing solid water and lung- and bone-equivalent heterogeneities. Photon and electron plans have been evaluated in both systems. The dose values at multiple distinct points of interest (POI) within the solid water, lung, and bone-equivalent slabs, as well as mean doses to several volumes of interest (VOI), have been re-calculated monthly using the available algorithms. Results: The computed doses vary slightly month-over-month. There have been more significant deviations following software upgrades, especially if the upgrade involved re-modeling of the beams. TG-100 guidance and the data presented here suggest an occurrence (O) of 2 depending on the frequency of re-commissioning the beams, severity (S) of 3, and detectability (D) of 2, giving an RPN of 12. Conclusion: Computerized treatment planning systems could pose a risk due to dosimetric errors and suboptimal treatment plans. The FMEA analysis presented here suggests that TPS QA should immediately follow software upgrades, but does not need to be performed every month.« less

  8. Template for updating regulations in QA manuals

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    White, M.G.; Banerjee, B.

    1992-01-01

    Recently, the U.S. Department of Energy (DOE) issued new quality assurance (QA) orders to reflect current policies for conduct and operation of DOE-authorized programs and facilities. Establishing traceability to new QA criteria and requirements from former multidraft orders, QA manuals, and guidance documentation for DOE-funded work can be confusing. Identified critical considerations still must be addressed. Most of the newly stated QA criteria can be cross referenced, where applicable, to former QA plans and manuals. Where additional criteria occur, new procedures may be required, together with revisions in QA plans and manuals.

  9. Sensitivity in error detection of patient specific QA tools for IMRT plans

    NASA Astrophysics Data System (ADS)

    Lat, S. Z.; Suriyapee, S.; Sanghangthum, T.

    2016-03-01

    The high complexity of dose calculation in treatment planning and accurate delivery of IMRT plan need high precision of verification method. The purpose of this study is to investigate error detection capability of patient specific QA tools for IMRT plans. The two H&N and two prostate IMRT plans with MapCHECK2 and portal dosimetry QA tools were studied. Measurements were undertaken for original and modified plans with errors introduced. The intentional errors composed of prescribed dose (±2 to ±6%) and position shifting in X-axis and Y-axis (±1 to ±5mm). After measurement, gamma pass between original and modified plans were compared. The average gamma pass for original H&N and prostate plans were 98.3% and 100% for MapCHECK2 and 95.9% and 99.8% for portal dosimetry, respectively. In H&N plan, MapCHECK2 can detect position shift errors starting from 3mm while portal dosimetry can detect errors started from 2mm. Both devices showed similar sensitivity in detection of position shift error in prostate plan. For H&N plan, MapCHECK2 can detect dose errors starting at ±4%, whereas portal dosimetry can detect from ±2%. For prostate plan, both devices can identify dose errors starting from ±4%. Sensitivity of error detection depends on type of errors and plan complexity.

  10. USGS QA Plan: Certification of digital airborne mapping products

    USGS Publications Warehouse

    Christopherson, J.

    2007-01-01

    To facilitate acceptance of new digital technologies in aerial imaging and mapping, the US Geological Survey (USGS) and its partners have launched a Quality Assurance (QA) Plan for Digital Aerial Imagery. This should provide a foundation for the quality of digital aerial imagery and products. It introduces broader considerations regarding processes employed by aerial flyers in collecting, processing and delivering data, and provides training and information for US producers and users alike.

  11. Statistical process control analysis for patient-specific IMRT and VMAT QA.

    PubMed

    Sanghangthum, Taweap; Suriyapee, Sivalee; Srisatit, Somyot; Pawlicki, Todd

    2013-05-01

    This work applied statistical process control to establish the control limits of the % gamma pass of patient-specific intensity modulated radiotherapy (IMRT) and volumetric modulated arc therapy (VMAT) quality assurance (QA), and to evaluate the efficiency of the QA process by using the process capability index (Cpml). A total of 278 IMRT QA plans in nasopharyngeal carcinoma were measured with MapCHECK, while 159 VMAT QA plans were undertaken with ArcCHECK. Six megavolts with nine fields were used for the IMRT plan and 2.5 arcs were used to generate the VMAT plans. The gamma (3%/3 mm) criteria were used to evaluate the QA plans. The % gamma passes were plotted on a control chart. The first 50 data points were employed to calculate the control limits. The Cpml was calculated to evaluate the capability of the IMRT/VMAT QA process. The results showed higher systematic errors in IMRT QA than VMAT QA due to the more complicated setup used in IMRT QA. The variation of random errors was also larger in IMRT QA than VMAT QA because the VMAT plan has more continuity of dose distribution. The average % gamma pass was 93.7% ± 3.7% for IMRT and 96.7% ± 2.2% for VMAT. The Cpml value of IMRT QA was 1.60 and VMAT QA was 1.99, which implied that the VMAT QA process was more accurate than the IMRT QA process. Our lower control limit for % gamma pass of IMRT is 85.0%, while the limit for VMAT is 90%. Both the IMRT and VMAT QA processes are good quality because Cpml values are higher than 1.0.

  12. Prospective clinical validation of independent DVH prediction for plan QA in automatic treatment planning for prostate cancer patients.

    PubMed

    Wang, Yibing; Heijmen, Ben J M; Petit, Steven F

    2017-12-01

    To prospectively investigate the use of an independent DVH prediction tool to detect outliers in the quality of fully automatically generated treatment plans for prostate cancer patients. A plan QA tool was developed to predict rectum, anus and bladder DVHs, based on overlap volume histograms and principal component analysis (PCA). The tool was trained with 22 automatically generated, clinical plans, and independently validated with 21 plans. Its use was prospectively investigated for 50 new plans by replanning in case of detected outliers. For rectum D mean , V 65Gy , V 75Gy , anus D mean , and bladder D mean , the difference between predicted and achieved was within 0.4 Gy or 0.3% (SD within 1.8 Gy or 1.3%). Thirteen detected outliers were re-planned, leading to moderate but statistically significant improvements (mean, max): rectum D mean (1.3 Gy, 3.4 Gy), V 65Gy (2.7%, 4.2%), anus D mean (1.6 Gy, 6.9 Gy), and bladder D mean (1.5 Gy, 5.1 Gy). The rectum V 75Gy of the new plans slightly increased (0.2%, p = 0.087). A high accuracy DVH prediction tool was developed and used for independent QA of automatically generated plans. In 28% of plans, minor dosimetric deviations were observed that could be improved by plan adjustments. Larger gains are expected for manually generated plans. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Test/QA plan for the validation of the verification protocol for high speed pesticide spray drift reduction technologies for row and field crops

    EPA Science Inventory

    This test/QA plan for evaluation the generic test protocol for high speed wind tunnel, representing aerial application, pesticide spray drift reduction technologies (DRT) for row and field crops is in conformance with EPA Requirements for Quality Assurance Project Plans (EPA QA/R...

  14. Test/QA plan for the validation of the verification protocol for low speed pesticide spray drift reduction technologies for row and field crops

    EPA Science Inventory

    This test/QA plan for evaluation the generic test protocol for high speed wind tunnel, representing aerial application, pesticide spray drift reduction technologies (DRT) for row and field crops is in conformance with EPA Requirements for Quality Assurance Project Plans (EPA QA/R...

  15. SU-F-T-308: Mobius FX Evaluation and Comparison Against a Commercial 4D Detector Array for VMAT Plan QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Vazquez Quino, L; Huerta Hernandez, C; Morrow, A

    2016-06-15

    Purpose: To evaluate the use of MobiusFX as a pre-treatment verification IMRT QA tool and compare it with a commercial 4D detector array for VMAT plan QA. Methods: 15 VMAT plan QA of different treatment sites were delivered and measured by traditional means with the 4D detector array ArcCheck (Sun Nuclear corporation) and at the same time measurement in linac treatment logs (Varian Dynalogs files) were analyzed from the same delivery with MobiusFX software (Mobius Medical Systems). VMAT plan QAs created in Eclipse treatment planning system (Varian) in a TrueBeam linac machine (Varian) were delivered and analyzed with the gammamore » analysis routine from SNPA software (Sun Nuclear corporation). Results: Comparable results in terms of the gamma analysis with 99.06% average gamma passing with 3%,3mm passing rate is observed in the comparison among MobiusFX, ArcCheck measurements, and the Treatment Planning System dose calculated. When going to a stricter criterion (1%,1mm) larger discrepancies are observed in different regions of the measurements with an average gamma of 66.24% between MobiusFX and ArcCheck. Conclusion: This work indicates the potential for using MobiusFX as a routine pre-treatment patient specific IMRT method for quality assurance purposes and its advantages as a phantom-less method which reduce the time for IMRT QA measurement. MobiusFX is capable of produce similar results of those by traditional methods used for patient specific pre-treatment verification VMAT QA. Even the gamma results comparing to the TPS are similar the analysis of both methods show that the errors being identified by each method are found in different regions. Traditional methods like ArcCheck are sensitive to setup errors and dose difference errors coming from the linac output. On the other hand linac log files analysis record different errors in the VMAT QA associated with the MLCs and gantry motion that by traditional methods cannot be detected.« less

  16. Poster - Thur Eve - 29: Detecting changes in IMRT QA using statistical process control.

    PubMed

    Drever, L; Salomons, G

    2012-07-01

    Statistical process control (SPC) methods were used to analyze 239 measurement based individual IMRT QA events. The selected IMRT QA events were all head and neck (H&N) cases with 70Gy in 35 fractions, and all prostate cases with 76Gy in 38 fractions planned between March 2009 and 2012. The results were used to determine if the tolerance limits currently being used for IMRT QA were able to indicate if the process was under control. The SPC calculations were repeated for IMRT QA of the same type of cases that were planned after the treatment planning system was upgraded from Eclipse version 8.1.18 to version 10.0.39. The initial tolerance limits were found to be acceptable for two of the three metrics tested prior to the upgrade. After the upgrade to the treatment planning system the SPC analysis found that the a priori limits were no longer capable of indicating control for 2 of the 3 metrics analyzed. The changes in the IMRT QA results were clearly identified using SPC, indicating that it is a useful tool for finding changes in the IMRT QA process. Routine application of SPC to IMRT QA results would help to distinguish unintentional trends and changes from the random variation in the IMRT QA results for individual plans. © 2012 American Association of Physicists in Medicine.

  17. SU-E-T-164: Clinical Implementation of ASi EPID Panels for QA of IMRT/VMAT Plans.

    PubMed

    Hosier, K; Wu, C; Beck, K; Radevic, M; Asche, D; Bareng, J; Kroner, A; Lehmann, J; Logsdon, M; Dutton, S; Rosenthal, S

    2012-06-01

    To investigate various issues for clinical implementation of aSi EPID panels for IMRT/VMAT QA. Six linacs are used in our clinic for EPID-based plan QA; two Varian Truebeams, two Varian 2100 series, two Elekta Infiniti series. Multiple corrections must be accounted for in the calibration of each panel for dosimetric use. Varian aSi panels are calibrated with standard dark field, flood field, and 40×40 diagonal profile for beam profile correction. Additional corrections to account for off-axis and support arm backscatter are needed for larger field sizes. Since Elekta iViewGT system does not export gantry angle with images, a third-party inclinometer must be physically mounted to back of linac gantry and synchronized with data acquisition via iViewGT PC clock. A T/2 offset correctly correlates image and gantry angle for arc plans due to iView image time stamp at the end of data acquisition for each image. For both Varian and Elekta panels, a 5 MU 10×10 calibration field is used to account for the nonlinear MU to dose response at higher energies. Acquired EPID images are deconvolved via a high pass filter in Fourier space and resultant fluence maps are used to reconstruct a 3D dose 'delivered' to patient using DosimetryCheck. Results are compared to patient 3D dose computed by TPS using a 3D-gamma analysis. 120 IMRT and 100 VMAT cases are reported. Two 3D gamma quantities (Gamma(V10) and Gamma(PTV)) are proposed for evaluating QA results. The Gamma(PTV) is sensitive to MLC offsets while Gamma(V10) is sensitive to gantry rotations. When a 3mm/3% criteria and 90% or higher 3D gamma pass rate is used, all IMRT and 90% of VMAT QA pass QA. After appropriate calibration of aSi panels and setup of image acquisition systems, EPID based 3D dose reconstruction method is found clinically feasible. © 2012 American Association of Physicists in Medicine.

  18. MO-G-BRE-02: A Survey of IMRT QA Practices for More Than 800 Institutions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pulliam, K; Kerns, J; Howell, R

    Purpose: A wide range of techniques and measurement devices are employed for IMRT QA, causing a large variation of accepted action limits and potential follow up for failing plans. Such procedures are not well established or accepted in the medical physics community. To achieve the goal of proving insight into current IMRT QA practices, we created an electronic IMRT QA survey. The survey was open to a variety of the most common QA devices and assessed the type of comparison to measurement, action limits, delivery methods, and clinical action for failing QA plans. Methods: We conducted an online survey throughmore » the Radiological Physics Center's (RPC) annual survey with the goal of ascertaining elements of routine patient-specific IMRT QA. A total of 874 institutions responded to the survey. The questions ranged from asking for action limits, dosimeter type(s) used, delivery techniques, and actions taken when a plan fails IMRT QA. Results: The most common (52%) planar gamma criteria was 3%/3 mm with a 95% of pixels passing criteria. The most common QA device were diode arrays (48%). The most common first response to a plan failing QA was to re-measure at the same point the point dose (89%), second was to re-measure at a new point (13%), and third was to analyze the plan in relative instead of absolute mode (10%) (Does not add to 100% as not all institutions placed a response for each QA follow-up option). Some institutions, however, claimed that they had never observed a plan failure. Conclusion: The survey provided insights into the way the community currently performs IMRT QA. This information will help in the push to standardize action limits among dosimeters.« less

  19. SU-F-T-643: Feasibility of Performing Patient Specific VMAT QA On Single Linac for Plans Treated in Beam-Matched Elekta Agility Linacs

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Leung, R; Lee, V; Cheung, S

    2016-06-15

    Purpose: The increasing application of VMAT demands a more efficient workflow and QA solution. This study aims to investigate the feasibility of performing VMAT QA measurements on one linac for plans treated on other beam-matched Elekta Agility linacs. Methods: A single model was used to create 24 clinically approved VMAT plans (12 head-and-neck and 12 prostate using 6MV and 10MV respectively) on Pinnacle v9.10 (Philips, Einhoven, Netherlands). All head-and-neck plans were delivered on three beam-matched machines while all prostate cases were delivered on two beam-matched 10MV Agility machines. All plans were delivered onto PTW Octavius 4D phantom with 1500 detectormore » array (PTW, Freiburg, Germany). Reconstructed volume doses were then compared with the Pinnacle reference plans in Verisoft 6.1 under 3%/3mm gamma criteria at local dose. Plans were considered clinically acceptable if >90% of the voxels passing the gamma criteria. Results: All measurements were passed (3D gamma passing rate >90%) and the result shows that the mean difference of 3D gamma of 12 head-and-neck cases is 1.2% with standard deviation of 0.6%. While for prostate cases, the mean difference of 3D gamma is 0.9% with standard deviation of 0.7%. Maximum difference of 3D gamma of all measurements between beam-matched machines is less than 2.5%. The differences of passing rates between different machines were statistically insignificant (p>0.05). Conclusion. The result suggests that ther Conclusion: The result suggests that there exists a 3D gamma threshold, in our case 92.5%, above which the VMAT QA performed in any one of beam-matched machine will also pass in another one. Therefore, VMAT QA efficiency may be increased and phantom set up time can be saved by implementing such method. A constant performance across all beam matched machines must be maintained to make this QA approach feasible.« less

  20. MO-A-16A-01: QA Procedures and Metrics: In Search of QA Usability

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sathiaseelan, V; Thomadsen, B

    Radiation therapy has undergone considerable changes in the past two decades with a surge of new technology and treatment delivery methods. The complexity of radiation therapy treatments has increased and there has been increased awareness and publicity about the associated risks. In response, there has been proliferation of guidelines for medical physicists to adopt to ensure that treatments are delivered safely. Task Group recommendations are copious, and clinical physicists' hours are longer, stretched to various degrees between site planning and management, IT support, physics QA, and treatment planning responsibilities.Radiation oncology has many quality control practices in place to ensure themore » delivery of high-quality, safe treatments. Incident reporting systems have been developed to collect statistics about near miss events at many radiation oncology centers. However, tools are lacking to assess the impact of these various control measures. A recent effort to address this shortcoming is the work of Ford et al (2012) who recently published a methodology enumerating quality control quantification for measuring the effectiveness of safety barriers. Over 4000 near-miss incidents reported from 2 academic radiation oncology clinics were analyzed using quality control quantification, and a profile of the most effective quality control measures (metrics) was identified.There is a critical need to identify a QA metric to help the busy clinical physicists to focus their limited time and resources most effectively in order to minimize or eliminate errors in the radiation treatment delivery processes. In this symposium the usefulness of workflows and QA metrics to assure safe and high quality patient care will be explored.Two presentations will be given:Quality Metrics and Risk Management with High Risk Radiation Oncology ProceduresStrategies and metrics for quality management in the TG-100 Era Learning Objectives: Provide an overview and the need for QA

  1. SU-F-T-287: A Preliminary Study On Patient Specific VMAT Verification Using a Phosphor-Screen Based Geometric QA System (Raven QA)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lee, M; Yi, B; Wong, J

    Purpose: The RavenQA system (LAP Laser, Germany) is a QA device with a phosphor screen detector for performing the QA tasks of TG-142. This study tested if it is feasible to use the system for the patient specific QA of the Volumetric Modulated Arc Therapy (VMAT). Methods: Water equivalent material (5cm) is attached to the front of the detector plate of the RavenQA for dosimetry purpose. Then the plate is attached to the gantry to synchronize the movement between the detector and the gantry. Since the detector moves together with gantry, The ’Reset gantry to 0’ function of the Eclipsemore » planning system (Varian, CA) is used to simulate the measurement situation when calculating dose of the detector plate. The same gantry setup is used when delivering the treatment beam for feasibility test purposes. Cumulative dose is acquired for each arc. The optical scatter component of each captured image from the CCD camera is corrected by deconvolving the 2D spatial invariant optical scatter kernel (OSK). We assume that the OSK is a 2D isotropic point spread function with inverse-squared decrease as a function of radius from the center. Results: Three cases of VMAT plans including head & neck, whole pelvis and abdomen-pelvis are tested. Setup time for measurements was less than 5 minutes. Passing rates of absolute gamma were 99.3, 98.2, 95.9 respectively for 3%/3mm criteria and 96.2, 97.1, 86.4 for 2%/2mm criteria. The abdomen-pelvis field has long treatment fields, 37cm, which are longer than the detector plate (25cm). This plan showed relatively lower passing rate than other plans. Conclusion: An algorithm for IMRT/VMAT verification using the RavenQA has been developed and tested. The model of spatially invariant OSK works well for deconvolution purpose. It is proved that the RavenQA can be used for the patient specific verification of VMAT. This work is funded in part by a Maryland Industrial Partnership Program grant to University of Maryland and to JPLC who owns

  2. SU-E-J-26: Analysis of Image Quality in CBCT QA Using a Treatment Planning System

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Xu, M

    2014-06-01

    Purpose: The objective of this study is to propose an alternative QA technique that analyzes imaging quality(IQ) in CBCT-QA processing. Methods: A catphan phantom was used to take CT imaging data set that were imported into a treatment planning system - Eclipse. The image quality was analyzed in terms of in-slice geometry resolution, Hounsfield numbers(HU) accuracy, mean-slice thickness, edge-to-center uniformity, low contrast resolution, and high contrast spatial resolution in Eclipse workstation. The CBCT-QA was also analyzed by OBI-workstation and a commercial software. Comparison was made to evaluation feasibility in a TPS environment. Results: The analysis of IQ was conducted inmore » Eclipse v10.0 TPS. In-slice geometric resolution was measured between 2-rods in section CTP404 and repeated for all 4 rods with the difference between expected and measured values less than +/−0.1 cm. For HU, the difference between expected and measured values in HU was found much less than +/−40. Mean slice thickness measured by a distance on the wire proportional to scanner increment multiplying by a factor of 0.42. After repeating measurements to 4 wires, the average difference between expected and measured values was less +/−0.124 mm in slice thickness. HU uniformity was measured in section CTP486 with the tolerance less than +/−40 HU. Low contrast resolution in section CTP515 and high contrast resolution in section CTP528 were found to be 7 disks in diameter of 4 mm and 6 lp/cm, respectively. Eclipse TPS results indicated a good agreement to those obtained in OBI workstation and ImagePro software for major parameters. Conclusion: An analysis of IQ was proposed as an alternative CBCT QA processing. Based upon measured data assessment, proposed method was accurate and consistent to IQ evaluation and TG142 guideline. The approach was to utilize TPS resource, which can be valuable to re-planning, verification, and delivery in adaptive therapy.« less

  3. SU-E-T-77: Comparison of 2D and 3D Gamma Analysis in Patient-Specific QA for Prostate VMAT Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Clemente, F; Perez, C

    2014-06-01

    Purpose: Patient-specific QA procedures for IMRT and VMAT are traditionally performed by comparing TPS calculations with measured single point values and plane dose distributions by means of gamma analysis. New QA devices permit us to calculate 3D dose distributions on patient anatomy as redundant secondary check and reconstruct it from measurements taken with 2D and 3D detector arrays. 3D dose calculations allow us to perform DVH-based comparisons with clinical relevance, as well as 3D gamma analysis. One of these systems (Compass, IBA Dosimetry) combines traditional 2D with new anatomical-based 3D gamma analysis. This work shows the ability of this systemmore » by comparing 2D and 3D gamma analysis in pre-treatment QA for several VMAT prostate plans. Methods: Compass is capable of calculating dose as secondary check from DICOM TPS data and reconstructing it from measurements taken by a 2D ion chamber array (MatriXX Evolution, IBA Dosimetry). Both 2D and 3D gamma tests are available to compare calculated and reconstructed dose in Compass with TPS RT Dose. Results: 15 VMAT prostate plans have been measured with Compass. Dose is reconstructed with Compass for these plans. 2D gamma comparisons can be done for any plane from dose matrix. Mean gamma passing rates for isocenter planes (axial, coronal, sagittal) are (99.7±0.2)%, (99.9±0.1)%, (99.9±0.1)% for reconstructed dose planes. 3D mean gamma passing rates are (98.5±1.7)% for PTVs, (99.1±1.5)% for rectum, (100.0±0.0)% for bladder, (99.6±0.7)% for femoral heads and (98.1±4.1)% for penile bulb. Conclusion: Compass is a powerful tool to perform a complete pre-treatment QA analysis, from 2D techniques to 3D DVH-based techniques with clinical relevance. All reported values for VMAT prostate plans are in good agreement with TPS values. This system permits us to ensure the accuracy in the delivery of VMAT treatments completing a full patient-specific QA program.« less

  4. WE-G-BRA-02: SafetyNet: Automating Radiotherapy QA with An Event Driven Framework

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hadley, S; Kessler, M; Litzenberg, D

    2015-06-15

    Purpose: Quality assurance is an essential task in radiotherapy that often requires many manual tasks. We investigate the use of an event driven framework in conjunction with software agents to automate QA and eliminate wait times. Methods: An in house developed subscription-publication service, EventNet, was added to the Aria OIS to be a message broker for critical events occurring in the OIS and software agents. Software agents operate without user intervention and perform critical QA steps. The results of the QA are documented and the resulting event is generated and passed back to EventNet. Users can subscribe to those eventsmore » and receive messages based on custom filters designed to send passing or failing results to physicists or dosimetrists. Agents were developed to expedite the following QA tasks: Plan Revision, Plan 2nd Check, SRS Winston-Lutz isocenter, Treatment History Audit, Treatment Machine Configuration. Results: Plan approval in the Aria OIS was used as the event trigger for plan revision QA and Plan 2nd check agents. The agents pulled the plan data, executed the prescribed QA, stored the results and updated EventNet for publication. The Winston Lutz agent reduced QA time from 20 minutes to 4 minutes and provided a more accurate quantitative estimate of radiation isocenter. The Treatment Machine Configuration agent automatically reports any changes to the Treatment machine or HDR unit configuration. The agents are reliable, act immediately, and execute each task identically every time. Conclusion: An event driven framework has inverted the data chase in our radiotherapy QA process. Rather than have dosimetrists and physicists push data to QA software and pull results back into the OIS, the software agents perform these steps immediately upon receiving the sentinel events from EventNet. Mr Keranen is an employee of Varian Medical Systems. Dr. Moran’s institution receives research support for her effort for a linear accelerator QA project

  5. SU-F-T-569: Implementation of a Patient Specific QA Method Using EBT-XD for CyberKnife SRS/SBRT Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zerouali, K; Aubry, J; Doucet, R

    2016-06-15

    Purpose: To implement the new EBT-XD Gafchromic films for accurate dosimetric and geometric validation of stereotactic radiosurgery (SRS) and stereotactic body radiation therapy (SBRT) CyberKnife (CK) patient specific QA. Methods: Film calibration was performed using a triplechannel film analysis on an Epson 10000XL scanner. Calibration films were irradiated using a Varian Clinac 21EX flattened beam (0 to 20 Gy), to ensure sufficient dose homogeneity. Films were scanned to a resolution of 0.3 mm, 24 hours post irradiation following a well-defined protocol. A set of 12 QA was performed for several types of CK plans: trigeminal neuralgia, brain metastasis, prostate andmore » lung tumors. A custom made insert for the CK head phantom has been manufactured to yield an accurate measured to calculated dose registration. When the high dose region was large enough, absolute dose was also measured with an ionization chamber. Dose calculation is performed using MultiPlan Ray-tracing algorithm for all cases since the phantom is mostly made from near water-equivalent plastic. Results: Good agreement (<2%) was found between the dose to the chamber and the film, when a chamber measurement was possible The average dose difference and standard deviations between film measurements and TPS calculations were respectively 1.75% and 3%. The geometric accuracy has been estimated to be <1 mm, combining robot positioning uncertainty and film registration to calculated dose. Conclusion: Patient specific QA measurements using EBT-XD films yielded a full 2D dose plane with high spatial resolution and acceptable dose accuracy. This method is particularly promising for trigeminal neuralgia plan QA, where the positioning of the spatial dose distribution is equally or more important than the absolute delivered dose to achieve clinical goals.« less

  6. SU-E-T-60: A Plan Quality Index in IMRT QA That Is Independent of the Acceptance Criteria

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kim, D; Kang, S; Kim, T

    2015-06-15

    Purpose: In IMRT QA, plan quality evaluation is made based on pass rate under preset acceptance criteria, mostly using gamma-values. This method is convenient but, its Result highly depends on what the acceptance criteria are and suffers from the lack of sensitivity in judging how good the plan is. In this study, we introduced a simple but effective plan quality index of IMRT QA based on dose difference only to supplement such shortcomings, and investigated its validity. Methods: The proposed index is a single value which is calculated mainly based on point-by-point comparison between planned and measured dose distributions, andmore » it becomes “1” in an ideal case. A systematic evaluation was performed with one-dimensional test dose distributions. For 3 hypothetical dose profiles, various displacements (in both dose and space) were introduced, the proposed index was calculated for each case, and the behavior of obtained indices was analyzed and compared with that of gamma evaluation. In addition, the feasibility of the index was assessed with clinical IMRT/VMAT/SBRT QA cases for different sites (prostate, head & neck, liver, lung, spine, and abdomen). Results: The proposed index showed more robust correlation with the amount of induced displacement compared to the gamma evaluation method. No matter what the acceptance criteria are (e.g., whether 3%/3mm or 2%/2mm), it was possible to clearly rank every case with the proposed index while it was difficult to do with the gamma evaluation method. Conclusion: IMRT plan quality can be evaluated quantitatively by the proposed index. It is considered that the proposed index would provide useful information for better judging the level of goodness of each plan and its Result is independent of the acceptance criteria. This work was supported by the Radiation Technology R&D program (No. 2013M2A2A7043498) and the Mid-career Researcher Program (2014R1A2A1A10050270) through the National Research Foundation of Korea funded

  7. SU-F-T-274: Modified Dose Calibration Methods for IMRT QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Luo, W; Westlund, S

    2016-06-15

    Purpose: To investigate IMRT QA uncertainties caused by dose calibration and modify widely used dose calibration procedures to improve IMRT QA accuracy and passing rate. Methods: IMRT QA dose measurement is calibrated using a calibration factor (CF) that is the ratio between measured value and expected value corresponding to the reference fields delivered on a phantom. Two IMRT QA phantoms were used for this study: a 30×30×30 cm3 solid water cube phantom (Cube), and the PTW Octavius phantom. CF was obtained by delivering 100 MUs to the phantoms with different reference fields ranging from 3×3 cm2 to 20×20 cm{sup 2}.more » For Cube, CFs were obtained using the following beam arrangements: 2-AP Field - chamber at dmax, 2-AP Field - chamber at isocenter, 4-beam box - chamber at isocenter, and 8 equally spaced fields and chamber at isocenter. The same plans were delivered on Octavius and CFs were derived for the dose at the isocenter using the above beam arrangements. The Octavius plans were evaluated with PTW-VeriSoft (Gamma criteria of 3%/3mm). Results: Four head and neck IMRT plans were included in this study. For point dose measurement with Cube, the CFs with 4-Field gave the best agreement between measurement and calculation within 4% for large field plans. All the measurement results agreed within 2% for a small field plan. Compared with calibration field sizes, 5×5 to 15×15 were more accurate than other field sizes. For Octavius, 4-Field calibration increased passing rate by up to 10% compared to AP calibration. Passing rate also increased by up to 4% with the increase of field size from 3×3 to 20×20. Conclusion: IMRT QA results are correlated with calibration methods used. The dose calibration using 4-beam box with field sizes from 5×5 to 20×20 can improve IMRT QA accuracy and passing rate.« less

  8. On the use of biomathematical models in patient-specific IMRT dose QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhen Heming; Nelms, Benjamin E.; Tome, Wolfgang A.

    2013-07-15

    Purpose: To investigate the use of biomathematical models such as tumor control probability (TCP) and normal tissue complication probability (NTCP) as new quality assurance (QA) metrics.Methods: Five different types of error (MLC transmission, MLC penumbra, MLC tongue and groove, machine output, and MLC position) were intentionally induced to 40 clinical intensity modulated radiation therapy (IMRT) patient plans (20 H and N cases and 20 prostate cases) to simulate both treatment planning system errors and machine delivery errors in the IMRT QA process. The changes in TCP and NTCP for eight different anatomic structures (H and N: CTV, GTV, both parotids,more » spinal cord, larynx; prostate: CTV, rectal wall) were calculated as the new QA metrics to quantify the clinical impact on patients. The correlation between the change in TCP/NTCP and the change in selected DVH values was also evaluated. The relation between TCP/NTCP change and the characteristics of the TCP/NTCP curves is discussed.Results:{Delta}TCP and {Delta}NTCP were summarized for each type of induced error and each structure. The changes/degradations in TCP and NTCP caused by the errors vary widely depending on dose patterns unique to each plan, and are good indicators of each plan's 'robustness' to that type of error.Conclusions: In this in silico QA study the authors have demonstrated the possibility of using biomathematical models not only as patient-specific QA metrics but also as objective indicators that quantify, pretreatment, a plan's robustness with respect to possible error types.« less

  9. Sci-Fri PM: Radiation Therapy, Planning, Imaging, and Special Techniques - 06: Patient-specific QA Procedure for Gated VMAT SABR Treatments using 10x Beam in Flattening-Filter Free Mode

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mestrovic, Ante; Chitsazzadeh, Shadi; Wells, Derek

    2016-08-15

    Purpose: To develop a highly sensitive patient specific QA procedure for gated VMAT stereotactic ablative radiotherapy (SABR) treatments. Methods: A platform was constructed to attach the translational stage of a Quasar respiratory motion phantom to a pinpoint ion chamber insert and move the ion chamber inside the ArcCheck. The Quasar phantom controller uses a patient-specific breathing pattern to translate the ion chamber in a superior-inferior direction inside the ArcCheck. With this system the ion chamber is used to QA the correct phase of the gated delivery and the ArcCheck diodes are used to QA the overall dose distribution. This novelmore » approach requires a single plan delivery for a complete QA of a gated plan. The sensitivity of the gating QA procedure was investigated with respect to the following parameters: PTV size, exhale duration, baseline drift, gating window size. Results: The difference between the measured dose to a point in the penumbra and the Eclipse calculated dose was under 2% for small residual motions. The QA procedure was independent of PTV size and duration of exhale. Baseline drift and gating window size, however, significantly affected the penumbral dose measurement, with differences of up to 30% compared to Eclipse. Conclusion: This study described a highly sensitive QA procedure for gated VMAT SABR treatments. The QA outcome was dependent on the gating window size and baseline drift. Analysis of additional patient breathing patterns is currently undergoing to determine a clinically relevant gating window size and an appropriate tolerance level for this procedure.« less

  10. SU-F-T-285: Evaluation of a Patient DVH-Based IMRT QA System

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhen, H; Redler, G; Chu, J

    2016-06-15

    Purpose: To evaluate the clinical performance of a patient DVH-based QA system for prostate VMAT QA. Methods: Mobius3D(M3D) is a QA software with an independent beam model and dose engine. The MobiusFX(MFX) add-on predicts patient dose using treatment machine log files. We commissioned the Mobius beam model in two steps. First, the stock beam model was customized using machine commissioning data, then verified against the TPS with 12 simple phantom plans and 7 clinical 3D plans. Secondly, the Dosimetric Leaf Gap(DLG) in the Mobius model was fine-tuned for VMAT treatment based on ion chamber measurements for 6 clinical VMAT plans.more » Upon successful commissioning, we retrospectively performed IMRT QA for 12 VMAT plans with the Mobius system as well as the ArcCHECK-3DVH system. Selected patient DVH values (PTV D95, D50; Bladder D2cc, Dmean; Rectum D2cc) were compared between TPS, M3D, MFX, and 3DVH. Results: During the first commissioning step, TPS and M3D calculated target Dmean for 3D plans agree within 0.7%±0.7%, with 3D gamma passing rates of 98%±2%. In the second commissioning step, the Mobius DLG was adjusted by 1.2mm from the stock value, reducing the average difference between MFX calculation and ion chamber measurement from 3.2% to 0.1%. In retrospective prostate VMAT QA, 5 of 60 MFX calculated DVH values have a deviation greater than 5% compared to TPS. One large deviation at high dose level was identified as a potential QA failure. This echoes the 3DVH QA result, which identified 2 instances of large DVH deviation on the same structure. For all DVH’s evaluated, M3D and MFX show high level of agreement (0.1%±0.2%), indicating that the observed deviation is likely from beam modelling differences rather than delivery errors. Conclusion: Mobius system provides a viable solution for DVH based VMAT QA, with the capability of separating TPS and delivery errors.« less

  11. MO-D-213-05: Sensitivity of Routine IMRT QA Metrics to Couch and Collimator Rotations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Alaei, P

    Purpose: To assess the sensitivity of gamma index and other IMRT QA metrics to couch and collimator rotations. Methods: Two brain IMRT plans with couch and/or collimator rotations in one or more of the fields were evaluated using the IBA MatriXX ion chamber array and its associated software (OmniPro-I’mRT). The plans were subjected to routine QA by 1) Creating a composite planar dose in the treatment planning system (TPS) with the couch/collimator rotations and 2) Creating the planar dose after “zeroing” the rotations. Plan deliveries to MatriXX were performed with all rotations set to zero on a Varian 21ex linearmore » accelerator. This in effect created TPS-created planar doses with an induced rotation error. Point dose measurements for the delivered plans were also performed in a solid water phantom. Results: The IMRT QA of the plans with couch and collimator rotations showed clear discrepancies in the planar dose and 2D dose profile overlays. The gamma analysis, however, did pass with the criteria of 3%/3mm (for 95% of the points), albeit with a lower percentage pass rate, when one or two of the fields had a rotation. Similar results were obtained with tighter criteria of 2%/2mm. Other QA metrics such as percentage difference or distance-to-agreement (DTA) histograms produced similar results. The point dose measurements did not obviously indicate the error due to location of dose measurement (on the central axis) and the size of the ion chamber used (0.6 cc). Conclusion: Relying on Gamma analysis, percentage difference, or DTA to determine the passing of an IMRT QA may miss critical errors in the plan delivery due to couch/collimator rotations. A combination of analyses for composite QA plans, or per-beam analysis, would detect these errors.« less

  12. SU-F-T-586: Pre-Treatment QA of InCise2 MLC Plans On a Cyberknife-M6 Using the Delta4 System in SBRT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Schmidhalter, D; Henzen, D; Malthaner, M

    Purpose: Performing pre-treatment quality assurance (QA) with the Delta4 system (ScandiDos Inc., Madison, WI) is well established for linac-based radiotherapy. This is not true when using a Cyberknife (Accuray Inc., Sunnyvale, CA) where, typically film-based QA is applied. The goal of this work was to test the feasibility to use the Delta4 system for pre-treatment QA for stereotactic body radiation therapy (SBRT) using a Cyberknife-M6 equipped with the InCise2 multileaf collimator (MLC). Methods: In order to perform measurements without accelerator pulse signal, the Tomotherapy option within the Delta4 software was used. Absolute calibration of the Delta4 phantom was performed usingmore » a 10×10 cm{sup 2} field shaped by the InCise2 MLC of the Cyberknife-M6. Five fiducials were attached to the Delta4 phantom in order to be able to track the phantom before and during measurements. For eight SBRT treatment plans (two liver, two prostate, one lung, three bone metastases) additional verification plans were recalculated on the Delta4 phantom using MultiPlan. Dicom data was exported from MultiPlan and was adapted in order to be compatible with the Delta4 software. The measured and calculated dose distributions were compared using the gamma analysis of the Delta4 system. Results: All eight SBRT plans were successfully measured with the aid of the Delta4 system. In the mean, 98.0±1.9%, 95.8±4.1% and 88.40±11.4% of measured dose points passed the gamma analysis using a global dose deviation criterion of 3% (100% corresponds to the dose maximum) and a distance-to-agreement criterion of 3 mm, 2 mm and 1 mm, respectively, and a threshold of 20%. Conclusion: Pre-treatment QA of SBRT plans using the Delta4 system on a Cyberknife-M6 is feasible. Measured dose distributions of SBRT plans showed clinically acceptable agreement with the corresponding calculated dose distributions.« less

  13. Toward optimizing patient-specific IMRT QA techniques in the accurate detection of dosimetrically acceptable and unacceptable patient plans

    PubMed Central

    McKenzie, Elizabeth M.; Balter, Peter A.; Stingo, Francesco C.; Jones, Jimmy; Followill, David S.; Kry, Stephen F.

    2014-01-01

    Purpose: The authors investigated the performance of several patient-specific intensity-modulated radiation therapy (IMRT) quality assurance (QA) dosimeters in terms of their ability to correctly identify dosimetrically acceptable and unacceptable IMRT patient plans, as determined by an in-house-designed multiple ion chamber phantom used as the gold standard. A further goal was to examine optimal threshold criteria that were consistent and based on the same criteria among the various dosimeters. Methods: The authors used receiver operating characteristic (ROC) curves to determine the sensitivity and specificity of (1) a 2D diode array undergoing anterior irradiation with field-by-field evaluation, (2) a 2D diode array undergoing anterior irradiation with composite evaluation, (3) a 2D diode array using planned irradiation angles with composite evaluation, (4) a helical diode array, (5) radiographic film, and (6) an ion chamber. This was done with a variety of evaluation criteria for a set of 15 dosimetrically unacceptable and 9 acceptable clinical IMRT patient plans, where acceptability was defined on the basis of multiple ion chamber measurements using independent ion chambers and a phantom. The area under the curve (AUC) on the ROC curves was used to compare dosimeter performance across all thresholds. Optimal threshold values were obtained from the ROC curves while incorporating considerations for cost and prevalence of unacceptable plans. Results: Using common clinical acceptance thresholds, most devices performed very poorly in terms of identifying unacceptable plans. Grouping the detector performance based on AUC showed two significantly different groups. The ion chamber, radiographic film, helical diode array, and anterior-delivered composite 2D diode array were in the better-performing group, whereas the anterior-delivered field-by-field and planned gantry angle delivery using the 2D diode array performed less well. Additionally, based on the AUCs, there

  14. Toward optimizing patient-specific IMRT QA techniques in the accurate detection of dosimetrically acceptable and unacceptable patient plans.

    PubMed

    McKenzie, Elizabeth M; Balter, Peter A; Stingo, Francesco C; Jones, Jimmy; Followill, David S; Kry, Stephen F

    2014-12-01

    The authors investigated the performance of several patient-specific intensity-modulated radiation therapy (IMRT) quality assurance (QA) dosimeters in terms of their ability to correctly identify dosimetrically acceptable and unacceptable IMRT patient plans, as determined by an in-house-designed multiple ion chamber phantom used as the gold standard. A further goal was to examine optimal threshold criteria that were consistent and based on the same criteria among the various dosimeters. The authors used receiver operating characteristic (ROC) curves to determine the sensitivity and specificity of (1) a 2D diode array undergoing anterior irradiation with field-by-field evaluation, (2) a 2D diode array undergoing anterior irradiation with composite evaluation, (3) a 2D diode array using planned irradiation angles with composite evaluation, (4) a helical diode array, (5) radiographic film, and (6) an ion chamber. This was done with a variety of evaluation criteria for a set of 15 dosimetrically unacceptable and 9 acceptable clinical IMRT patient plans, where acceptability was defined on the basis of multiple ion chamber measurements using independent ion chambers and a phantom. The area under the curve (AUC) on the ROC curves was used to compare dosimeter performance across all thresholds. Optimal threshold values were obtained from the ROC curves while incorporating considerations for cost and prevalence of unacceptable plans. Using common clinical acceptance thresholds, most devices performed very poorly in terms of identifying unacceptable plans. Grouping the detector performance based on AUC showed two significantly different groups. The ion chamber, radiographic film, helical diode array, and anterior-delivered composite 2D diode array were in the better-performing group, whereas the anterior-delivered field-by-field and planned gantry angle delivery using the 2D diode array performed less well. Additionally, based on the AUCs, there was no significant difference

  15. GIS FOR QA PROFESSIONALS

    EPA Science Inventory

    GIS scientists and QA Professionals have combined their efforts to create this one day course that provides the QA community with a basic understanding of Geographic Information Systems (GIS). The course emphasizes the QA Aspects of GIS so that the QA Professional is better prep...

  16. QA4, a language for artificial intelligence.

    NASA Technical Reports Server (NTRS)

    Derksen, J. A. C.

    1973-01-01

    Introduction of a language for problem solving and specifically robot planning, program verification, and synthesis and theorem proving. This language, called question-answerer 4 (QA4), embodies many features that have been found useful for constructing problem solvers but have to be programmed explicitly by the user of a conventional language. The most important features of QA4 are described, and examples are provided for most of the material introduced. Language features include backtracking, parallel processing, pattern matching, set manipulation, and pattern-triggered function activation. The language is most convenient for use in an interactive way and has extensive trace and edit facilities.

  17. MO-FG-202-09: Virtual IMRT QA Using Machine Learning: A Multi-Institutional Validation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Valdes, G; Scheuermann, R; Solberg, T

    Purpose: To validate a machine learning approach to Virtual IMRT QA for accurately predicting gamma passing rates using different QA devices at different institutions. Methods: A Virtual IMRT QA was constructed using a machine learning algorithm based on 416 IMRT plans, in which QA measurements were performed using diode-array detectors and a 3%local/3mm with 10% threshold. An independent set of 139 IMRT measurements from a different institution, with QA data based on portal dosimetry using the same gamma index and 10% threshold, was used to further test the algorithm. Plans were characterized by 90 different complexity metrics. A weighted poisonmore » regression with Lasso regularization was trained to predict passing rates using the complexity metrics as input. Results: In addition to predicting passing rates with 3% accuracy for all composite plans using diode-array detectors, passing rates for portal dosimetry on per-beam basis were predicted with an error <3.5% for 120 IMRT measurements. The remaining measurements (19) had large areas of low CU, where portal dosimetry has larger disagreement with the calculated dose and, as such, large errors were expected. These beams need to be further modeled to correct the under-response in low dose regions. Important features selected by Lasso to predict gamma passing rates were: complete irradiated area outline (CIAO) area, jaw position, fraction of MLC leafs with gaps smaller than 20 mm or 5mm, fraction of area receiving less than 50% of the total CU, fraction of the area receiving dose from penumbra, weighted Average Irregularity Factor, duty cycle among others. Conclusion: We have demonstrated that the Virtual IMRT QA can predict passing rates using different QA devices and across multiple institutions. Prediction of QA passing rates could have profound implications on the current IMRT process.« less

  18. Comparison of four commercial devices for RapidArc and sliding window IMRT QA

    PubMed Central

    Chandraraj, Varatharaj; Manickam, Ravikumar; Esquivel, Carlos; Supe, Sanjay S.; Papanikolaou, Nikos

    2011-01-01

    For intensity‐modulated radiation therapy, evaluation of the measured dose against the treatment planning calculated dose is essential in the context of patient‐specific quality assurance. The complexity of volumetric arc radiotherapy delivery attributed to its dynamic and synchronization nature require new methods and potentially new tools for the quality assurance of such techniques. In the present study, we evaluated and compared the dosimetric performance of EDR2 film and three other commercially available quality assurance devices: IBA I'MatriXX array, PTW Seven29 array and the Delta 4 array. The evaluation of these dosimetric systems was performed for RapidArc and IMRT deliveries using a Varian NovalisTX linear accelerator. The plans were generated using the Varian Eclipse treatment planning system. Our results showed that all four QA techniques yield equivalent results. All patient QAs passed our institutional clinical criteria of gamma index based on a 3% dose difference and 3 mm distance to agreement. In addition, the Bland‐Altman analysis was performed which showed that all the calculated gamma values of all three QA devices were within 5% from those of the film. The results showed that the four QA systems used in this patient‐specific IMRT QA analysis are equivalent. We concluded that the dosimetric systems under investigation can be used interchangeably for routine patient specific QA. PACS numbers: 87.55.Qr, 87.56.Fc

  19. Test/QA plan for the verification testing of selective catalytic reduction control technologies for highway, nonroad use heavy-duty diesel engines

    EPA Science Inventory

    This ETV test/QA plan for heavy-duty diesel engine testing at the Southwest Research Institute’s Department of Emissions Research (DER) describes how the Federal Test Procedure (FTP), as listed in 40 CFR Part 86 for highway engines and 40 CFR Part 89 for nonroad engines, will be ...

  20. Tolerance design of patient-specific range QA using the DMAIC framework in proton therapy.

    PubMed

    Rah, Jeong-Eun; Shin, Dongho; Manger, Ryan P; Kim, Tae Hyun; Oh, Do Hoon; Kim, Dae Yong; Kim, Gwe-Ya

    2018-02-01

    To implement the DMAIC (Define-Measure-Analyze-Improve-Control) can be used for customizing the patient-specific QA by designing site-specific range tolerances. The DMAIC framework (process flow diagram, cause and effect, Pareto chart, control chart, and capability analysis) were utilized to determine the steps that need focus for improving the patient-specific QA. The patient-specific range QA plans were selected according to seven treatment site groups, a total of 1437 cases. The process capability index, C pm was used to guide the tolerance design of patient site-specific range. For prostate field, our results suggested that the patient range measurements were capable at the current tolerance level of ±1 mm in clinical proton plans. For other site-specific ranges, we analyzed that the tolerance tends to be overdesigned to insufficient process capability calculated by the patient-specific QA data. The customized tolerances were calculated for treatment sites. Control charts were constructed to simulate the patient QA time before and after the new tolerances were implemented. It is found that the total simulation QA time was decreased on average of approximately 20% after establishing new site-specific range tolerances. We simulated the financial impact of this project. The QA failure for whole process in proton therapy would lead up to approximately 30% increase in total cost. DMAIC framework can be used to provide an effective QA by setting customized tolerances. When tolerance design is customized, the quality is reasonably balanced with time and cost demands. © 2017 American Association of Physicists in Medicine.

  1. SU-F-T-226: QA Management for a Large Institution with Multiple Campuses for FMEA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tang, G; Chan, M; Lovelock, D

    2016-06-15

    Purpose: To redesign our radiation therapy QA program with the goal to improve quality, efficiency, and consistency among a growing number of campuses at a large institution. Methods: A QA committee was established with at least one physicist representing each of our six campuses (22 linacs). Weekly meetings were scheduled to advise on and update current procedures, to review end-to-end and other test results, and to prepare composite reports for internal and external audits. QA procedures for treatment and imaging equipment were derived from TG Reports 142 and 66, practice guidelines, and feedback from ACR evaluations. The committee focused onmore » reaching a consensus on a single QA program among all campuses using the same type of equipment and reference data. Since the recommendations for tolerances referenced to baseline data were subject to interpretation in some instances, the committee reviewed the characteristics of all machines and quantified any variations before choosing between treatment planning system (i.e. treatment planning system commissioning data that is representative for all machines) or machine-specific values (i.e. commissioning data of the individual machines) as baseline data. Results: The configured QA program will be followed strictly by all campuses. Inventory of available equipment has been compiled, and additional equipment acquisitions for the QA program are made as needed. Dosimetric characteristics are evaluated for all machines using the same methods to ensure consistency of beam data where possible. In most cases, baseline data refer to treatment planning system commissioning data but machine-specific values are used as reference where it is deemed appropriate. Conclusion: With a uniform QA scheme, variations in QA procedures are kept to a minimum. With a centralized database, data collection and analysis are simplified. This program will facilitate uniformity in patient treatments and analysis of large amounts of QA data

  2. SU-F-T-489: 4-Years Experience of QA in TomoTherapy MVCT: What Do We Look Out For?

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lee, F; Chan, K

    2016-06-15

    Purpose: To evaluate the QA results of TomoTherapy MVCT from March 2012 to February 2016, and to identify issues that may affect consistency in HU numbers and reconstructed treatment dose in MVCT. Methods: Monthly QA was performed on our TomoHD system. Phantom with rod inserts of various mass densities was imaged in MVCT and compared to baseline to evaluate HU number consistency. To evaluate treatment dose reconstructed by delivered sinogram and MVCT, a treatment plan was designed on a humanoid skull phantom. The phantom was imaged with MVCT and treatment plan was delivered to obtain the sinogram. The dose reconstructedmore » with the Planned Adaptive software was compared to the dose in the original plan. The QA tolerance for HU numbers was ±30 HU, and ±2% for discrepancy between original plan dose and reconstructed dose. Tolerances were referenced to AAPM TG148. Results: Several technical modifications or maintenance activities to the system have been identified which affected QA Results: 1) Upgrade in console system software which added a weekly HU calibration procedure; 2) Linac or MLC replacement leading to change in Accelerator Output Machine (AOM) parameters; 3) Upgrade in planning system algorithm affecting MVCT dose reconstruction. These events caused abrupt changes in QA results especially for the reconstructed dose. In the past 9 months, when no such modifications were done to the system, reconstructed dose was consistent with maximum deviation from baseline less than 0.6%. The HU number deviated less than 5HU. Conclusion: Routine QA is essential for MVCT, especially if the MVCT is used for daily dose reconstruction to monitor delivered dose to patients. Several technical events which may affect consistency of this are software changes, linac or MLC replacement. QA results reflected changes which justify re-calibration or system adjustment. In normal circumstances, the system should be relatively stable and quarterly QA may be sufficient.« less

  3. SU-E-T-392: Evaluation of Ion Chamber/film and Log File Based QA to Detect Delivery Errors

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nelson, C; Mason, B; Kirsner, S

    2015-06-15

    Purpose: Ion chamber and film (ICAF) is a method used to verify patient dose prior to treatment. More recently, log file based QA has been shown as an alternative for measurement based QA. In this study, we delivered VMAT plans with and without errors to determine if ICAF and/or log file based QA was able to detect the errors. Methods: Using two VMAT patients, the original treatment plan plus 7 additional plans with delivery errors introduced were generated and delivered. The erroneous plans had gantry, collimator, MLC, gantry and collimator, collimator and MLC, MLC and gantry, and gantry, collimator, andmore » MLC errors. The gantry and collimator errors were off by 4{sup 0} for one of the two arcs. The MLC error introduced was one in which the opening aperture didn’t move throughout the delivery of the field. For each delivery, an ICAF measurement was made as well as a dose comparison based upon log files. Passing criteria to evaluate the plans were ion chamber less and 5% and film 90% of pixels pass the 3mm/3% gamma analysis(GA). For log file analysis 90% of voxels pass the 3mm/3% 3D GA and beam parameters match what was in the plan. Results: Two original plans were delivered and passed both ICAF and log file base QA. Both ICAF and log file QA met the dosimetry criteria on 4 of the 12 erroneous cases analyzed (2 cases were not analyzed). For the log file analysis, all 12 erroneous plans alerted a mismatch in delivery versus what was planned. The 8 plans that didn’t meet criteria all had MLC errors. Conclusion: Our study demonstrates that log file based pre-treatment QA was able to detect small errors that may not be detected using an ICAF and both methods of were able to detect larger delivery errors.« less

  4. Organization of the qa Gene Cluster in NEUROSPORA CRASSA: Direction of Transcription of the qa-3 Gene

    PubMed Central

    Strøman, Per; Reinert, William; Case, Mary E.; Giles, Norman H.

    1979-01-01

    In Neurospora crassa, the enzyme quinate (shikimate) dehydrogenase catalyzes the first reaction in the inducible quinic acid catabolic pathway and is encoded in the qa-3 gene of the qa cluster. In this cluster, the order of genes has been established as qa-1 qa-3 qa-4 qa-2. Amino-terminal sequences have been determined for purified quinate dehydrogenase from wild type and from UV-induced revertants in two different qa-3 mutants. These two mutants (M16 and M45) map at opposite ends of the qa-3 locus. In addition, mapping data (Case et al. 1978) indicate that the end of the qa-3 gene specified by M45 is closer to the adjacent qa-1 gene than is the end specified by the M16 mutant site. In one of the revertants (R45 from qa-3 mutant M45), the aminoterminal sequence for the first ten amino acids is identical to that of wild type. The other revertant (R1 from qa-3 mutant M16) differs from wild type at the amino-terminal end by a single altered residue at position three in the sequence. The observed change involves the substitution of an isoleucine in M16-R1 for a proline in wild type. This substitution requires a two-nucleotide change in the corresponding wild-type codon.——The combined genetic and biochemical data indicate that the qa-3 mutants M16 and M45 carry amino acid substitutions near the amino-terminal and carboxyl-terminal ends of the quinate dehydrogenase enzyme, respectively. On this basis we conclude that transcription of the qa-3 gene proceeds from the end specified by the M16 mutant site in the direction of the qa-1 gene. It appears probable that transcription is initiated from a promoter site within the qa cluster, possibly immediately adjacent to the qa-3 gene. PMID:159203

  5. WE-AB-201-02: TPS Commissioning and QA: A Process Orientation and Application of Control Charts

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sharpe, M.

    Treatment planning systems (TPS) are a cornerstone of modern radiation therapy. Errors in their commissioning or use can have a devastating impact on many patients. To support safe and high quality care, medical physicists must conduct efficient and proper commissioning, good clinical integration, and ongoing quality assurance (QA) of the TPS. AAPM Task Group 53 and related publications have served as seminal benchmarks for TPS commissioning and QA over the past two decades. Over the same time, continuing innovations have made the TPS even more complex and more central to the clinical process. Medical goals are now expressed in termsmore » of the dose and margins around organs and tissues that are delineated from multiple imaging modalities (CT, MR and PET); and even temporally resolved (i.e., 4D) imaging. This information is passed on to optimization algorithms to establish accelerator movements that are programmed directly for IMRT, VMAT and stereotactic treatments. These advances have made commissioning and QA of the TPS much more challenging. This education session reviews up-to-date experience and guidance on this subject; including the recently published AAPM Medical Physics Practice Guideline (MPPG) #5 “Commissioning and QA of Treatment Planning Dose Calculations: Megavoltage Photon and Electron Beams”. Treatment Planning System Commissioning and QA: Challenges and Opportunities (Greg Salomons) This session will provide some key background and review publications describing prominent incidents relating to TPS commissioning and QA. Traditional approaches have been hardware and feature oriented. They aim to establish a functional configuration and establish specifications for regular testing of features (like dose calculation) to assure stable operation and detect failures. With the advent of more complex systems, more patient-specific testing has also been adopted. A number of actual TPS defects will be presented along with heuristics for identifying

  6. SU-F-T-264: VMAT QA with 2D Radiation Measuring Equipment Attached to Gantry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fung, A

    2016-06-15

    Purpose: To introduce a method of VMAT QA by 2D measuring device. The 2D device is attached to the gantry throughout measurement duration. This eliminates error caused by the angular dependence of the radiation detectors. Methods: A 2D radiation measuring device was attached to the gantry of linear accelerator. The center of the detector plane was at the isocenter. For each patient plan, two verification plans were created for QA purpose. One was like an ordinary VMAT plan, to be used for radiation delivery. The other is a plan with gantry angle fixed at zero, so the dose distribution asmore » seen by the rotating 2D device. Points above 10% dose threshold were analyzed. Data is in tolerance if it fits within the 3 mm or 3% dose gamma criteria. For each patient, the plan was passed when 95% of all the points in the 2D matrix fit the gamma criteria. The following statistics were calculated: number of patient plans passed, percentage of all points passed, average percentage difference of all points. Results: VMAT QA was performed for patients during one year in our department, and the results were analyzed. All irradiation was with 6 MV photon beam. Each plan has calculated and measured doses compared. After collecting one year’s result, with 81 patient plans analyzed, all (100%) of the plans passed the gamma criteria. Of the points analyzed from all plans, 98.8% of all points passed. Conclusion: This method of attaching a 2D measuring device on the linac gantry proves to be an accurate way for VMAT QA. It is simple to use and low cost, and it eliminates the problem of directional dependence.« less

  7. QA/QC in the laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hood, F.C.

    1992-05-01

    Quality assurance and quality control (QA/QC) of analytical chemistry laboratory activities are essential to the validity and usefulness of resultant data. However, in themselves, conventional QA/QC measures will not always ensure that fraudulent data are not generated. Conventional QA/QC measures are based on the assumption that work will be done in good faith; to assure against fraudulent practices, QA/QC measures must be tailored to specific analyses protocols in anticipation of intentional misapplication of those protocols. Application of specific QA/QC measures to ensure against fraudulent practices result in an increased administrative burden being placed on the analytical process; accordingly, in keepingmore » with graded QA philosophy, data quality objectives must be used to identify specific points of concern for special control to minimize the administrative impact.« less

  8. SU-F-T-271: Comparing IMRT QA Pass Rates Before and After MLC Calibration

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mazza, A; Perrin, D; Fontenot, J

    Purpose: To compare IMRT QA pass rates before and after an in-house MLC leaf calibration procedure. Methods: The MLC leaves and backup jaws on four Elekta linear accelerators with MLCi2 heads were calibrated using the EPID-based RIT Hancock Test as the means for evaluation. The MLCs were considered to be successfully calibrated when they could pass the Hancock Test with criteria of 1 mm jaw position tolerance, and 1 mm leaf position tolerance. IMRT QA results were collected pre- and postcalibration and analyzed using gamma analysis with 3%/3mm DTA criteria. AAPM TG-119 test plans were also compared pre- and post-calibration,more » at both 2%/2mm DTA and 3%/3mm DTA. Results: A weighted average was performed on the results for all four linear accelerators. The pre-calibration IMRT QA pass rate was 98.3 ± 0.1%, compared with the post-calibration pass rate of 98.5 ± 0.1%. The TG-119 test plan results showed more of an improvement, particularly at the 2%/2mm criteria. The averaged results were 89.1% pre and 96.1% post for the C-shape plan, 94.8% pre and 97.1% post for the multi-target plan, 98.6% pre and 99.7% post for the prostate plan, 94.7% pre and 94.8% post for the head/neck plan. Conclusion: The patient QA results did not show statistically significant improvement at the 3%/3mm DTA criteria after the MLC calibration procedure. However, the TG-119 test cases did show significant improvement at the 2%/2mm level.« less

  9. A new Tla region antigen Qa-11, similar to Qa-2 and associated with B-type beta 2-microglobulin.

    PubMed

    van de Meugheuvel, W; van Seventer, G; Demant, P

    1985-04-01

    A new antigen, Qa-11, is detected as a 40,000 dalton band in the SDS-PAGE of immunoprecipitates of radiolabeled lymphocyte membrane preparations. In C57BL H-2 congenic strains, its presence is controlled by a gene in the Tla region. In strains with genetic background other than C57BL it is not expressed. Tests with recombinant inbred strains and with H-3 congenic strains show that, in addition to the Tla region, a gene linked to or identical with the beta 2-microglobulin-b-allele is required for the expression of Qa-11 as well. The mobility of the Qa-11 antigen in SDS-PAGE and in isoelectrofocusing is the same as that of Qa-2 antigen. The Cleveland peptide maps of Qa-2 and Qa-11 are identical as well. This finding, that the Tla region controlled Qa-11 antigen is structurally very similar to the Qa-2 antigen, contrasts with the fact that Tla region products do not react with anti-Qa-2 sera. This paradox could be explained by a separate Qa-11 region between Qa-2 and Tla. Alternatively, it is possible that the Qa-11 antigen is the result of the action of a modifying gene in the Tla region upon a Qa-2 gene product, or that the structural gene for Qa-11 is located in the Qa-2 region and a Tla region gene controls its expression.

  10. Tolerance limits and methodologies for IMRT measurement-based verification QA: Recommendations of AAPM Task Group No. 218.

    PubMed

    Miften, Moyed; Olch, Arthur; Mihailidis, Dimitris; Moran, Jean; Pawlicki, Todd; Molineu, Andrea; Li, Harold; Wijesooriya, Krishni; Shi, Jie; Xia, Ping; Papanikolaou, Nikos; Low, Daniel A

    2018-04-01

    Patient-specific IMRT QA measurements are important components of processes designed to identify discrepancies between calculated and delivered radiation doses. Discrepancy tolerance limits are neither well defined nor consistently applied across centers. The AAPM TG-218 report provides a comprehensive review aimed at improving the understanding and consistency of these processes as well as recommendations for methodologies and tolerance limits in patient-specific IMRT QA. The performance of the dose difference/distance-to-agreement (DTA) and γ dose distribution comparison metrics are investigated. Measurement methods are reviewed and followed by a discussion of the pros and cons of each. Methodologies for absolute dose verification are discussed and new IMRT QA verification tools are presented. Literature on the expected or achievable agreement between measurements and calculations for different types of planning and delivery systems are reviewed and analyzed. Tests of vendor implementations of the γ verification algorithm employing benchmark cases are presented. Operational shortcomings that can reduce the γ tool accuracy and subsequent effectiveness for IMRT QA are described. Practical considerations including spatial resolution, normalization, dose threshold, and data interpretation are discussed. Published data on IMRT QA and the clinical experience of the group members are used to develop guidelines and recommendations on tolerance and action limits for IMRT QA. Steps to check failed IMRT QA plans are outlined. Recommendations on delivery methods, data interpretation, dose normalization, the use of γ analysis routines and choice of tolerance limits for IMRT QA are made with focus on detecting differences between calculated and measured doses via the use of robust analysis methods and an in-depth understanding of IMRT verification metrics. The recommendations are intended to improve the IMRT QA process and establish consistent, and comparable IMRT QA

  11. SU-F-T-558: ArcCheck for Patient Specific QA in Stereotactic Ablative Radiotherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ramachandran, P; RMIT University, Bundoora; Tajaldeen, A

    2016-06-15

    Purpose: Stereotactic Ablative Radiotherapy (SABR) is one of the most preferred treatment techniques for early stage lung cancer. This technique has been extended to other treatment sites like Spine, Liver, Scapula, Sternum etc., This has resulted in increased physics QA time on machine. In this study, we’ve tested the feasibility of using ArcCheck as an alternative method to replace film dosimetry. Methods: Twelve patients with varied diagnosis of Lung, Liver, scapula, sternum and Spine undergoing SABR were selected for this study. Pre-treatment QA was performed for all the patients which include ionization chamber and film dosimetry. The required gamma criteriamore » for each SABR plan to pass QA and proceed to treatment is 95% (3%,1mm). In addition to this routine process, the treatment plans were exported on to an ArcCheck phantom. The planned and measured dose from the ArcCheck device were compared using four different gamma criteria: 2%,2 mm, 3%,2 mm, 3%,1 mm and 3%, 3 mm. In addition to this, we’ve also introduced errors to gantry, collimator and couch angle to assess sensitivity of the ArcCheck with potential delivery errors. Results: The ArcCheck mean passing rates for all twelve cases were 76.1%±9.7% for gamma criteria 3%,1 mm, 89.5%±5.3% for 2%,2 mm, 92.6%±4.2% for 3%,2 mm, and 97.6%±2.4% for 3%,3 mm gamma criteria. When SABR spine cases are excluded, we observe ArcCheck passing rates higher than 95% for all the studied cases with 3%, 3mm, and ArcCheck results in acceptable agreement with the film gamma results. Conclusion: Our ArcCheck results at 3%, 3 mm were found to correlate well with our non-SABR spine routine patient specific QA results (3%,1 mm). We observed significant reduction in QA time on using ArcCheck for SABR QA. This study shows that ArcCheck could replace film dosimetry for all sites except SABR spine.« less

  12. Test/QA plan for the verification testing of diesel exhaust catalysts, particulate filters and engine modification control technologies for highway and nonroad use diesel engines

    EPA Science Inventory

    This ETV test/QA plan for heavy-duty diesel engine testing at the Southwest Research Institute’s Department of Emissions Research (DER) describes how the Federal Test Procedure (FTP), as listed in 40 CFR Part 86 for highway engines and 40 CFR Part 89 for nonroad engines, will be ...

  13. A Varian DynaLog file-based procedure for patient dose-volume histogram-based IMRT QA.

    PubMed

    Calvo-Ortega, Juan F; Teke, Tony; Moragues, Sandra; Pozo, Miquel; Casals-Farran, Joan

    2014-03-06

    In the present study, we describe a method based on the analysis of the dynamic MLC log files (DynaLog) generated by the controller of a Varian linear accelerator in order to perform patient-specific IMRT QA. The DynaLog files of a Varian Millennium MLC, recorded during an IMRT treatment, can be processed using a MATLAB-based code in order to generate the actual fluence for each beam and so recalculate the actual patient dose distribution using the Eclipse treatment planning system. The accuracy of the DynaLog-based dose reconstruction procedure was assessed by introducing ten intended errors to perturb the fluence of the beams of a reference plan such that ten subsequent erroneous plans were generated. In-phantom measurements with an ionization chamber (ion chamber) and planar dose measurements using an EPID system were performed to investigate the correlation between the measured dose changes and the expected ones detected by the reconstructed plans for the ten intended erroneous cases. Moreover, the method was applied to 20 cases of clinical plans for different locations (prostate, lung, breast, and head and neck). A dose-volume histogram (DVH) metric was used to evaluate the impact of the delivery errors in terms of dose to the patient. The ionometric measurements revealed a significant positive correlation (R² = 0.9993) between the variations of the dose induced in the erroneous plans with respect to the reference plan and the corresponding changes indicated by the DynaLog-based reconstructed plans. The EPID measurements showed that the accuracy of the DynaLog-based method to reconstruct the beam fluence was comparable with the dosimetric resolution of the portal dosimetry used in this work (3%/3 mm). The DynaLog-based reconstruction method described in this study is a suitable tool to perform a patient-specific IMRT QA. This method allows us to perform patient-specific IMRT QA by evaluating the result based on the DVH metric of the planning CT image (patient

  14. IMRT QA: Selecting gamma criteria based on error detection sensitivity

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Steers, Jennifer M.; Fraass, Benedick A., E-mail: benedick.fraass@cshs.org

    Purpose: The gamma comparison is widely used to evaluate the agreement between measurements and treatment planning system calculations in patient-specific intensity modulated radiation therapy (IMRT) quality assurance (QA). However, recent publications have raised concerns about the lack of sensitivity when employing commonly used gamma criteria. Understanding the actual sensitivity of a wide range of different gamma criteria may allow the definition of more meaningful gamma criteria and tolerance limits in IMRT QA. We present a method that allows the quantitative determination of gamma criteria sensitivity to induced errors which can be applied to any unique combination of device, delivery technique,more » and software utilized in a specific clinic. Methods: A total of 21 DMLC IMRT QA measurements (ArcCHECK®, Sun Nuclear) were compared to QA plan calculations with induced errors. Three scenarios were studied: MU errors, multi-leaf collimator (MLC) errors, and the sensitivity of the gamma comparison to changes in penumbra width. Gamma comparisons were performed between measurements and error-induced calculations using a wide range of gamma criteria, resulting in a total of over 20 000 gamma comparisons. Gamma passing rates for each error class and case were graphed against error magnitude to create error curves in order to represent the range of missed errors in routine IMRT QA using 36 different gamma criteria. Results: This study demonstrates that systematic errors and case-specific errors can be detected by the error curve analysis. Depending on the location of the error curve peak (e.g., not centered about zero), 3%/3 mm threshold = 10% at 90% pixels passing may miss errors as large as 15% MU errors and ±1 cm random MLC errors for some cases. As the dose threshold parameter was increased for a given %Diff/distance-to-agreement (DTA) setting, error sensitivity was increased by up to a factor of two for select cases. This increased sensitivity with increasing

  15. IMRT QA: Selecting gamma criteria based on error detection sensitivity.

    PubMed

    Steers, Jennifer M; Fraass, Benedick A

    2016-04-01

    The gamma comparison is widely used to evaluate the agreement between measurements and treatment planning system calculations in patient-specific intensity modulated radiation therapy (IMRT) quality assurance (QA). However, recent publications have raised concerns about the lack of sensitivity when employing commonly used gamma criteria. Understanding the actual sensitivity of a wide range of different gamma criteria may allow the definition of more meaningful gamma criteria and tolerance limits in IMRT QA. We present a method that allows the quantitative determination of gamma criteria sensitivity to induced errors which can be applied to any unique combination of device, delivery technique, and software utilized in a specific clinic. A total of 21 DMLC IMRT QA measurements (ArcCHECK®, Sun Nuclear) were compared to QA plan calculations with induced errors. Three scenarios were studied: MU errors, multi-leaf collimator (MLC) errors, and the sensitivity of the gamma comparison to changes in penumbra width. Gamma comparisons were performed between measurements and error-induced calculations using a wide range of gamma criteria, resulting in a total of over 20 000 gamma comparisons. Gamma passing rates for each error class and case were graphed against error magnitude to create error curves in order to represent the range of missed errors in routine IMRT QA using 36 different gamma criteria. This study demonstrates that systematic errors and case-specific errors can be detected by the error curve analysis. Depending on the location of the error curve peak (e.g., not centered about zero), 3%/3 mm threshold = 10% at 90% pixels passing may miss errors as large as 15% MU errors and ±1 cm random MLC errors for some cases. As the dose threshold parameter was increased for a given %Diff/distance-to-agreement (DTA) setting, error sensitivity was increased by up to a factor of two for select cases. This increased sensitivity with increasing dose threshold was consistent

  16. SU-E-T-29: A Web Application for GPU-Based Monte Carlo IMRT/VMAT QA with Delivered Dose Verification

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Folkerts, M; University of California, San Diego, La Jolla, CA; Graves, Y

    Purpose: To enable an existing web application for GPU-based Monte Carlo (MC) 3D dosimetry quality assurance (QA) to compute “delivered dose” from linac logfile data. Methods: We added significant features to an IMRT/VMAT QA web application which is based on existing technologies (HTML5, Python, and Django). This tool interfaces with python, c-code libraries, and command line-based GPU applications to perform a MC-based IMRT/VMAT QA. The web app automates many complicated aspects of interfacing clinical DICOM and logfile data with cutting-edge GPU software to run a MC dose calculation. The resultant web app is powerful, easy to use, and is ablemore » to re-compute both plan dose (from DICOM data) and delivered dose (from logfile data). Both dynalog and trajectorylog file formats are supported. Users upload zipped DICOM RP, CT, and RD data and set the expected statistic uncertainty for the MC dose calculation. A 3D gamma index map, 3D dose distribution, gamma histogram, dosimetric statistics, and DVH curves are displayed to the user. Additional the user may upload the delivery logfile data from the linac to compute a 'delivered dose' calculation and corresponding gamma tests. A comprehensive PDF QA report summarizing the results can also be downloaded. Results: We successfully improved a web app for a GPU-based QA tool that consists of logfile parcing, fluence map generation, CT image processing, GPU based MC dose calculation, gamma index calculation, and DVH calculation. The result is an IMRT and VMAT QA tool that conducts an independent dose calculation for a given treatment plan and delivery log file. The system takes both DICOM data and logfile data to compute plan dose and delivered dose respectively. Conclusion: We sucessfully improved a GPU-based MC QA tool to allow for logfile dose calculation. The high efficiency and accessibility will greatly facilitate IMRT and VMAT QA.« less

  17. WE-D-BRA-06: IMRT QA with ArcCHECK: The MD Anderson Experience

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Aristophanous, M; Suh, Y; Chi, P

    Purpose: The objective of this project was to report our initial IMRT QA results and experience with the SunNuclear ArcCHECK. Methods: Three thousand one-hundred and sixteen cases were treated with IMRT or VMAT at our institution between October 2013 and September 2014. All IMRT/VMAT treatment plans underwent Quality Assurance (QA) using ArcCHECK prior to therapy. For clinical evaluation, a Gamma analysis is performed following QA delivery using the SNC Patient software (Sun Nuclear Corp) at the 3%/3mm level. QA Gamma pass rates were analyzed based on categories of treatment site, technique, and type of MLCs. Our current clinical threshold formore » passing a QA (Tclin) is set at a Gamma pass rate greater than 90%. We recorded the percent of failures for each category, as well as the Gamma pass rate threshold that would Result in 95% of QAs to pass (T95). Results: Using Tclin a failure rate of 5.9% over all QAs was observed. The highest failure rate was observed for gynecological (22%) and the lowest for CNS (0.9%) treatments. T95 was 91% over all QAs and ranged from 73% (gynecological) to 96.5% (CNS) for individual treatments sites. T95 was lower for IMRT and non-HD (high definition) MLCs at 88.5% and 94.5%, respectively, compared to 92.4% and 97.1% for VMAT and HD MLC treatments, respectively. There was a statistically significant difference between the passing rates for IMRT vs. VMAT and for HD MLCs vs. non-HD MLCs (p-values << 0.01). Gynecological, IMRT, and HD MLC treatments typically include more plans with larger field sizes. Conclusion: On average, Tclin with ArcCHECK was consistent with T95, as well as the 90% action level reported in TG-119. However, significant variations between the examined categories suggest a link between field size and QA passing rates and may warrant field size-specific passing rate thresholds.« less

  18. QA/QC in the laboratory. Session F

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hood, F.C.

    1992-05-01

    Quality assurance and quality control (QA/QC) of analytical chemistry laboratory activities are essential to the validity and usefulness of resultant data. However, in themselves, conventional QA/QC measures will not always ensure that fraudulent data are not generated. Conventional QA/QC measures are based on the assumption that work will be done in good faith; to assure against fraudulent practices, QA/QC measures must be tailored to specific analyses protocols in anticipation of intentional misapplication of those protocols. Application of specific QA/QC measures to ensure against fraudulent practices result in an increased administrative burden being placed on the analytical process; accordingly, in keepingmore » with graded QA philosophy, data quality objectives must be used to identify specific points of concern for special control to minimize the administrative impact.« less

  19. Test/QA plan for the verification testing of alternative or reformulated liquid fuels, fuel additives, fuel emulsions, and lubricants for highway and nonroad use heavy-duty diesel engines

    EPA Science Inventory

    This Environmental Technology Verification Program test/QA plan for heavy-duty diesel engine testing at the Southwest Research Institute’s Department of Emissions Research describes how the Federal Test Procedure (FTP), as listed in 40 CFR Part 86 for highway engines and 40 CFR P...

  20. SU-E-T-760: Tolerance Design for Site-Specific Range in Proton Patient QA Process Using the Six Sigma Model

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lah, J; Shin, D; Kim, G

    Purpose: To show how tolerance design and tolerancing approaches can be used to predict and improve the site-specific range in patient QA process in implementing the Six Sigma. Methods: In this study, patient QA plans were selected according to 6 site-treatment groups: head &neck (94 cases), spine (76 cases), lung (89 cases), liver (53 cases), pancreas (55 cases), and prostate (121 cases), treated between 2007 and 2013. We evaluated a model of the Six Sigma that determines allowable deviations in design parameters and process variables in patient-specific QA, where possible, tolerance may be loosened, then customized if it necessary tomore » meet the functional requirements. A Six Sigma problem-solving methodology is known as DMAIC phases, which are used stand for: Define a problem or improvement opportunity, Measure process performance, Analyze the process to determine the root causes of poor performance, Improve the process by fixing root causes, Control the improved process to hold the gains. Results: The process capability for patient-specific range QA is 0.65 with only ±1 mm of tolerance criteria. Our results suggested the tolerance level of ±2–3 mm for prostate and liver cases and ±5 mm for lung cases. We found that customized tolerance between calculated and measured range reduce that patient QA plan failure and almost all sites had failure rates less than 1%. The average QA time also improved from 2 hr to less than 1 hr for all including planning and converting process, depth-dose measurement and evaluation. Conclusion: The objective of tolerance design is to achieve optimization beyond that obtained through QA process improvement and statistical analysis function detailing to implement a Six Sigma capable design.« less

  1. Improving spot-scanning proton therapy patient specific quality assurance with HPlusQA, a second-check dose calculation engine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mackin, Dennis; Li, Yupeng; Taylor, Michael B.

    Purpose: The purpose of this study was to validate the use of HPlusQA, spot-scanning proton therapy (SSPT) dose calculation software developed at The University of Texas MD Anderson Cancer Center, as second-check dose calculation software for patient-specific quality assurance (PSQA). The authors also showed how HPlusQA can be used within the current PSQA framework.Methods: The authors compared the dose calculations of HPlusQA and the Eclipse treatment planning system with 106 planar dose measurements made as part of PSQA. To determine the relative performance and the degree of correlation between HPlusQA and Eclipse, the authors compared calculated with measured point doses.more » Then, to determine how well HPlusQA can predict when the comparisons between Eclipse calculations and the measured dose will exceed tolerance levels, the authors compared gamma index scores for HPlusQA versus Eclipse with those of measured doses versus Eclipse. The authors introduce the αβγ transformation as a way to more easily compare gamma scores.Results: The authors compared measured and calculated dose planes using the relative depth, z/R × 100%, where z is the depth of the measurement and R is the proton beam range. For relative depths than less than 80%, both Eclipse and HPlusQA calculations were within 2 cGy of dose measurements on average. When the relative depth was greater than 80%, the agreement between the calculations and measurements fell to 4 cGy. For relative depths less than 10%, the Eclipse and HPlusQA dose discrepancies showed a negative correlation, −0.21. Otherwise, the correlation between the dose discrepancies was positive and as large as 0.6. For the dose planes in this study, HPlusQA correctly predicted when Eclipse had and had not calculated the dose to within tolerance 92% and 79% of the time, respectively. In 4 of 106 cases, HPlusQA failed to predict when the comparison between measurement and Eclipse's calculation had exceeded the tolerance levels of

  2. Improving spot-scanning proton therapy patient specific quality assurance with HPlusQA, a second-check dose calculation engine.

    PubMed

    Mackin, Dennis; Li, Yupeng; Taylor, Michael B; Kerr, Matthew; Holmes, Charles; Sahoo, Narayan; Poenisch, Falk; Li, Heng; Lii, Jim; Amos, Richard; Wu, Richard; Suzuki, Kazumichi; Gillin, Michael T; Zhu, X Ronald; Zhang, Xiaodong

    2013-12-01

    The purpose of this study was to validate the use of HPlusQA, spot-scanning proton therapy (SSPT) dose calculation software developed at The University of Texas MD Anderson Cancer Center, as second-check dose calculation software for patient-specific quality assurance (PSQA). The authors also showed how HPlusQA can be used within the current PSQA framework. The authors compared the dose calculations of HPlusQA and the Eclipse treatment planning system with 106 planar dose measurements made as part of PSQA. To determine the relative performance and the degree of correlation between HPlusQA and Eclipse, the authors compared calculated with measured point doses. Then, to determine how well HPlusQA can predict when the comparisons between Eclipse calculations and the measured dose will exceed tolerance levels, the authors compared gamma index scores for HPlusQA versus Eclipse with those of measured doses versus Eclipse. The authors introduce the αβγ transformation as a way to more easily compare gamma scores. The authors compared measured and calculated dose planes using the relative depth, z∕R × 100%, where z is the depth of the measurement and R is the proton beam range. For relative depths than less than 80%, both Eclipse and HPlusQA calculations were within 2 cGy of dose measurements on average. When the relative depth was greater than 80%, the agreement between the calculations and measurements fell to 4 cGy. For relative depths less than 10%, the Eclipse and HPlusQA dose discrepancies showed a negative correlation, -0.21. Otherwise, the correlation between the dose discrepancies was positive and as large as 0.6. For the dose planes in this study, HPlusQA correctly predicted when Eclipse had and had not calculated the dose to within tolerance 92% and 79% of the time, respectively. In 4 of 106 cases, HPlusQA failed to predict when the comparison between measurement and Eclipse's calculation had exceeded the tolerance levels of 3% for dose and 3 mm for

  3. Specific application for Oak Ridge National Laboratory dismantlement of Building 3004. Appendix A -- Quality assurance plan; Appendix B -- Records management plan

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    This quality assurance (QA) plan defines the QA requirements for the dismantlement and removal of Building 3004 at Oak Ridge National Laboratory (ORNL). The building is a four-story wooden trained structure with wooden siding, which resides approximately 150 ft west of the Bulk Shielding Reactor, and only several feet away from the visitors entrance to the Graphite Reactor museum. Complete descriptions and sketches are in the Performance Specification document for this project. This project is being conducted as a non-CERCLA maintenance action. This plan is an appendix to the QA plan for the ORNL Environmental Restoration (ER) Program. ORNL/ER-225, whichmore » is the source of the project QA requirements, tailors those QA requirements to the specific needs of this project as defined in ORNL/ER-225. Project-specific description and organization are also provided in this plan. Appendix B, Records Management Plan, is included.« less

  4. 77 FR 61046 - The Amendment of the Designation of Al-Qa'ida in the Arabian Peninsula, aka Al-Qa'ida of Jihad...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-05

    ... DEPARTMENT OF STATE [Public Notice 8054] The Amendment of the Designation of Al-Qa'ida in the Arabian Peninsula, aka Al-Qa'ida of Jihad Organization in the Arabian Peninsula, aka Tanzim Qa'idat al-Jihad fi Jazirat al-Arab, aka Al- Qa'ida in Yemen, aka Al-Qa'ida in the South Arabian Peninsula, as a...

  5. 77 FR 61046 - The Review and Amendment of the Designation of Al-Qa'ida in the Arabian Peninsula, aka Al-Qa'ida...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-05

    ... DEPARTMENT OF STATE [Public Notice 8055] The Review and Amendment of the Designation of Al-Qa'ida in the Arabian Peninsula, aka Al-Qa'ida of Jihad Organization in the Arabian Peninsula, aka Tanzim Qa'idat al-Jihad fi Jazirat al-Arab, aka Al- Qa'ida in Yemen, aka Al-Qa'ida in the South Arabian Peninsula...

  6. A virtual dosimetry audit - Towards transferability of gamma index analysis between clinical trial QA groups.

    PubMed

    Hussein, Mohammad; Clementel, Enrico; Eaton, David J; Greer, Peter B; Haworth, Annette; Ishikura, Satoshi; Kry, Stephen F; Lehmann, Joerg; Lye, Jessica; Monti, Angelo F; Nakamura, Mitsuhiro; Hurkmans, Coen; Clark, Catharine H

    2017-12-01

    Quality assurance (QA) for clinical trials is important. Lack of compliance can affect trial outcome. Clinical trial QA groups have different methods of dose distribution verification and analysis, all with the ultimate aim of ensuring trial compliance. The aim of this study was to gain a better understanding of different processes to inform future dosimetry audit reciprocity. Six clinical trial QA groups participated. Intensity modulated treatment plans were generated for three different cases. A range of 17 virtual 'measurements' were generated by introducing a variety of simulated perturbations (such as MLC position deviations, dose differences, gantry rotation errors, Gaussian noise) to three different treatment plan cases. Participants were blinded to the 'measured' data details. Each group analysed the datasets using their own gamma index (γ) technique and using standardised parameters for passing criteria, lower dose threshold, γ normalisation and global γ. For the same virtual 'measured' datasets, different results were observed using local techniques. For the standardised γ, differences in the percentage of points passing with γ < 1 were also found, however these differences were less pronounced than for each clinical trial QA group's analysis. These variations may be due to different software implementations of γ. This virtual dosimetry audit has been an informative step in understanding differences in the verification of measured dose distributions between different clinical trial QA groups. This work lays the foundations for audit reciprocity between groups, particularly with more clinical trials being open to international recruitment. Copyright © 2017 Elsevier B.V. All rights reserved.

  7. GEOSPATIAL IT/IM QA CHECKLIST

    EPA Science Inventory

    Quality assurance (QA) of information technology (IT) and Information Management (IM) systems help to ensure that the end product is of known quality and integrity. As the complexity of IT & IM processes increase, so does the need for regular QA evaluation.

    The areas revi...

  8. SU-G-201-01: An Automated Treatment Plan Quality Assurance Program for High-Dose Rate (HDR) Brachytherapy with a VaginalCylinder Applicator

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhou, Y; Tan, J; Jiang, S

    Purpose: Plan specific quality assurance (QA) is an important step in high dose rate (HDR) brachytherapy to ensure the integrity of a treatment plan. The conventional approach is to assemble a set of plan screen-captures in a document and have an independent plan-checker to verify it. Not only is this approach cumbersome and time-consuming, using a document also limits the items that can be verified, hindering plan quality and patient safety. We have initiated efforts to develop a web-based HDR brachytherapy QA system called AutoBrachy QA, for comprehensive and efficient QA. This abstract reports a new plugin in this systemmore » for the QA of a cylinder HDR brachytherapy treatment. Methods: A cylinder plan QA module was developed using Python. It was plugged into our AutoBrachy QA system. This module extracted information from CT images and treatment plan. Image processing techniques were employed to obtain geometric parameters, e.g. cylinder diameter. A comprehensive set of eight geometrical and eight dosimetric features of the plan were validated against user specified planning parameter, such as prescription value, treatment depth and length, etc. A PDF document was generated, consisting of a summary QA sheet with all the QA results, as well as images showing plan details. Results: The cylinder QA program has been implemented in our clinic. To date, it has been used in 11 patient cases and was able to successfully perform QA tests in all of them. The QA program reduced the average plan QA time from 7 min using conventional manual approach to 0.5 min. Conclusion: Being a new module in our AutoBrachy QA system, an automated treatment plan QA module for cylinder HDR brachytherapy has been successfully developed and clinically implemented. This module improved clinical workflow and plan integrity compared to the conventional manual approach.« less

  9. SU-E-T-88: Comprehensive Automated Daily QA for Hypo- Fractionated Treatments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    McGuinness, C; Morin, O

    2014-06-01

    Purpose: The trend towards more SBRT treatments with fewer high dose fractions places increased importance on daily QA. Patient plan specific QA with 3%/3mm gamma analysis and daily output constancy checks may not be enough to guarantee the level of accuracy required for SBRT treatments. But increasing the already extensive amount of QA procedures that are required is a daunting proposition. We performed a feasibility study for more comprehensive automated daily QA that could improve the diagnostic capabilities of QA without increasing workload. Methods: We performed the study on a Siemens Artiste linear accelerator using the integrated flat panel EPID.more » We included square fields, a picket fence, overlap and representative IMRT fields to measure output, flatness, symmetry, beam center, and percent difference from the standard. We also imposed a set of machine errors: MLC leaf position, machine output, and beam steering to compare with the standard. Results: Daily output was consistent within +/− 1%. Change in steering current by 1.4% and 2.4% resulted in a 3.2% and 6.3% change in flatness. 1 and 2mm MLC leaf offset errors were visibly obvious in difference plots, but passed a 3%/3mm gamma analysis. A simple test of transmission in a picket fence can catch a leaf offset error of a single leaf by 1mm. The entire morning QA sequence is performed in less than 30 minutes and images are automatically analyzed. Conclusion: Automated QA procedures could be used to provide more comprehensive information about the machine with less time and human involvement. We have also shown that other simple tests are better able to catch MLC leaf position errors than a 3%/3mm gamma analysis commonly used for IMRT and modulated arc treatments. Finally, this information could be used to watch trends of the machine and predict problems before they lead to costly machine downtime.« less

  10. A novel technique for VMAT QA with EPID in cine mode on a Varian TrueBeam linac

    NASA Astrophysics Data System (ADS)

    Liu, Bo; Adamson, Justus; Rodrigues, Anna; Zhou, Fugen; Yin, Fang-fang; Wu, Qiuwen

    2013-10-01

    Volumetric modulated arc therapy (VMAT) is a relatively new treatment modality for dynamic photon radiation therapy. Pre-treatment quality assurance (QA) is necessary and many efforts have been made to apply electronic portal imaging device (EPID)-based IMRT QA methods to VMAT. It is important to verify the gantry rotation speed during delivery as this is a new variable that is also modulated in VMAT. In this paper, we present a new technique to perform VMAT QA using an EPID. The method utilizes EPID cine mode and was tested on Varian TrueBeam in research mode. The cine images were acquired during delivery and converted to dose matrices after profile correction and dose calibration. A sub-arc corresponding to each cine image was extracted from the original plan and its portal image prediction was calculated. Several analyses were performed including 3D γ analysis (2D images + gantry angle axis), 2D γ analysis, and other statistical analyses. The method was applied to 21 VMAT photon plans of 3 photon energies. The accuracy of the cine image information was investigated. Furthermore, this method's sensitivity to machine delivery errors was studied. The pass rate (92.8 ± 1.4%) for 3D γ analysis was comparable to those from Delta4 system (99.9 ± 0.1%) under similar criteria (3%, 3 mm, 5% threshold and 2° angle to agreement) at 6 MV. The recorded gantry angle and start/stop MUs were found to have sufficient accuracy for clinical QA. Machine delivery errors can be detected through combined analyses of 3D γ, gantry angle, and percentage dose difference. In summary, we have developed and validated a QA technique that can simultaneously verify the gantry angle and delivered MLC fluence for VMAT treatment.This technique is efficient and its accuracy is comparable to other QA methods.

  11. TH-AB-201-12: Using Machine Log-Files for Treatment Planning and Delivery QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stanhope, C; Liang, J; Drake, D

    2016-06-15

    Purpose: To determine the segment reduction and dose resolution necessary for machine log-files to effectively replace current phantom-based patient-specific quality assurance, while minimizing computational cost. Methods: Elekta’s Log File Convertor R3.2 records linac delivery parameters (dose rate, gantry angle, leaf position) every 40ms. Five VMAT plans [4 H&N, 1 Pulsed Brain] comprised of 2 arcs each were delivered on the ArcCHECK phantom. Log-files were reconstructed in Pinnacle on the phantom geometry using 1/2/3/4° control point spacing and 2/3/4mm dose grid resolution. Reconstruction effectiveness was quantified by comparing 2%/2mm gamma passing rates of the original and log-file plans. Modulation complexity scoresmore » (MCS) were calculated for each beam to correlate reconstruction accuracy and beam modulation. Percent error in absolute dose for each plan-pair combination (log-file vs. ArcCHECK, original vs. ArcCHECK, log-file vs. original) was calculated for each arc and every diode greater than 10% of the maximum measured dose (per beam). Comparing standard deviations of the three plan-pair distributions, relative noise of the ArcCHECK and log-file systems was elucidated. Results: The original plans exhibit a mean passing rate of 95.1±1.3%. The eight more modulated H&N arcs [MCS=0.088±0.014] and two less modulated brain arcs [MCS=0.291±0.004] yielded log-file pass rates most similar to the original plan when using 1°/2mm [0.05%±1.3% lower] and 2°/3mm [0.35±0.64% higher] log-file reconstructions respectively. Log-file and original plans displayed percent diode dose errors 4.29±6.27% and 3.61±6.57% higher than measurement. Excluding the phantom eliminates diode miscalibration and setup errors; log-file dose errors were 0.72±3.06% higher than the original plans – significantly less noisy. Conclusion: For log-file reconstructed VMAT arcs, 1° control point spacing and 2mm dose resolution is recommended, however, less modulated arcs may

  12. Boiling incipience and convective boiling of neon and nitrogen

    NASA Technical Reports Server (NTRS)

    Papell, S. S.; Hendricks, R. C.

    1977-01-01

    Forced convection and subcooled boiling heat transfer data for liquid nitrogen and liquid neon were obtained in support of a design study for a 30 tesla cryomagnet cooled by forced convection of liquid neon. This design precludes nucleate boiling in the flow channels as they are too small to handle vapor flow. Consequently, it was necessary to determine boiling incipience under the operating conditions of the magnet system. The cryogen data obtained over a range of system pressures, fluid flow rates, and applied heat fluxes were used to develop correlations for predicting boiling incipience and convective boiling heat transfer coefficients in uniformly heated flow channels. The accuracy of the correlating equations was then evaluated. A technique was also developed to calculate the position of boiling incipience in a uniformly heated flow channel. Comparisons made with the experimental data showed a prediction accuracy of plus or minus 15 percent

  13. SU-D-BRC-02: Application of Six Sigma Approach to Improve the Efficiency of Patient-Specific QA in Proton Therapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    LAH, J; Shin, D; Manger, R

    Purpose: To show how the Six Sigma DMAIC (Define-Measure-Analyze-Improve-Control) can be used for improving and optimizing the efficiency of patient-specific QA process by designing site-specific range tolerances. Methods: The Six Sigma tools (process flow diagram, cause and effect, capability analysis, Pareto chart, and control chart) were utilized to determine the steps that need focus for improving the patient-specific QA process. The patient-specific range QA plans were selected according to 7 treatment site groups, a total of 1437 cases. The process capability index, Cpm was used to guide the tolerance design of patient site-specific range. We also analyzed the financial impactmore » of this project. Results: Our results suggested that the patient range measurements were non-capable at the current tolerance level of ±1 mm in clinical proton plans. The optimized tolerances were calculated for treatment sites. Control charts for the patient QA time were constructed to compare QA time before and after the new tolerances were implemented. It is found that overall processing time was decreased by 24.3% after establishing new site-specific range tolerances. The QA failure for whole process in proton therapy would lead up to a 46% increase in total cost. This result can also predict how costs are affected by changes in adopting the tolerance design. Conclusion: We often believe that the quality and performance of proton therapy can easily be improved by merely tightening some or all of its tolerance requirements. This can become costly, however, and it is not necessarily a guarantee of better performance. The tolerance design is not a task to be undertaken without careful thought. The Six Sigma DMAIC can be used to improve the QA process by setting optimized tolerances. When tolerance design is optimized, the quality is reasonably balanced with time and cost demands.« less

  14. SU-E-CAMPUS-T-04: Statistical Process Control for Patient-Specific QA in Proton Beams

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    LAH, J; SHIN, D; Kim, G

    Purpose: To evaluate and improve the reliability of proton QA process, to provide an optimal customized level using the statistical process control (SPC) methodology. The aim is then to suggest the suitable guidelines for patient-specific QA process. Methods: We investigated the constancy of the dose output and range to see whether it was within the tolerance level of daily QA process. This study analyzed the difference between the measured and calculated ranges along the central axis to suggest the suitable guidelines for patient-specific QA in proton beam by using process capability indices. In this study, patient QA plans were classifiedmore » into 6 treatment sites: head and neck (41 cases), spinal cord (29 cases), lung (28 cases), liver (30 cases), pancreas (26 cases), and prostate (24 cases). Results: The deviations for the dose output and range of daily QA process were ±0.84% and ±019%, respectively. Our results show that the patient-specific range measurements are capable at a specification limit of ±2% in all treatment sites except spinal cord cases. In spinal cord cases, comparison of process capability indices (Cp, Cpm, Cpk ≥1, but Cpmk ≤1) indicated that the process is capable, but not centered, the process mean deviates from its target value. The UCL (upper control limit), CL (center line) and LCL (lower control limit) for spinal cord cases were 1.37%, −0.27% and −1.89%, respectively. On the other hands, the range differences in prostate cases were good agreement between calculated and measured values. The UCL, CL and LCL for prostate cases were 0.57%, −0.11% and −0.78%, respectively. Conclusion: SPC methodology has potential as a useful tool to customize an optimal tolerance levels and to suggest the suitable guidelines for patient-specific QA in clinical proton beam.« less

  15. Clinical validation of an in-house EPID dosimetry system for IMRT QA at the Prince of Wales Hospital

    NASA Astrophysics Data System (ADS)

    Tyler, M.; Vial, P.; Metcalfe, P.; Downes, S.

    2013-06-01

    In this study a simple method using standard flood-field corrected Electronic Portal Imaging Device (EPID) images for routine Intensity Modulated Radiation Therapy (IMRT) Quality Assurance (QA) was investigated. The EPID QA system was designed and tested on a Siemens Oncor Impression linear accelerator with an OptiVue 1000ST EPID panel (Siemens Medical Solutions USA, Inc, USA) and an Elekta Axesse linear accelerator with an iViewGT EPID (Elekta AB, Sweden) for 6 and 10 MV IMRT fields with Step-and-Shoot and dynamic-MLC delivery. Two different planning systems were used for patient IMRT field generation for comparison with the measured EPID fluences. All measured IMRT plans had >95% agreement to the planning fluences (using 3 cGy / 3 mm Gamma Criteria) and were comparable to the pass-rates calculated using a 2-D diode array dosimeter.

  16. Monte Carlo based, patient-specific RapidArc QA using Linac log files.

    PubMed

    Teke, Tony; Bergman, Alanah M; Kwa, William; Gill, Bradford; Duzenli, Cheryl; Popescu, I Antoniu

    2010-01-01

    A Monte Carlo (MC) based QA process to validate the dynamic beam delivery accuracy for Varian RapidArc (Varian Medical Systems, Palo Alto, CA) using Linac delivery log files (DynaLog) is presented. Using DynaLog file analysis and MC simulations, the goal of this article is to (a) confirm that adequate sampling is used in the RapidArc optimization algorithm (177 static gantry angles) and (b) to assess the physical machine performance [gantry angle and monitor unit (MU) delivery accuracy]. Ten clinically acceptable RapidArc treatment plans were generated for various tumor sites and delivered to a water-equivalent cylindrical phantom on the treatment unit. Three Monte Carlo simulations were performed to calculate dose to the CT phantom image set: (a) One using a series of static gantry angles defined by 177 control points with treatment planning system (TPS) MLC control files (planning files), (b) one using continuous gantry rotation with TPS generated MLC control files, and (c) one using continuous gantry rotation with actual Linac delivery log files. Monte Carlo simulated dose distributions are compared to both ionization chamber point measurements and with RapidArc TPS calculated doses. The 3D dose distributions were compared using a 3D gamma-factor analysis, employing a 3%/3 mm distance-to-agreement criterion. The dose difference between MC simulations, TPS, and ionization chamber point measurements was less than 2.1%. For all plans, the MC calculated 3D dose distributions agreed well with the TPS calculated doses (gamma-factor values were less than 1 for more than 95% of the points considered). Machine performance QA was supplemented with an extensive DynaLog file analysis. A DynaLog file analysis showed that leaf position errors were less than 1 mm for 94% of the time and there were no leaf errors greater than 2.5 mm. The mean standard deviation in MU and gantry angle were 0.052 MU and 0.355 degrees, respectively, for the ten cases analyzed. The accuracy and

  17. SU-F-T-260: Using Portal Image Device for Pre-Treatment QA in Volumetric Modulated Arc Plans with Flattening Filter Free (FFF) Beams

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Qu, H; Qi, P; Yu, N

    Purpose: To implement and validate a method of using electronic portal image device (EPID) for pre-treatment quality assurance (QA) of volumetric modulated arc therapy (VMAT) plans using flattering filter free (FFF) beams for stereotactic body radiotherapy (SBRT). Methods: On Varian Edge with 6MV FFF beam, open field (from 2×2 cm to 20×20 cm) EPID images were acquired with 200 monitor unit (MU) at the image device to radiation source distance of 150cm. With 10×10 open field and calibration unit (CU) provided by vendor to EPID image pixel, a dose conversion factor was determined by dividing the center dose calculated frommore » the treatment planning system (TPS) to the corresponding CU readout on the image. Water phantom measured beam profile and the output factors for various field sizes were further correlated to those of EPID images. The dose conversion factor and correction factors were then used for converting the portal images to the planner dose distributions of clinical fields. A total of 28 VMAT fields of 14 SBRT plans (8 lung, 2 prostate, 2 liver and 2 spine) were measured. With 10% low threshold cutoff, the delivered dose distributions were compared to the reference doses calculated in water phantom from the TPS. A gamma index analysis was performed for the comparison in percentage dose difference/distance-to-agreement specifications. Results: The EPID device has a linear response to the open fields with increasing MU. For the clinical fields, the gamma indices between the converted EPID dose distributions and the TPS calculated 2D dose distributions were 98.7%±1.1%, 94.0%±3.4% and 70.3%±7.7% for the criteria of 3%/3mm, 2%/2mm and 1%/1mm, respectively. Conclusion: Using a portal image device, a high resolution and high accuracy portal dosimerty was achieved for pre-treatment QA verification for SBRT VMAT plans with FFF beams.« less

  18. Analysis of Qa-1bPeptide Binding Specificity and the Capacity of Cd94/Nkg2a to Discriminate between Qa-1–Peptide Complexes

    PubMed Central

    Kraft, Jennifer R.; Vance, Russell E.; Pohl, Jan; Martin, Amy M.; Raulet, David H.; Jensen, Peter E.

    2000-01-01

    The major histocompatibility complex class Ib protein, Qa-1b, serves as a ligand for murine CD94/NKG2A natural killer (NK) cell inhibitory receptors. The Qa-1b peptide-binding site is predominantly occupied by a single nonameric peptide, Qa-1 determinant modifier (Qdm), derived from the leader sequence of H-2D and L molecules. Five anchor residues were identified in this study by measuring the peptide-binding affinities of substituted Qdm peptides in experiments with purified recombinant Qa-1b. A candidate peptide-binding motif was determined by sequence analysis of peptides eluted from Qa-1 that had been folded in the presence of random peptide libraries or pools of Qdm derivatives randomized at specific anchor positions. The results indicate that Qa-1b can bind a diverse repertoire of peptides but that Qdm has an optimal primary structure for binding Qa-1b. Flow cytometry experiments with Qa-1b tetramers and NK target cell lysis assays demonstrated that CD94/NKG2A discriminates between Qa-1b complexes containing peptides with substitutions at nonanchor positions P4, P5, or P8. Our findings suggest that it may be difficult for viruses to generate decoy peptides that mimic Qdm and raise the possibility that competitive replacement of Qdm with other peptides may provide a novel mechanism for activation of NK cells. PMID:10974028

  19. Boiling incipience and convective boiling of neon and nitrogen

    NASA Technical Reports Server (NTRS)

    Papell, S. S.; Hendricks, R. C.

    1977-01-01

    Forced convection and subcooled boiling heat transfer data for liquid nitrogen and liquid neon were obtained in support of a design study for a 30 tesla cryomagnet cooled by forced convection of liquid neon. The cryogen data obtained over a range of system pressures, fluid flow rates, and applied heat fluxes were used to develop correlations for predicting boiling incipience and convective boiling heat transfer coefficients in uniformly heated flow channels. The accuracy of the correlating equations was then evaluated. A technique was also developed to calculate the position of boiling incipience in a uniformly heated flow channel. Comparisons made with the experimental data showed a prediction accuracy of + or - 15 percent.

  20. Transforming an EPA QA/R-2 quality management plan into an ISO 9002 quality management system.

    PubMed

    Kell, R A; Hedin, C M; Kassakhian, G H; Reynolds, E S

    2001-01-01

    The Environmental Protection Agency's (EPA) Office of Emergency and Remedial Response (OERR) requires environmental data of known quality to support Superfund hazardous waste site projects. The Quality Assurance Technical Support (QATS) Program is operated by Shaw Environmental and Infrastructure, Inc. to provide EPA's Analytical Operations Center (AOC) with performance evaluation samples, reference materials, on-site laboratory auditing capabilities, data audits (including electronic media data audits), methods development, and other support services. The new QATS contract awarded in November 2000 required that the QATS Program become ISO 9000 certified. In a first for an EPA contractor, the QATS staff and management successfully transformed EPA's QA/R-2 type Quality Management Plan into a Quality Management System (QMS) that complies with the requirements of the internationally recognized ISO 9002 standard and achieved certification in the United States, Canada, and throughout Europe. The presentation describes how quality system elements of ISO 9002 were implemented on an already existing quality system. The psychological and organizational challenges of the culture change in QATS' day-to-day operations will be discussed for the benefit of other ISO 9000 aspirants.

  1. SU-E-J-78: Adaptive Planning Workflow in a Pencil Beam Scanning Proton Therapy Center

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Blakey, M; Price, S; Robison, B

    Purpose: The susceptibility of proton therapy to changes in patient setup and anatomy necessitates an adaptive planning process. With the right planning tools and clinical workflow, an adaptive plan can be created in a timely manner without adding significant workload to the treatment planning staff. Methods: In our center, a weekly QA CT is performed on most patients to assess setup, anatomy change, and tumor response. The QA CT is fused to the treatment planning CT, the contours are transferred via deformable registration, and the plan dose is recalculated on the QA CT. A physicist assesses the dose distribution, andmore » an adaptive plan is requested based on tumor coverage or OAR dose changes. After the physician confirms or alters the deformed contours, a dosimetrist develops an adaptive plan using our TPS adaptation module. The plan is assessed for robustness and is then reviewed by the physician. Patient QA is performed within three days following the first adapted treatment. Results: Of the patients who received QA CTs, 19% required at least one adaptive plan (18.5% H&N, 18.5% brain, 11.1% breast, 14.8% chestwall, 14.8% lung, 18.5% pelvis and 3.8% abdomen). Of these patients, 14% went on a break, while the remainder was treated with the previous plan during the re-planning process. Adaptive plans were performed based on tumor shrinkage, anatomy change or positioning uncertainties for 37.9%, 44.8%, and 17.3% of the patients, respectively. On average, 3 full days are required between the QA CT and the first adapted plan treatment. Conclusion: Adaptive planning is a crucial component of proton therapy and should be applied to any site when the QA CT shows significant deviation from the plan. With an efficient workflow, an adaptive plan can be applied without delaying patient treatment or burdening the dosimetry and medical physics team.« less

  2. Secondary pool boiling effects

    NASA Astrophysics Data System (ADS)

    Kruse, C.; Tsubaki, A.; Zuhlke, C.; Anderson, T.; Alexander, D.; Gogos, G.; Ndao, S.

    2016-02-01

    A pool boiling phenomenon referred to as secondary boiling effects is discussed. Based on the experimental trends, a mechanism is proposed that identifies the parameters that lead to this phenomenon. Secondary boiling effects refer to a distinct decrease in the wall superheat temperature near the critical heat flux due to a significant increase in the heat transfer coefficient. Recent pool boiling heat transfer experiments using femtosecond laser processed Inconel, stainless steel, and copper multiscale surfaces consistently displayed secondary boiling effects, which were found to be a result of both temperature drop along the microstructures and nucleation characteristic length scales. The temperature drop is a function of microstructure height and thermal conductivity. An increased microstructure height and a decreased thermal conductivity result in a significant temperature drop along the microstructures. This temperature drop becomes more pronounced at higher heat fluxes and along with the right nucleation characteristic length scales results in a change of the boiling dynamics. Nucleation spreads from the bottom of the microstructure valleys to the top of the microstructures, resulting in a decreased surface superheat with an increasing heat flux. This decrease in the wall superheat at higher heat fluxes is reflected by a "hook back" of the traditional boiling curve and is thus referred to as secondary boiling effects. In addition, a boiling hysteresis during increasing and decreasing heat flux develops due to the secondary boiling effects. This hysteresis further validates the existence of secondary boiling effects.

  3. A novel approach to EPID-based 3D volumetric dosimetry for IMRT and VMAT QA

    NASA Astrophysics Data System (ADS)

    Alhazmi, Abdulaziz; Gianoli, Chiara; Neppl, Sebastian; Martins, Juliana; Veloza, Stella; Podesta, Mark; Verhaegen, Frank; Reiner, Michael; Belka, Claus; Parodi, Katia

    2018-06-01

    Intensity modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT) are relatively complex treatment delivery techniques and require quality assurance (QA) procedures. Pre-treatment dosimetric verification represents a fundamental QA procedure in daily clinical routine in radiation therapy. The purpose of this study is to develop an EPID-based approach to reconstruct a 3D dose distribution as imparted to a virtual cylindrical water phantom to be used for plan-specific pre-treatment dosimetric verification for IMRT and VMAT plans. For each depth, the planar 2D dose distributions acquired in air were back-projected and convolved by depth-specific scatter and attenuation kernels. The kernels were obtained by making use of scatter and attenuation models to iteratively estimate the parameters from a set of reference measurements. The derived parameters served as a look-up table for reconstruction of arbitrary measurements. The summation of the reconstructed 3D dose distributions resulted in the integrated 3D dose distribution of the treatment delivery. The accuracy of the proposed approach was validated in clinical IMRT and VMAT plans by means of gamma evaluation, comparing the reconstructed 3D dose distributions with Octavius measurement. The comparison was carried out using (3%, 3 mm) criteria scoring 99% and 96% passing rates for IMRT and VMAT, respectively. An accuracy comparable to the one of the commercial device for 3D volumetric dosimetry was demonstrated. In addition, five IMRT and five VMAT were validated against the 3D dose calculation performed by the TPS in a water phantom using the same passing rate criteria. The median passing rates within the ten treatment plans was 97.3%, whereas the lowest was 95%. Besides, the reconstructed 3D distribution is obtained without predictions relying on forward dose calculation and without external phantom or dosimetric devices. Thus, the approach provides a fully automated, fast and easy QA

  4. SU-F-T-227: A Comprehensive Patient Specific, Structure Specific, Pre-Treatment 3D QA Protocol for IMRT, SBRT and VMAT - Clinical Experience

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gueorguiev, G; Cotter, C; Young, M

    2016-06-15

    Purpose: To present a 3D QA method and clinical results for 550 patients. Methods: Five hundred and fifty patient treatment deliveries (400 IMRT, 75 SBRT and 75 VMAT) from various treatment sites, planned on Raystation treatment planning system (TPS), were measured on three beam-matched Elekta linear accelerators using IBA’s COMPASS system. The difference between TPS computed and delivered dose was evaluated in 3D by applying three statistical parameters to each structure of interest: absolute average dose difference (AADD, 6% allowed difference), absolute dose difference greater than 6% (ADD6, 4% structure volume allowed to fail) and 3D gamma test (3%/3mm DTA,more » 4% structure volume allowed to fail). If the allowed value was not met for a given structure, manual review was performed. The review consisted of overlaying dose difference or gamma results with the patient CT, scrolling through the slices. For QA to pass, areas of high dose difference or gamma must be small and not on consecutive slices. For AADD to manually pass QA, the average dose difference in cGy must be less than 50cGy. The QA protocol also includes DVH analysis based on QUANTEC and TG-101 recommended dose constraints. Results: Figures 1–3 show the results for the three parameters per treatment modality. Manual review was performed on 67 deliveries (27 IMRT, 22 SBRT and 18 VMAT), for which all passed QA. Results show that statistical parameter AADD may be overly sensitive for structures receiving low dose, especially for the SBRT deliveries (Fig.1). The TPS computed and measured DVH values were in excellent agreement and with minimum difference. Conclusion: Applying DVH analysis and different statistical parameters to any structure of interest, as part of the 3D QA protocol, provides a comprehensive treatment plan evaluation. Author G. Gueorguiev discloses receiving travel and research funding from IBA for unrelated to this project work. Author B. Crawford discloses receiving travel

  5. Deep nets vs expert designed features in medical physics: An IMRT QA case study.

    PubMed

    Interian, Yannet; Rideout, Vincent; Kearney, Vasant P; Gennatas, Efstathios; Morin, Olivier; Cheung, Joey; Solberg, Timothy; Valdes, Gilmer

    2018-03-30

    The purpose of this study was to compare the performance of Deep Neural Networks against a technique designed by domain experts in the prediction of gamma passing rates for Intensity Modulated Radiation Therapy Quality Assurance (IMRT QA). A total of 498 IMRT plans across all treatment sites were planned in Eclipse version 11 and delivered using a dynamic sliding window technique on Clinac iX or TrueBeam Linacs. Measurements were performed using a commercial 2D diode array, and passing rates for 3%/3 mm local dose/distance-to-agreement (DTA) were recorded. Separately, fluence maps calculated for each plan were used as inputs to a convolution neural network (CNN). The CNNs were trained to predict IMRT QA gamma passing rates using TensorFlow and Keras. A set of model architectures, inspired by the convolutional blocks of the VGG-16 ImageNet model, were constructed and implemented. Synthetic data, created by rotating and translating the fluence maps during training, was created to boost the performance of the CNNs. Dropout, batch normalization, and data augmentation were utilized to help train the model. The performance of the CNNs was compared to a generalized Poisson regression model, previously developed for this application, which used 78 expert designed features. Deep Neural Networks without domain knowledge achieved comparable performance to a baseline system designed by domain experts in the prediction of 3%/3 mm Local gamma passing rates. An ensemble of neural nets resulted in a mean absolute error (MAE) of 0.70 ± 0.05 and the domain expert model resulted in a 0.74 ± 0.06. Convolutional neural networks (CNNs) with transfer learning can predict IMRT QA passing rates by automatically designing features from the fluence maps without human expert supervision. Predictions from CNNs are comparable to a system carefully designed by physicist experts. © 2018 American Association of Physicists in Medicine.

  6. When water does not boil at the boiling point.

    PubMed

    Chang, Hasok

    2007-03-01

    Every schoolchild learns that, under standard pressure, pure water always boils at 100 degrees C. Except that it does not. By the late 18th century, pioneering scientists had already discovered great variations in the boiling temperature of water under fixed pressure. So, why have most of us been taught that the boiling point of water is constant? And, if it is not constant, how can it be used as a 'fixed point' for the calibration of thermometers? History of science has the answers.

  7. 222-S Laboratory Quality Assurance Plan. Revision 1

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Meznarich, H.K.

    1995-07-31

    This Quality Assurance Plan provides,quality assurance (QA) guidance, regulatory QA requirements (e.g., 10 CFR 830.120), and quality control (QC) specifications for analytical service. This document follows the U.S Department of Energy (DOE) issued Hanford Analytical Services Quality Assurance Plan (HASQAP). In addition, this document meets the objectives of the Quality Assurance Program provided in the WHC-CM-4-2, Section 2.1. Quality assurance elements required in the Guidelines and Specifications for Preparing Quality Assurance Program Plans (QAMS-004) and Interim Guidelines and Specifications for Preparing Quality Assurance Project Plans (QAMS-005) from the US Environmental Protection Agency (EPA) are covered throughout this document. A qualitymore » assurance index is provided in the Appendix A. This document also provides and/or identifies the procedural information that governs laboratory operations. The personnel of the 222-S Laboratory and the Standards Laboratory including managers, analysts, QA/QC staff, auditors, and support staff shall use this document as guidance and instructions for their operational and quality assurance activities. Other organizations that conduct activities described in this document for the 222-S Laboratory shall follow this QA/QC document.« less

  8. SU-E-CAMPUS-T-05: Validation of High-Resolution 3D Patient QA for Proton Pencil Beam Scanning and IMPT by Polymer Gel Dosimetry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cardin, A; Avery, S; Ding, X

    2014-06-15

    Purpose: Validation of high-resolution 3D patient QA for proton pencil beam scanning and IMPT by polymer gel dosimetry. Methods: Four BANG3Pro polymer gel dosimeters (manufactured by MGS Research Inc, Madison, CT) were used for patient QA at the Robert's Proton Therapy Center (RPTC, Philadelphia, PA). All dosimeters were sealed in identical thin-wall Pyrex glass spheres. Each dosimeter contained a set of markers for 3D registration purposes. The dosimeters were mounted in a consistent and reproducible manner using a custom build holder. Two proton pencil beam scanning plans were designed using Varian Eclipse™ treatment planning system: 1) A two-field intensity modulatedmore » proton therapy (IMPT) plan and 2) one single field uniform dose (SFUD) plan. The IMPT fields were evaluated as a composite plan and individual fields, the SFUD plan was delivered as a single field plan.Laser CT scanning was performed using the manufacturer's OCTOPUS-IQ axial transmission laser CT scanner using a 1 mm slice thickness. 3D registration, analysis, and OD/cm to absorbed dose calibrations were perfomed using DICOM RT-Dose and CT files, and software developed by the manufacturer. 3D delta index, a metric equivalent to the gamma tool, was used for dose comparison. Results: Very good agreement with single IMPT fields and with SFUD was obtained. Composite IMPT fields had a less satisfactory agreement. The single fields had 3D delta index passing rates (3% dose difference, 3 mm DTA) of 98.98% and 94.91%. The composite 3D delta index passing rate was 80.80%. The SFUD passing rate was 93.77%. Required shifts of the dose distributions were less than 4 mm. Conclusion: A formulation of the BANG3Pro polymer gel dosimeter, suitable for 3D QA of proton patient plans is established and validated. Likewise, the mailed QA analysis service provided by the manufacturer is a practical option when required resources are unavailable. We fully disclose that the subject of this research regards a

  9. Lattice Boltzmann modeling of boiling heat transfer: The boiling curve and the effects of wettability

    DOE PAGES

    Li, Q.; Kang, Q. J.; Francois, M. M.; ...

    2015-03-03

    A hybrid thermal lattice Boltzmann (LB) model is presented to simulate thermal multiphase flows with phase change based on an improved pseudopotential LB approach (Li et al., 2013). The present model does not suffer from the spurious term caused by the forcing-term effect, which was encountered in some previous thermal LB models for liquid–vapor phase change. Using the model, the liquid–vapor boiling process is simulated. The boiling curve together with the three boiling stages (nucleate boiling, transition boiling, and film boiling) is numerically reproduced in the LB community for the first time. The numerical results show that the basic featuresmore » and the fundamental characteristics of boiling heat transfer are well captured, such as the severe fluctuation of transient heat flux in the transition boiling and the feature that the maximum heat transfer coefficient lies at a lower wall superheat than that of the maximum heat flux. Moreover, the effects of the heating surface wettability on boiling heat transfer are investigated. It is found that an increase in contact angle promotes the onset of boiling but reduces the critical heat flux, and makes the boiling process enter into the film boiling regime at a lower wall superheat, which is consistent with the findings from experimental studies.« less

  10. Zero boil-off system testing

    NASA Astrophysics Data System (ADS)

    Plachta, D. W.; Johnson, W. L.; Feller, J. R.

    2016-03-01

    Cryogenic propellants such as liquid hydrogen (LH2) and liquid oxygen (LO2) are a part of NASA's future space exploration plans due to their high specific impulse for rocket motors of upper stages. However, the low storage temperatures of LH2 and LO2 cause substantial boil-off losses for long duration missions. These losses can be eliminated by incorporating high performance cryocooler technology to intercept heat load to the propellant tanks and modulating the cryocooler temperature to control tank pressure. The technology being developed by NASA is the reverse turbo-Brayton cycle cryocooler and its integration to the propellant tank through a distributed cooling tubing network coupled to the tank wall. This configuration was recently tested at NASA Glenn Research Center in a vacuum chamber and cryoshroud that simulated the essential thermal aspects of low Earth orbit, its vacuum and temperature. This test series established that the active cooling system integrated with the propellant tank eliminated boil-off and robustly controlled tank pressure.

  11. Zero Boil-Off System Testing

    NASA Technical Reports Server (NTRS)

    Plachta, D. W.; Johnson, W. L.; Feller, J. R.

    2015-01-01

    Cryogenic propellants such as liquid hydrogen (LH2) and liquid oxygen (LO2) are a part of NASA's future space exploration plans due to their high specific impulse for rocket motors of upper stages. However, the low storage temperatures of LH2 and LO2 cause substantial boil-off losses for long duration missions. These losses can be eliminated by incorporating high performance cryocooler technology to intercept heat load to the propellant tanks and modulating the cryocooler temperature to control tank pressure. The technology being developed by NASA is the reverse turbo-Brayton cycle cryocooler and its integration to the propellant tank through a distributed cooling tubing network coupled to the tank wall. This configuration was recently tested at NASA Glenn Research Center in a vacuum chamber and cryoshroud that simulated the essential thermal aspects of low Earth orbit, its vacuum and temperature. This test series established that the active cooling system integrated with the propellant tank eliminated boil-off and robustly controlled tank pressure.

  12. SU-E-T-100: Designing a QA Tool for Enhance Dynamic Wedges Based On Dynalog Files

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yousuf, A; Hussain, A

    2014-06-01

    Purpose: A robust quality assurance (QA) program for computer controlled enhanced dynamic wedge (EDW) has been designed and tested. Calculations to perform such QA test is based upon the EDW dynamic log files generated during dose delivery. Methods: Varian record and verify system generates dynamic log (dynalog) files during dynamic dose delivery. The system generated dynalog files contain information such as date and time of treatment, energy, monitor units, wedge orientation, and type of treatment. It also contains the expected calculated segmented treatment tables (STT) and the actual delivered STT for the treatment delivery as a verification record. These filesmore » can be used to assess the integrity and precision of the treatment plan delivery. The plans were delivered with a 6 MV beam from a Varian linear accelerator. For available EDW angles (10°, 15°, 20°, 25°, 30°, 45°, and 60°) Varian STT values were used to manually calculate monitor units for each segment. It can also be used to calculate the EDW factors. Independent verification of fractional MUs per segment was performed against those generated from dynalog files. The EDW factors used to calculate MUs in TPS were dosimetrically verified in solid water phantom with semiflex chamber on central axis. Results: EDW factors were generated from the STT provided by Varian and verified against practical measurements. The measurements were in agreement of the order of 1 % to the calculated EDW data. Variation between the MUs per segment obtained from dynalog files and those manually calculated was found to be less than 2%. Conclusion: An efficient and easy tool to perform routine QA procedure of EDW is suggested. The method can be easily implemented in any institution without a need for expensive QA equipment. An error of the order of ≥2% can be easily detected.« less

  13. Investigating ion recombination effects in a liquid-filled ionization chamber array used for IMRT QA measurements

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Knill, Cory, E-mail: knillcor@gmail.com; Snyder, Michael; Rakowski, Joseph T.

    Purpose: PTW’s Octavius 1000 SRS array performs IMRT quality assurance (QA) measurements with liquid-filled ionization chambers (LICs) to allow closer detector spacing and higher resolution, compared to air-filled QA devices. However, reduced ion mobility in LICs relative to air leads to increased ion recombination effects and reduced collection efficiencies that are dependent on Linac pulse frequency and pulse dose. These pulse parameters are variable during an IMRT delivery, which affects QA results. In this study, (1) 1000 SRS collection efficiencies were measured as a function of pulse frequency and pulse dose, (2) two methods were developed to correct changes inmore » collection efficiencies during IMRT QA measurements, and the effects of these corrections on QA pass rates were compared. Methods: To obtain collection efficiencies, the OCTAVIUS 1000 SRS was used to measure open fields of varying pulse frequency, pulse dose, and beam energy with results normalized to air-filled chamber measurements. Changes in ratios of 1000 SRS to chamber measured dose were attributed to changing collection efficiencies, which were then correlated to pulse parameters using regression analysis. The usefulness of the derived corrections was then evaluated using 6 MV and 10FFF SBRT RapidArc plans delivered to the OCTAVIUS 4D system using a TrueBeam (Varian Medical Systems) linear accelerator equipped with a high definition multileaf collimator. For the first correction, MATLAB software was developed that calculates pulse frequency and pulse dose for each detector, using measurement and DICOM RT Plan files. Pulse information is converted to collection efficiency, and measurements are corrected by multiplying detector dose by ratios of calibration to measured collection efficiencies. For the second correction the MU/min in the daily 1000 SRS calibration was chosen to match the average MU/min of the volumetric modulated arc therapy plan. Effects of the two corrections on QA

  14. MO-FG-202-04: Gantry-Resolved Linac QA for VMAT: A Comprehensive and Efficient System Using An Electronic Portal Imaging Device

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zwan, B J; University of Newcastle, Newcastle, NSW; Barnes, M

    2016-06-15

    Purpose: To automate gantry-resolved linear accelerator (linac) quality assurance (QA) for volumetric modulated arc therapy (VMAT) using an electronic portal imaging device (EPID). Methods: A QA system for VMAT was developed that uses an EPID, frame-grabber assembly and in-house developed image processing software. The system relies solely on the analysis of EPID image frames acquired without the presence of a phantom. Images were acquired at 8.41 frames per second using a frame grabber and ancillary acquisition computer. Each image frame was tagged with a gantry angle from the linac’s on-board gantry angle encoder. Arc-dynamic QA plans were designed to assessmore » the performance of each individual linac component during VMAT. By analysing each image frame acquired during the QA deliveries the following eight machine performance characteristics were measured as a function of gantry angle: MLC positional accuracy, MLC speed constancy, MLC acceleration constancy, MLC-gantry synchronisation, beam profile constancy, dose rate constancy, gantry speed constancy, dose-gantry angle synchronisation and mechanical sag. All tests were performed on a Varian iX linear accelerator equipped with a 120 leaf Millennium MLC and an aS1000 EPID (Varian Medical Systems, Palo Alto, CA, USA). Results: Machine performance parameters were measured as a function of gantry angle using EPID imaging and compared to machine log files and the treatment plan. Data acquisition is currently underway at 3 centres, incorporating 7 treatment units, at 2 weekly measurement intervals. Conclusion: The proposed system can be applied for streamlined linac QA and commissioning for VMAT. The set of test plans developed can be used to assess the performance of each individual components of the treatment machine during VMAT deliveries as a function of gantry angle. The methodology does not require the setup of any additional phantom or measurement equipment and the analysis is fully automated to allow

  15. National Ignition Facility quality assurance program plan revision 2

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wolfe, C R

    1998-06-01

    NIF Project activities will be conducted in a manner consistent with the guidance and direction of the DOE Order on Quality Assurance (414.1), the LLNL QA Program, and the Laser Directorate QA Plan. Quality assurance criteria will be applied in a graded manner to achieve a balance between the rigor of application of QA measures and the scale, cost, and complexity of the work involved. Accountability for quality is everyone's, extending from the Project Manager through established lines of authority to all Project personnel, who are responsible for the requisite quality of their own work. The NLF QA Program willmore » be implemented by personnel conducting their activities to meet requirements and expectations, according to established plans and procedures that reflect the way business is to be conducted on the Project.« less

  16. Region 7 Quality Management Plan

    EPA Pesticide Factsheets

    To document adherence to EPA Order 5360.1 A2, EPA requires each organizational unitto develop a quality management plan per the specifications in EPA Requirements for QualityManagement Plans, EPA QA R-2.

  17. owl-qa | Informatics Technology for Cancer Research (ITCR)

    Cancer.gov

    owl-qa is an OWL-based QA tool for cancer study CDEs. The tool uses the combination of the NCI Thesaurus and additional disjointness axioms to detect potential errors and duplications in the data element definitions. The tool comprises three modules: Data Integration and Services Module; Compositional Expression Transformation Module; and OWL-based Quality Assurance Module.

  18. 40 CFR 98.44 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98.44 Section 98.44 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electricity Generation § 98.44 Monitoring and QA/QC...

  19. 40 CFR 98.44 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98.44 Section 98.44 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electricity Generation § 98.44 Monitoring and QA/QC...

  20. 40 CFR 98.44 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98.44 Section 98.44 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electricity Generation § 98.44 Monitoring and QA/QC...

  1. 40 CFR 98.44 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98.44 Section 98.44 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electricity Generation § 98.44 Monitoring and QA/QC...

  2. 40 CFR 98.44 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98.44 Section 98.44 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electricity Generation § 98.44 Monitoring and QA/QC...

  3. SU-F-T-587: Quality Assurance of Stereotactic Radiosurgery (SRS) and Stereotactic Body Radiation Therapy (SBRT) for Patient Specific Plans: A Comparison Between MATRIXX and Delta4 QA Devices

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tsai, YC; Lu, SH; Chen, LH

    2016-06-15

    Purpose: Patient-specific quality assurance (QA) is necessary to accurately deliver high dose radiation to the target, especially for stereotactic radiosurgery (SRS) and stereotactic body radiation therapy (SBRT). Unlike previous 2 dimensional (D) array QA devices, Delta{sup 4} can verify the dose delivery in 3D. In this study, the difference between calculated and measured dose distribution was compared with two QA devices (MATRIXX and Delta{sup 4}) to evaluate the delivery accuracy. Methods: Twenty-seven SRS/SBRT plans with VMAT were verified with point-dose and dose-map analysis. We use an ion chamber (A1SL, 0.053cc) for point-dose measurement. For verification of the dose map, themore » differences between the calculated and measured doses were analyzed with a gamma index using MATRIXX and Delta{sup 4} devices. The passing criteria for gamma evaluation were set at 3 mm for distance-to-agreement (DTA) and 3% for dose-difference. A gamma index less than 1 was defined as the verification passing the criteria and satisfying at least 95% of the points. Results: The mean prescribed dose and fraction was 40 ± 14.41 Gy (range: 16–60) and 10 ± 2.35 fractions (range: 1–8), respectively. In point dose analysis, the differences between the calculated and measured doses were all less than 5% (mean: 2.12 ± 1.13%; range: −0.55% to 4.45%). In dose-map analysis, the average passing rates were 99.38 ± 0.96% (range: 95.31–100%) and 100 ± 0.12% (range: 99.5%–100%) for MATRIXX and Delta{sup 4}, respectively. Even using criteria of 2%/2 mm, the passing rate of Delta{sup 4} was still more than 95% (mean: 99 ± 1.08%; range: 95.6%–100%). Conclusion: Both MATRIXX and Delta{sup 4} offer accurate and efficient verification for SRS/SBRT plans. The results measured by MATRIXX and Delta{sup 4} dosimetry systems are similar for SRS/SBRT performed with the VMAT technique.« less

  4. SU-E-T-04: 3D Dose Based Patient Compensator QA Procedure for Proton Radiotherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zou, W; Reyhan, M; Zhang, M

    2015-06-15

    Purpose: In proton double-scattering radiotherapy, compensators are the essential patient specific devices to contour the distal dose distribution to the tumor target. Traditional compensator QA is limited to checking the drilled surface profiles against the plan. In our work, a compensator QA process was established that assess the entire compensator including its internal structure for patient 3D dose verification. Methods: The fabricated patient compensators were CT scanned. Through mathematical image processing and geometric transformations, the CT images of the proton compensator were combined with the patient simulation CT images into a new series of CT images, in which the imagedmore » compensator is placed at the planned location along the corresponding beam line. The new CT images were input into the Eclipse treatment planning system. The original plan was calculated to the combined CT image series without the plan compensator. The newly computed patient 3D dose from the combined patientcompensator images was verified against the original plan dose. Test plans include the compensators with defects intentionally created inside the fabricated compensators. Results: The calculated 3D dose with the combined compensator and patient CT images reflects the impact of the fabricated compensator to the patient. For the test cases in which no defects were created, the dose distributions were in agreement between our method and the corresponding original plans. For the compensator with the defects, the purposely changed material and a purposely created internal defect were successfully detected while not possible with just the traditional compensator profiles detection methods. Conclusion: We present here a 3D dose verification process to qualify the fabricated proton double-scattering compensator. Such compensator detection process assesses the patient 3D impact of the fabricated compensator surface profile as well as the compensator internal material and structure

  5. Implementation of the validation testing in MPPG 5.a "Commissioning and QA of treatment planning dose calculations-megavoltage photon and electron beams".

    PubMed

    Jacqmin, Dustin J; Bredfeldt, Jeremy S; Frigo, Sean P; Smilowitz, Jennifer B

    2017-01-01

    The AAPM Medical Physics Practice Guideline (MPPG) 5.a provides concise guidance on the commissioning and QA of beam modeling and dose calculation in radiotherapy treatment planning systems. This work discusses the implementation of the validation testing recommended in MPPG 5.a at two institutions. The two institutions worked collaboratively to create a common set of treatment fields and analysis tools to deliver and analyze the validation tests. This included the development of a novel, open-source software tool to compare scanning water tank measurements to 3D DICOM-RT Dose distributions. Dose calculation algorithms in both Pinnacle and Eclipse were tested with MPPG 5.a to validate the modeling of Varian TrueBeam linear accelerators. The validation process resulted in more than 200 water tank scans and more than 50 point measurements per institution, each of which was compared to a dose calculation from the institution's treatment planning system (TPS). Overall, the validation testing recommended in MPPG 5.a took approximately 79 person-hours for a machine with four photon and five electron energies for a single TPS. Of the 79 person-hours, 26 person-hours required time on the machine, and the remainder involved preparation and analysis. The basic photon, electron, and heterogeneity correction tests were evaluated with the tolerances in MPPG 5.a, and the tolerances were met for all tests. The MPPG 5.a evaluation criteria were used to assess the small field and IMRT/VMAT validation tests. Both institutions found the use of MPPG 5.a to be a valuable resource during the commissioning process. The validation testing in MPPG 5.a showed the strengths and limitations of the TPS models. In addition, the data collected during the validation testing is useful for routine QA of the TPS, validation of software upgrades, and commissioning of new algorithms. © 2016 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of

  6. Boiling of the interface between two immiscible liquids below the bulk boiling temperatures of both components.

    PubMed

    Pimenova, Anastasiya V; Goldobin, Denis S

    2014-11-01

    We consider the problem of boiling of the direct contact of two immiscible liquids. An intense vapour formation at such a direct contact is possible below the bulk boiling points of both components, meaning an effective decrease of the boiling temperature of the system. Although the phenomenon is known in science and widely employed in technology, the direct contact boiling process was thoroughly studied (both experimentally and theoretically) only for the case where one of liquids is becoming heated above its bulk boiling point. On the contrary, we address the case where both liquids remain below their bulk boiling points. In this paper we construct the theoretical description of the boiling process and discuss the actualisation of the case we consider for real systems.

  7. 40 CFR 98.64 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98.64 Section 98.64 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Aluminum Production § 98.64 Monitoring and QA/QC requirements...

  8. 40 CFR 98.334 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Zinc Production § 98.334 Monitoring and QA/QC requirements. If..., belt weigh feeders, weighed purchased quantities in shipments or containers, combination of bulk...

  9. 40 CFR 98.334 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Zinc Production § 98.334 Monitoring and QA/QC requirements. If..., belt weigh feeders, weighed purchased quantities in shipments or containers, combination of bulk...

  10. 40 CFR 98.334 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Zinc Production § 98.334 Monitoring and QA/QC requirements. If..., belt weigh feeders, weighed purchased quantities in shipments or containers, combination of bulk...

  11. 40 CFR 98.334 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Zinc Production § 98.334 Monitoring and QA/QC requirements. If..., belt weigh feeders, weighed purchased quantities in shipments or containers, combination of bulk...

  12. 40 CFR 98.64 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98.64 Section 98.64 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Aluminum Production § 98.64 Monitoring and QA/QC requirements...

  13. 40 CFR 98.84 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98.84 Section 98.84 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Cement Production § 98.84 Monitoring and QA/QC requirements...

  14. SU-E-T-11: A Cloud Based CT and LINAC QA Data Management System

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wiersma, R; Grelewicz, Z; Belcher, A

    Purpose: The current status quo of QA data management consists of a mixture of paper-based forms and spreadsheets for recording the results of daily, monthly, and yearly QA tests for both CT scanners and LINACs. Unfortunately, such systems suffer from a host of problems as, (1) records can be easily lost or destroyed, (2) data is difficult to access — one must physically hunt down records, (3) poor or no means of historical data analysis, and (4) no remote monitoring of machine performance off-site. To address these issues, a cloud based QA data management system was developed and implemented. Methods:more » A responsive tablet interface that optimizes clinic workflow with an easy-to-navigate interface accessible from any web browser was implemented in HTML/javascript/CSS to allow user mobility when entering QA data. Automated image QA was performed using a phantom QA kit developed in Python that is applicable to any phantom and is currently being used with the Gammex ACR, Las Vegas, Leeds, and Catphan phantoms for performing automated CT, MV, kV, and CBCT QAs, respectively. A Python based resource management system was used to distribute and manage intensive CPU tasks such as QA phantom image analysis or LaTeX-to-PDF QA report generation to independent process threads or different servers such that website performance is not affected. Results: To date the cloud QA system has performed approximately 185 QA procedures. Approximately 200 QA parameters are being actively tracked by the system on a monthly basis. Electronic access to historical QA parameter information was successful in proactively identifying a Linac CBCT scanner’s performance degradation. Conclusion: A fully comprehensive cloud based QA data management system was successfully implemented for the first time. Potential machine performance issues were proactively identified that would have been otherwise missed by a paper or spreadsheet based QA system.« less

  15. Daily QA of linear accelerators using only EPID and OBI

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sun, Baozhou, E-mail: bsun@radonc.wustl.edu; Goddu, S. Murty; Yaddanapudi, Sridhar

    2015-10-15

    Purpose: As treatment delivery becomes more complex, there is a pressing need for robust quality assurance (QA) tools to improve efficiency and comprehensiveness while simultaneously maintaining high accuracy and sensitivity. This work aims to present the hardware and software tools developed for comprehensive QA of linear accelerator (LINAC) using only electronic portal imaging devices (EPIDs) and kV flat panel detectors. Methods: A daily QA phantom, which includes two orthogonally positioned phantoms for QA of MV-beams and kV onboard imaging (OBI) is suspended from the gantry accessory holder to test both geometric and dosimetric components of a LINAC and an OBI.more » The MV component consists of a 0.5 cm water-equivalent plastic sheet incorporating 11 circular steel plugs for transmission measurements through multiple thicknesses and one resolution plug for MV-image quality testing. The kV-phantom consists of a Leeds phantom (TOR-18 FG phantom supplied by Varian) for testing low and high contrast resolutions. In the developed process, the existing LINAC tools were used to automate daily acquisition of MV and kV images and software tools were developed for simultaneous analysis of these images. A method was developed to derive and evaluate traditional QA parameters from these images [output, flatness, symmetry, uniformity, TPR{sub 20/10}, and positional accuracy of the jaws and multileaf collimators (MLCs)]. The EPID-based daily QA tools were validated by performing measurements on a detuned 6 MV beam to test its effectiveness in detecting errors in output, symmetry, energy, and MLC positions. The developed QA process was clinically commissioned, implemented, and evaluated on a Varian TrueBeam LINAC (Varian Medical System, Palo Alto, CA) over a period of three months. Results: Machine output constancy measured with an EPID (as compared against a calibrated ion-chamber) is shown to be within ±0.5%. Beam symmetry and flatness deviations measured using an EPID

  16. SU-E-T-184: Clinical VMAT QA Practice Using LINAC Delivery Log Files

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnston, H; Jacobson, T; Gu, X

    2015-06-15

    Purpose: To evaluate the accuracy of volumetric modulated arc therapy (VMAT) treatment delivery dose clouds by comparing linac log data to doses measured using an ionization chamber and film. Methods: A commercial IMRT quality assurance (QA) process utilizing a DICOM-RT framework was tested for clinical practice using 30 prostate and 30 head and neck VMAT plans. Delivered 3D VMAT dose distributions were independently checked using a PinPoint ionization chamber and radiographic film in a solid water phantom. DICOM RT coordinates were used to extract the corresponding point and planar doses from 3D log file dose distributions. Point doses were evaluatedmore » by computing the percent error between log file and chamber measured values. A planar dose evaluation was performed for each plan using a 2D gamma analysis with 3% global dose difference and 3 mm isodose point distance criteria. The same analysis was performed to compare treatment planning system (TPS) doses to measured values to establish a baseline assessment of agreement. Results: The mean percent error between log file and ionization chamber dose was 1.0%±2.1% for prostate VMAT plans and −0.2%±1.4% for head and neck plans. The corresponding TPS calculated and measured ionization chamber values agree within 1.7%±1.6%. The average 2D gamma passing rates for the log file comparison to film are 98.8%±1.0% and 96.2%±4.2% for the prostate and head and neck plans, respectively. The corresponding passing rates for the TPS comparison to film are 99.4%±0.5% and 93.9%±5.1%. Overall, the point dose and film data indicate that log file determined doses are in excellent agreement with measured values. Conclusion: Clinical VMAT QA practice using LINAC treatment log files is a fast and reliable method for patient-specific plan evaluation.« less

  17. 40 CFR 98.94 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electronics Manufacturing § 98.94 Monitoring and QA/QC requirements. (a) For calendar year 2011 monitoring, you may follow the provisions in paragraphs (a)(1) through...

  18. 40 CFR 98.94 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electronics Manufacturing § 98.94 Monitoring and QA/QC requirements. (a) For calendar year 2011 monitoring, you may follow the provisions in paragraphs (a)(1) through...

  19. 40 CFR 98.94 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electronics Manufacturing § 98.94 Monitoring and QA/QC requirements. (a) For calendar year 2011 monitoring, you may follow the provisions in paragraphs (a)(1) through...

  20. 40 CFR 98.94 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Electronics Manufacturing § 98.94 Monitoring and QA/QC...-specific heel factors for each container type for each gas used, according to the procedures in paragraphs...

  1. SU-G-TeP4-04: An Automated Monte Carlo Based QA Framework for Pencil Beam Scanning Treatments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shin, J; Jee, K; Clasie, B

    2016-06-15

    Purpose: Prior to treating new PBS field, multiple (three) patient-field-specific QA measurements are performed: two 2D dose distributions at shallow depth (M1) and at the tumor depth (M2) with treatment hardware at zero gantry angle; one 2D dose distribution at iso-center (M3) without patient specific devices at the planned gantry angle. This patient-specific QA could be simplified by the use of MC model. The results of MC model commissioning for a spot-scanning system and the fully automated TOPAS/MC-based QA framework will be presented. Methods: We have developed in-house MC interface to access a TPS (Astroid) database from a computer clustermore » remotely. Once a plan is identified, the interface downloads information for the MC simulations, such as patient images, apertures points, and fluence maps and initiates calculations in both the patient and QA geometries. The resulting calculations are further analyzed to evaluate the TPS dose accuracy and the PBS delivery. Results: The Monte Carlo model of our system was validated within 2.0 % accuracy over the whole range of the dose distribution (proximal/shallow part, as well as target dose part) due to the location of the measurements. The averaged range difference after commissioning was 0.25 mm over entire treatment ranges, e.g., 6.5 cm to 31.6 cm. Conclusion: As M1 depths range typically from 1 cm to 4 cm from the phantom surface, The Monte Carlo model of our system was validated within +− 2.0 % in absolute dose level over a whole treatment range. The averaged range difference after commissioning was 0.25 mm over entire treatment ranges, e.g., 6.5 cm to 31.6 cm. This work was supported by NIH/NCI under CA U19 21239.« less

  2. The use and QA of biologically related models for treatment planning: short report of the TG-166 of the therapy physics committee of the AAPM.

    PubMed

    Allen Li, X; Alber, Markus; Deasy, Joseph O; Jackson, Andrew; Ken Jee, Kyung-Wook; Marks, Lawrence B; Martel, Mary K; Mayo, Charles; Moiseenko, Vitali; Nahum, Alan E; Niemierko, Andrzej; Semenenko, Vladimir A; Yorke, Ellen D

    2012-03-01

    Treatment planning tools that use biologically related models for plan optimization and/or evaluation are being introduced for clinical use. A variety of dose-response models and quantities along with a series of organ-specific model parameters are included in these tools. However, due to various limitations, such as the limitations of models and available model parameters, the incomplete understanding of dose responses, and the inadequate clinical data, the use of biologically based treatment planning system (BBTPS) represents a paradigm shift and can be potentially dangerous. There will be a steep learning curve for most planners. The purpose of this task group is to address some of these relevant issues before the use of BBTPS becomes widely spread. In this report, the authors (1) discuss strategies, limitations, conditions, and cautions for using biologically based models and parameters in clinical treatment planning; (2) demonstrate the practical use of the three most commonly used commercially available BBTPS and potential dosimetric differences between biologically model based and dose-volume based treatment plan optimization and evaluation; (3) identify the desirable features and future directions in developing BBTPS; and (4) provide general guidelines and methodology for the acceptance testing, commissioning, and routine quality assurance (QA) of BBTPS.

  3. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Carbon Dioxide § 98.424 Monitoring and QA/QC... determine quantity in accordance with this paragraph. (i) Reporters that supply CO2 in containers using...

  4. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Carbon Dioxide § 98.424 Monitoring and QA/QC... determine quantity in accordance with this paragraph. (i) Reporters that supply CO2 in containers using...

  5. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 21 2011-07-01 2011-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Carbon Dioxide § 98.424 Monitoring and QA/QC... determine quantity in accordance with this paragraph. (i) Reporters that supply CO2 in containers using...

  6. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Carbon Dioxide § 98.424 Monitoring and QA/QC... determine quantity in accordance with this paragraph. (i) Reporters that supply CO2 in containers using...

  7. SU-F-T-567: Sensitivity and Reproducibility of the Portal Imaging Panel for Routine FFF QC and Patient Plan Dose Measurements

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Willett, A; Gilmore, M; Rowbottom, C

    2016-06-15

    Purpose: The purpose of this work was to see if the EPID is a viable alternative to other QA devices for routine FFF QA and plan dose measurements. Methods: Sensitivity measurements were made to assess response to small changes in field size and beam steering. QA plans were created where field size was varied from baseline values (5–5.5cm, 20–20.5cm). Beam steering was adjusted by altering values in service mode (Symmetry 0–3%). Plans were measured using the Varian portal imager (aS1200 DMI panel), QA3 (Sun Nuclear), and Starcheck Maxi (PTW). FFF beam parameters as stated in Fogliata et al were calculated.more » Constancy measurements were taken using all 3 QC devices to measure a MLC defined 20×20cm field. Two clinical SABR patient plans were measured on a Varian Edge linac, using the Portal Dosimetry module in ARIA, and results compared with analysis made using Delta4 (ScandiDos). Results: The EPID and the Starcheck performed better at detecting clinically relevant changes in field size with the QA3 performing better when detecting similar changes in beam symmetry. Consistency measurements with the EPID and Starcheck were equivalent, with comparable standard deviations. Clinical plan measurements on the EPID compared well with Delta4 results at 3%/1mm. Conclusion: Our results show that for FFF QA measurements such as field size and symmetry, using the EPID is a viable alternative to other QA devices. The EPID could potentially be used for QC measurements with a focus on geometric accuracy, such as MLC positional QA, due to its high resolution compared to other QA devices (EPID 0.34mm, Starcheck 3mm, QA3 5mm). Good agreement between Delta4 and portal dosimetry also indicated the EPID may be a suitable alternative for measurement of clinical plans.« less

  8. 40 CFR 98.474 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Injection of Carbon Dioxide § 98.474 Monitoring and QA/QC.... (2) You must determine the quarterly mass or volume of contents in all containers if you receive CO2...

  9. 40 CFR 98.474 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Injection of Carbon Dioxide § 98.474 Monitoring and QA/QC.... (2) You must determine the quarterly mass or volume of contents in all containers if you receive CO2...

  10. 40 CFR 98.474 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Injection of Carbon Dioxide § 98.474 Monitoring and QA/QC.... (2) You must determine the quarterly mass or volume of contents in all containers if you receive CO2...

  11. SU-E-T-268: Differences in Treatment Plan Quality and Delivery Between Two Commercial Treatment Planning Systems for Volumetric Arc-Based Radiation Therapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chen, S; Zhang, H; Zhang, B

    2015-06-15

    Purpose: To clinically evaluate the differences in volumetric modulated arc therapy (VMAT) treatment plan and delivery between two commercial treatment planning systems. Methods: Two commercial VMAT treatment planning systems with different VMAT optimization algorithms and delivery approaches were evaluated. This study included 16 clinical VMAT plans performed with the first system: 2 spine, 4 head and neck (HN), 2 brain, 4 pancreas, and 4 pelvis plans. These 16 plans were then re-optimized with the same number of arcs using the second treatment planning system. Planning goals were invariant between the two systems. Gantry speed, dose rate modulation, MLC modulation, planmore » quality, number of monitor units (MUs), VMAT quality assurance (QA) results, and treatment delivery time were compared between the 2 systems. VMAT QA results were performed using Mapcheck2 and analyzed with gamma analysis (3mm/3% and 2mm/2%). Results: Similar plan quality was achieved with each VMAT optimization algorithm, and the difference in delivery time was minimal. Algorithm 1 achieved planning goals by highly modulating the MLC (total distance traveled by leaves (TL) = 193 cm average over control points per plan), while maintaining a relatively constant dose rate (dose-rate change <100 MU/min). Algorithm 2 involved less MLC modulation (TL = 143 cm per plan), but greater dose-rate modulation (range = 0-600 MU/min). The average number of MUs was 20% less for algorithm 2 (ratio of MUs for algorithms 2 and 1 ranged from 0.5-1). VMAT QA results were similar for all disease sites except HN plans. For HN plans, the average gamma passing rates were 88.5% (2mm/2%) and 96.9% (3mm/3%) for algorithm 1 and 97.9% (2mm/2%) and 99.6% (3mm/3%) for algorithm 2. Conclusion: Both VMAT optimization algorithms achieved comparable plan quality; however, fewer MUs were needed and QA results were more robust for Algorithm 2, which more highly modulated dose rate.« less

  12. 40 CFR 98.424 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Carbon Dioxide § 98.424 Monitoring and QA/QC... containers shall measure the mass in each CO2 container using weigh bills, scales, or load cells and sum the...

  13. WE-AB-BRB-10: Filmless QA of CyberKnife MLC-Collimated and Iris-Collimated Fields

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gersh, J; Spectrum Medical Physics, LLC, Greenville, SC

    Purpose: Current methods of CK field shape QA is based on the use of radiochromic film. Though accurate results can be attained, these methods are prone to error, time consuming, and expensive. The techniques described herein perform similar QA using the FOIL Detector (Field, Output, and Image Localization). A key feature of this in-house QA solution, and central to this study, is an aSi flat-panel detector which provides the user with the means to perform accurate, immediate, and quantitative field analysis. Methods: The FOIL detector is automatically aligned in the CK beam using fiducial markers implanted within the detector case.more » Once the system is aligned, a treatment plan is delivered which irradiates the flat-panel imager using the field being tested. The current study tests each of the clinically-used fields shaped using the Iris variable-aperture collimation system using a plan which takes 6 minutes to deliver. The user is immediately provided with field diameter and beam profile, as well as a comparison to baseline values. Additionally, the detector is used to acquire and analyze leaf positions of the InCise multi-leaf collimation system. Results: Using a 6-minute plan consisting of 11 beams of 25MU-per-beam, the FOIL detector provided the user with a quantitative analysis of all clinically-used field shapes. The FOIL detector was also able to clearly resolve field edge junctions in a picket fence test, including slight over-travel of individual leaves as well as inter-leaf leakage. Conclusion: The FOIL system provided comparable field diameter and profile data when compared to methods using film; providing results much faster and with 5% of the MU used for film. When used with the MLC system, the FOIL detector provided the means for immediate quantification of the performance of the system through analysis of leaf positions in a picket fence test field. Author is the President/Owner of Spectrum Medical Physics, LLC, a company which maintains

  14. The influence of surface roughness and solution concentration on pool boiling process in Diethanolamine aqueous solution

    NASA Astrophysics Data System (ADS)

    Khoshechin, Mohsen; Salimi, Farhad; Jahangiri, Alireza

    2018-04-01

    In this research, the effect of surface roughness and concentration of solution on bubble departing frequency and nucleation site density for pool boiling of water/diethanolamine (DEA) binary solution were investigated experimentally. In this investigation, boiling heat transfer coefficient, bubble departing frequency and nucleation site density have been experimentally investigated in various concentrations and heat fluxes. Microstructured surfaces with a wide range of well-defined surface roughness were fabricated, and a heat flux between 1.5-86 kW/m2 was achieved under atmospheric conditions. The Results indicated that surface roughness and concentration of solution increase the bubble departing frequency and nucleation site density with increasing heat flux. The boiling heat transfer coefficient in mixtures of water/DEA increases with increasing concentration of DEA in water. The experimental results were compared with predictions of several used correlations in the literatures. Results showed that the boiling heat transfer coefficients of this case study are much higher than the predicted values by major existing correlations and models. The excellent agreement for bubble departing frequency found between the models of Jackob and Fritz (1966) and experimental data and also the nucleation site density were in close agreement with the model of Paul (1983) data. f bubble departure frequency, 1/s or Hz N Number of nucleation sites per area per time R c Minimum cavity size, m D c critical diameter, m g gravitational acceleration, m/s2 ρ density, kg/m3 T temperature, °c ΔT temperature difference, °c d d vapor bubble diameter, m h fg enthalpy of vaporization, J/kg R Roughness, μm Ja Jakob number cp specific heat, J/kg °c Pr Prandtl number Ar Archimedes number h Heat transfer coefficient, J/(m2 °c) tg time it takes to grow a bubble, s q/A heat flux (kW/m2) tw time required to heat the layer, s gc Correction coefficient of incompatible units R a Surface

  15. CMU OAQA at TREC 2015 LiveQA: Discovering the Right Answer with Clues

    DTIC Science & Technology

    2015-11-20

    QA) system that was evaluated in the TREC 2015 LiveQA Challenge. This system answers real-user questions freshly submitted to the Yahoo ! Answers...questions on the Yahoo ! Answers site 1, which have not yet received a human answer. As per the requirements for this track, participants must deploy their... Yahoo ! Answers. We also designed and im- plemented a new data model and novel relevance ranking methods for LiveQA. During the official run, our QA web

  16. SU-F-T-251: The Quality Assurance for the Heavy Patient Load Department in the Developing Country: The Primary Experience of An Entire Workflow QA Process Management in Radiotherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Xie, J; Wang, J; Peng, J

    Purpose: To implement an entire workflow quality assurance (QA) process in the radiotherapy department and to reduce the error rates of radiotherapy based on the entire workflow management in the developing country. Methods: The entire workflow QA process management starts from patient registration to the end of last treatment including all steps through the entire radiotherapy process. Error rate of chartcheck is used to evaluate the the entire workflow QA process. Two to three qualified senior medical physicists checked the documents before the first treatment fraction of every patient. Random check of the treatment history during treatment was also performed.more » A total of around 6000 patients treatment data before and after implementing the entire workflow QA process were compared from May, 2014 to December, 2015. Results: A systemic checklist was established. It mainly includes patient’s registration, treatment plan QA, information exporting to OIS(Oncology Information System), documents of treatment QAand QA of the treatment history. The error rate derived from the chart check decreases from 1.7% to 0.9% after our the entire workflow QA process. All checked errors before the first treatment fraction were corrected as soon as oncologist re-confirmed them and reinforce staff training was accordingly followed to prevent those errors. Conclusion: The entire workflow QA process improved the safety, quality of radiotherapy in our department and we consider that our QA experience can be applicable for the heavily-loaded radiotherapy departments in developing country.« less

  17. Patient-Specific QA of Spot-Scanning Proton Beams using Radiochromic Film.

    PubMed

    Chan, Maria F; Chen, Chin-Cheng; Shi, Chengyu; Li, Jingdong; Tang, Xiaoli; Li, Xiang; Mah, Dennis

    2017-05-01

    Radiochromic film for spot-scanning QA provides high spatial resolution and efficiency gains from one-shot irradiation for multiple depths. However, calibration can be a tedious procedure which may limit widespread use. Moreover, since there may be an energy dependence, which manifests as a depth dependence, this may require additional measurements for each patient. We present a one-scan protocol to simplify the procedure. A calibration using an EBT3 film, exposed by a 6-level step-wedge plan on a Proteus ® PLUS proton system (IBA, Belgium), was performed at depths of 18, 20, 24cm using Plastic Water ® (CIRS, Norfolk, VA). The calibration doses ranged from 65-250 cGy(RBE) (relative biological effectiveness) for proton energies of 170-200 MeV. A clinical prostate+nodes plan was used for validation. The planar doses at selected depths were measured with EBT3 films and analyzed using One-scan protocol (one-scan digitization of QA film and at least one film exposed to a known dose). The gamma passing rates, dose-difference maps, and profiles of 2D planar doses measured with EBT3 film and IBA MatriXX-PT, versus the RayStation TPS calculations were analyzed and compared. The EBT3 film measurement results matched well with the TPS calculation data with an average passing rate of ~95% for 2%/2mm and slightly lower passing rates were obtained from an ion chamber array detector. We were able to demonstrate that the use of a proton step-wedge provided clinically acceptable results and minimized variations between film-scanner orientation, inter-scan, and scanning conditions. Furthermore, for relative dosimetry (calibration is not done at the time of experiment) it could be derived from no more than two films exposed to known doses (one could be zero) for rescaling the master calibration curve at each depth. The sensitivity of the calibration to depth variations has been explored. One-scan protocol results appear to be comparable to that of the ion chamber array detector

  18. Size-exclusion chromatography for the determination of the boiling point distribution of high-boiling petroleum fractions.

    PubMed

    Boczkaj, Grzegorz; Przyjazny, Andrzej; Kamiński, Marian

    2015-03-01

    The paper describes a new procedure for the determination of boiling point distribution of high-boiling petroleum fractions using size-exclusion chromatography with refractive index detection. Thus far, the determination of boiling range distribution by chromatography has been accomplished using simulated distillation with gas chromatography with flame ionization detection. This study revealed that in spite of substantial differences in the separation mechanism and the detection mode, the size-exclusion chromatography technique yields similar results for the determination of boiling point distribution compared with simulated distillation and novel empty column gas chromatography. The developed procedure using size-exclusion chromatography has a substantial applicability, especially for the determination of exact final boiling point values for high-boiling mixtures, for which a standard high-temperature simulated distillation would have to be used. In this case, the precision of final boiling point determination is low due to the high final temperatures of the gas chromatograph oven and an insufficient thermal stability of both the gas chromatography stationary phase and the sample. Additionally, the use of high-performance liquid chromatography detectors more sensitive than refractive index detection allows a lower detection limit for high-molar-mass aromatic compounds, and thus increases the sensitivity of final boiling point determination. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  19. (Boiling water reactor (BWR) CORA experiments)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ott, L.J.

    To participate in the 1990 CORA Workshop at Kernforschungszentrum Karlsruhe (KfK) GmbH, Karlsruhe, FRG, on October 1--4, and to participate in detailed discussions on October 5 with the KfK CORA Boiling Water Reactor (BWR) experiments. The traveler attended the 1990 CORA Workshop at KfK, FRG. Participation included the presentation of a paper on work performed by the Boiling Water Reactor Core Melt Progression Phenomena Program at Oak Ridge National Laboratory (ORNL) on posttest analyses of CORA BWR experiments. The Statement of Work (November 1989) for the BWR Core Melt Progression Phenomena Program provides for pretest and posttest analyses of themore » BWR CORA experiments performed at KfK. Additionally, it is intended that ORNL personnel participate in the planning process for future CORA BWR experiments. For these purposes, meetings were held with KfK staff to discuss such topics as (1) experimental test schedule, (2) BWR test conduct, (3) perceived BWR experimental needs, and (4) KfK operational staff needs with respect to ORNL support. 19 refs.« less

  20. MO-FG-CAMPUS-TeP1-01: An Efficient Method of 3D Patient Dose Reconstruction Based On EPID Measurements for Pre-Treatment Patient Specific QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    David, R; Lee, C; Calvary Mater Newcastle, Newcastle

    Purpose: To demonstrate an efficient and clinically relevant patient specific QA method by reconstructing 3D patient dose from 2D EPID images for IMRT plans. Also to determine the usefulness of 2D QA metrics when assessing 3D patient dose deviations. Methods: Using the method developed by King et al (Med Phys 39(5),2839–2847), EPID images of IMRT fields were acquired in air and converted to dose at 10 cm depth (SAD setup) in a flat virtual water phantom. Each EPID measured dose map was then divided by the corresponding treatment planning system (TPS) dose map calculated with an identical setup, to derivemore » a 2D “error matrix”. For each field, the error matrix was used to adjust the doses along the respective ray lines in the original patient 3D dose. All field doses were combined to derive a reconstructed 3D patient dose for quantitative analysis. A software tool was developed to efficiently implement the entire process and was tested with a variety of IMRT plans for 2D (virtual flat phantom) and 3D (in-patient) QA analysis. Results: The method was tested on 60 IMRT plans. The mean (± standard deviation) 2D gamma (2%,2mm) pass rate (2D-GPR) was 97.4±3.0% and the mean 2D gamma index (2D-GI) was 0.35±0.06. The 3D PTV mean dose deviation was 1.8±0.8%. The analysis showed very weak correlations between both the 2D-GPR and 2D-GI when compared with PTV mean dose deviations (R2=0.3561 and 0.3632 respectively). Conclusion: Our method efficiently calculates 3D patient dose from 2D EPID images, utilising all of the advantages of an EPID-based dosimetry system. In this study, the 2D QA metrics did not predict the 3D patient dose deviation. This tool allows reporting of the 3D volumetric dose parameters thus providing more clinically relevant patient specific QA.« less

  1. Subcooled forced convection boiling of trichlorotrifluoroethane

    NASA Technical Reports Server (NTRS)

    Dougall, R. S.; Panian, D. J.

    1972-01-01

    Experimental heat-transfer data were obtained for the forced-convection boiling of trichlorotrifluoroethane (R-113 or Freon-113) in a vertical annular test annular test section. The 97 data points obtained covered heat transfer by forced convection, local boiling, and fully-developed boiling. Correlating methods were obtained which accurately predicted the heat flux as a function of wall superheat (boiling curve) over the range of parameters studied.

  2. Optimizing the Combination of Smoking and Boiling on Quality of Korean Traditional Boiled Loin (M. longissimus dorsi)

    PubMed Central

    Choi, Yun-Sang; Kim, Hyun-Wook; Kim, Young-Boong; Kim, Cheon-Jei

    2015-01-01

    The combined effects of smoking and boiling on the proximate composition, technological quality traits, shear force, and sensory characteristics of the Korean traditional boiled loin were studied. Cooking loss, processing loss, and shear force were lower in the smoked/boiled samples than those in the control (without smoking treatment) (p<0.05). The results showed that the boiled loin samples between the control and treatment did not differ significantly in protein, fat, or ash contents, or pH values (p>0.05). The treated samples had higher score for overall acceptability than the control (p<0.05). Thus, these results show that the Korean traditional boiled loin treated with smoking for 60 min before boiling had improved physicochemical properties and sensory characteristics. PMID:26761822

  3. Hubble Space Telescope: SRM/QA observations and lessons learned

    NASA Technical Reports Server (NTRS)

    Rodney, George A.

    1990-01-01

    The Hubble Space Telescope (HST) Optical Systems Board of Investigation was established on July 2, 1990 to review, analyze, and evaluate the facts and circumstances regarding the manufacture, development, and testing of the HST Optical Telescope Assembly (OTA). Specifically, the board was tasked to ascertain what caused the spherical aberration and how it escaped notice until on-orbit operation. The error that caused the on-orbit spherical aberration in the primary mirror was traced to the assembly process of the Reflective Null Corrector, one of the three Null Correctors developed as special test equipment (STE) to measure and test the primary mirror. Therefore, the safety, reliability, maintainability, and quality assurance (SRM&QA) investigation covers the events and the overall product assurance environment during the manufacturing phase of the primary mirror and Null Correctors (from 1978 through 1981). The SRM&QA issues that were identified during the HST investigation are summarized. The crucial product assurance requirements (including nonconformance processing) for the HST are examined. The history of Quality Assurance (QA) practices at Perkin-Elmer (P-E) for the period under investigation are reviewed. The importance of the information management function is discussed relative to data retention/control issues. Metrology and other critical technical issues also are discussed. The SRM&QA lessons learned from the investigation are presented along with specific recommendations. Appendix A provides the MSFC SRM&QA report. Appendix B provides supplemental reference materials. Appendix C presents the findings of the independent optical consultants, Optical Research Associates (ORA). Appendix D provides further details of the fault-tree analysis portion of the investigation process.

  4. The myth of the boiling point.

    PubMed

    Chang, Hasok

    2008-01-01

    Around 1800, many reputable scientists reported significant variations in the temperature of pure water boiling under normal atmospheric pressure. The reported variations included a difference of over 1 degree C between boiling in metallic and glass vessels (Gay-Lussac), and "superheating" up to 112 degrees C on extracting dissolved air out of water (De Luc). I have confirmed most of these observations in my own experiments, many of which are described in this paper. Water boils at the "boiling point" only under very particular circumstances. Our common-sense intuition about the fixedness of the boiling point is only sustained by our limited experience.

  5. Universality of oscillating boiling in Leidenfrost transition

    NASA Astrophysics Data System (ADS)

    Tran, Tuan; Khavari, Mohammad

    2017-11-01

    The Leidenfrost transition leads a boiling system to the boiling crisis, a state in which the liquid loses contact with the heated surface due to excessive vapor generation. Here, using experiments of liquid droplets boiling on a heated surface, we report a new phenomenon, termed oscillating boiling, at the Leidenfrost transition. We show that oscillating boiling results from the competition between two effects: separation of liquid from the heated surface due to localized boiling, and rewetting. We argue theoretically that the Leidenfrost transition can be predicted based on its link with the oscillating boiling phenomenon, and verify the prediction experimentally for various liquids. This work was funded by Nanyang Technological University and A*STAR, Singapore.

  6. THE IMPORTANCE OF A SUCCESSFUL QUALITY ASSURANCE (QA) PROGRAM FROM A RESEARCH MANAGER'S PERSPECTIVE

    EPA Science Inventory

    The paper discusses the Air Pollution Prevention and Control Division's Quality Assurance (QA) program and the approaches used to meet QA requirements in the Division. The presentation is a technical manager's perspective of the Division's requirements for and approach to QA in i...

  7. SU-F-T-295: MLCs Performance and Patient-Specific IMRT QA Using Log File Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Osman, A; American University of Biuret Medical Center, Biuret; Maalej, N

    2016-06-15

    Purpose: To analyze the performance of the multi-leaf collimators (MLCs) from the log files recorded during the intensity modulated radiotherapy (IMRT) treatment and to construct the relative fluence maps and do the gamma analysis to compare the planned and executed MLCs movement. Methods: We developed a program to extract and analyze the data from dynamic log files (dynalog files) generated from sliding window IMRT delivery treatments. The program extracts the planned and executed (actual or delivered) MLCs movement, calculates and compares the relative planned and executed fluences. The fluence maps were used to perform the gamma analysis (with 3% dosemore » difference and 3 mm distance to agreement) for 3 IMR patients. We compared our gamma analysis results with those obtained from portal dose image prediction (PDIP) algorithm performed using the EPID. Results: For 3 different IMRT patient treatments, the maximum difference between the planned and the executed MCLs positions was 1.2 mm. The gamma analysis results of the planned and delivered fluences were in good agreement with the gamma analysis from portal dosimetry. The maximum difference for number of pixels passing the gamma criteria (3%/3mm) was 0.19% with respect to portal dosimetry results. Conclusion: MLC log files can be used to verify the performance of the MLCs. Patientspecific IMRT QA based on MLC movement log files gives similar results to EPID dosimetry results. This promising method for patient-specific IMRT QA is fast, does not require dose measurements in a phantom, can be done before the treatment and for every fraction, and significantly reduces the IMRT workload. The author would like to thank King Fahd University of petroleum and Minerals for the support.« less

  8. SU-E-T-646: Quality Assurance of Truebeam Multi-Leaf Collimator Using a MLC QA Phantom

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, J; Lu, J; Hong, D

    2015-06-15

    Purpose: To perform a routine quality assurance procedure for Truebeam multi-leaf collimator (MLC) using MLC QA phantom, verify the stability and reliability of MLC during the treatment. Methods: MLC QA phantom is a specialized phantom for MLC quality assurance (QA), and contains five radio-opaque spheres that are embedded in an “L” shape. The phantom was placed isocentrically on the Truebeam treatment couch for the tests. A quality assurance plan was setted up in the Eclipse v10.0, the fields that need to be delivered in order to acquire the necessary images, the MLC shapes can then be obtained by the images.more » The images acquired by the electronic portal imaging device (EPID), and imported into the PIPSpro software for the analysis. The tests were delivered twelve weeks (once a week) to verify consistency of the delivery, and the images are acquired in the same manner each time. Results: For the Leaf position test, the average position error was 0.23mm±0.02mm (range: 0.18mm∼0.25mm). The Leaf width was measured at the isocenter, the average error was 0.06mm±0.02mm (range: 0.02mm∼0.08mm) for the Leaf width test. Multi-Port test showed the dynamic leaf shift error, the average error was 0.28mm±0.03mm (range: 0.2mm∼0.35mm). For the leaf transmission test, the average inter-leaf leakage value was 1.0%±0.17% (range: 0.8%∼1.3%) and the average inter-bank leakage value was 32.6%±2.1% (range: 30.2%∼36.1%). Conclusion: By the test of 12 weeks, the MLC system of the Truebeam is running in a good condition and the MLC system can be steadily and reliably carried out during the treatment. The MLC QA phantom is a useful test tool for the MLC QA.« less

  9. [A Quality Assurance (QA) System with a Web Camera for High-dose-rate Brachytherapy].

    PubMed

    Hirose, Asako; Ueda, Yoshihiro; Oohira, Shingo; Isono, Masaru; Tsujii, Katsutomo; Inui, Shouki; Masaoka, Akira; Taniguchi, Makoto; Miyazaki, Masayoshi; Teshima, Teruki

    2016-03-01

    The quality assurance (QA) system that simultaneously quantifies the position and duration of an (192)Ir source (dwell position and time) was developed and the performance of this system was evaluated in high-dose-rate brachytherapy. This QA system has two functions to verify and quantify dwell position and time by using a web camera. The web camera records 30 images per second in a range from 1,425 mm to 1,505 mm. A user verifies the source position from the web camera at real time. The source position and duration were quantified with the movie using in-house software which was applied with a template-matching technique. This QA system allowed verification of the absolute position in real time and quantification of dwell position and time simultaneously. It was evident from the verification of the system that the mean of step size errors was 0.31±0.1 mm and that of dwell time errors 0.1±0.0 s. Absolute position errors can be determined with an accuracy of 1.0 mm at all dwell points in three step sizes and dwell time errors with an accuracy of 0.1% in more than 10.0 s of the planned time. This system is to provide quick verification and quantification of the dwell position and time with high accuracy at various dwell positions without depending on the step size.

  10. Study of impacts of different evaluation criteria on gamma pass rates in VMAT QA using MatriXX and EPID

    NASA Astrophysics Data System (ADS)

    Noufal, Manthala Padannayil; Abdullah, Kallikuzhiyil Kochunny; Niyas, Puzhakkal; Subha, Pallimanhayil Abdul Raheem

    2017-12-01

    Aim: This study evaluates the impacts of using different evaluation criteria on gamma pass rates in two commercially available QA methods employed for the verification of VMAT plans using different hypothetical planning target volumes (PTVs) and anatomical regions. Introduction: Volumetric modulated arc therapy (VMAT) is a widely accepted technique to deliver highly conformal treatment in a very efficient manner. As their level of complexity is high in comparison to intensity-modulated radiotherapy (IMRT), the implementation of stringent quality assurance (QA) before treatment delivery is of paramount importance. Material and Methods: Two sets of VMAT plans were generated using Eclipse planning systems, one with five different complex hypothetical three-dimensional PTVs and one including three anatomical regions. The verification of these plans was performed using a MatriXX ionization chamber array embedded inside a MultiCube phantom and a Varian EPID dosimetric system attached to a Clinac iX. The plans were evaluated based on the 3%/3 mm, 2%/2 mm, and 1%/1 mm global gamma criteria and with three low-dose threshold values (0%, 10%, and 20%). Results: The gamma pass rates were above 95% in all VMAT plans, when the 3%/3mm gamma criterion was used and no threshold was applied. In both systems, the pass rates decreased as the criteria become stricter. Higher pass rates were observed when no threshold was applied and they tended to decrease for 10% and 20% thresholds. Conclusion: The results confirm the suitability of the equipments used and the validity of the plans. The study also confirmed that the threshold settings greatly affect the gamma pass rates, especially for lower gamma criteria.

  11. ETD QA CORE TEAM: AN ELOQUENT SOLUTION TO A COMPLEX PROBLEM

    EPA Science Inventory

    ETD QA CORE TEAM: AN ELOQUENT SOLUTION TO A COMPLEX PROBLEMThomas J. Hughes, QA and Records Manager, Experimental Toxicology Division (ETD), National Health and Environmental Effects Research Laboratory (NHEERL), ORD, U.S. EPA, RTP, NC 27709

    ETD is the largest health divis...

  12. The unusually strong hydrogen bond between the carbonyl of Q(A) and His M219 in the Rhodobacter sphaeroides reaction center is not essential for efficient electron transfer from Q(A)(-) to Q(B).

    PubMed

    Breton, Jacques; Lavergne, Jérôme; Wakeham, Marion C; Nabedryk, Eliane; Jones, Michael R

    2007-06-05

    In native reaction centers (RCs) from photosynthetic purple bacteria the primary quinone (QA) and the secondary quinone (QB) are interconnected via a specific His-Fe-His bridge. In Rhodobacter sphaeroides RCs the C4=O carbonyl of QA forms a very strong hydrogen bond with the protonated Npi of His M219, and the Ntau of this residue is in turn coordinated to the non-heme iron atom. The second carbonyl of QA is engaged in a much weaker hydrogen bond with the backbone N-H of Ala M260. In previous work, a Trp side chain was introduced by site-directed mutagenesis at the M260 position in the RC of Rb. sphaeroides, resulting in a complex that is completely devoid of QA and therefore nonfunctional. A photochemically competent derivative of the AM260W mutant was isolated that contains a Cys side chain at the M260 position (denoted AM260(W-->C)). In the present work, the interactions between the carbonyl groups of QA and the protein in the AM260(W-->C) suppressor mutant have been characterized by light-induced FTIR difference spectroscopy of the photoreduction of QA. The QA-/QA difference spectrum demonstrates that the strong interaction between the C4=O carbonyl of QA and His M219 is lost in the mutant, and the coupled CO and CC modes of the QA- semiquinone are also strongly perturbed. In parallel, a band assigned to the perturbation of the C5-Ntau mode of His M219 upon QA- formation in the native RC is lacking in the spectrum of the mutant. Furthermore, a positive band between 2900 and 2400 cm-1 that is related to protons fluctuating within a network of highly polarizable hydrogen bonds in the native RC is reduced in amplitude in the mutant. On the other hand, the QB-/QB FTIR difference spectrum is essentially the same as for the native RC. The kinetics of electron transfer from QA- to QB were measured by the flash-induced absorption changes at 780 nm. Compared to native RCs the absorption transients are slowed by a factor of about 2 for both the slow phase (in the

  13. Odd-Boiled Eggs

    ERIC Educational Resources Information Center

    Kaminsky, Kenneth; Scheman, Naomi

    2010-01-01

    At a Shabbat lunch in Madrid not long ago, the conversation turned to the question of boiling eggs. One of the guests mentioned that a Dutch rabbi he knew had heard that in order to make it more likely that boiled eggs be kosher, you should add an egg to the pot if the number you began with was even. According to the laws of Kashruth, Jews may not…

  14. 21 CFR 872.6710 - Boiling water sterilizer.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 21 Food and Drugs 8 2011-04-01 2011-04-01 false Boiling water sterilizer. 872.6710 Section 872...) MEDICAL DEVICES DENTAL DEVICES Miscellaneous Devices § 872.6710 Boiling water sterilizer. (a) Identification. A boiling water sterilizer is an AC-powered device that consists of a container for boiling water...

  15. 21 CFR 872.6710 - Boiling water sterilizer.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false Boiling water sterilizer. 872.6710 Section 872...) MEDICAL DEVICES DENTAL DEVICES Miscellaneous Devices § 872.6710 Boiling water sterilizer. (a) Identification. A boiling water sterilizer is an AC-powered device that consists of a container for boiling water...

  16. 21 CFR 872.6710 - Boiling water sterilizer.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 21 Food and Drugs 8 2013-04-01 2013-04-01 false Boiling water sterilizer. 872.6710 Section 872...) MEDICAL DEVICES DENTAL DEVICES Miscellaneous Devices § 872.6710 Boiling water sterilizer. (a) Identification. A boiling water sterilizer is an AC-powered device that consists of a container for boiling water...

  17. 21 CFR 872.6710 - Boiling water sterilizer.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 21 Food and Drugs 8 2012-04-01 2012-04-01 false Boiling water sterilizer. 872.6710 Section 872...) MEDICAL DEVICES DENTAL DEVICES Miscellaneous Devices § 872.6710 Boiling water sterilizer. (a) Identification. A boiling water sterilizer is an AC-powered device that consists of a container for boiling water...

  18. Revision 2 of the Enbridge Quality Assurance Project Plan

    EPA Pesticide Factsheets

    This Quality Assurance Project Plan (QAPP) presents Revision 2 of the organization, objectives, planned activities, and specific quality assurance/quality control (QA/QC) procedures associated with the Enbridge Marshall Pipeline Release Project.

  19. Boils

    MedlinePlus

    ... boil is an infection that affects groups of hair follicles and nearby skin tissue. Related conditions include: Carbunculosis ... found on the skin's surface. Damage to the hair follicle allows the infection to grow deeper into the ...

  20. Pool Boiling Experiment Has Five Successful Flights

    NASA Technical Reports Server (NTRS)

    Chiaramonte, Fran

    1997-01-01

    The Pool Boiling Experiment (PBE) is designed to improve understanding of the fundamental mechanisms that constitute nucleate pool boiling. Nucleate pool boiling is a process wherein a stagnant pool of liquid is in contact with a surface that can supply heat to the liquid. If the liquid absorbs enough heat, a vapor bubble can be formed. This process occurs when a pot of water boils. On Earth, gravity tends to remove the vapor bubble from the heating surface because it is dominated by buoyant convection. In the orbiting space shuttle, however, buoyant convection has much less of an effect because the forces of gravity are very small. The Pool Boiling Experiment was initiated to provide insight into this nucleate boiling process, which has many earthbound applications in steamgeneration power plants, petroleum plants, and other chemical plants. In addition, by using the test fluid R-113, the Pool Boiling Experiment can provide some basic understanding of the boiling behavior of cryogenic fluids without the large cost of an experiment using an actual cryogen.

  1. Technical Note: Response time evolution of XR-QA2 GafChromic™ film models.

    PubMed

    Aldelaijan, Saad; Tomic, Nada; Papaconstadopoulos, Pavlos; Schneider, James; Seuntjens, Jan; Shih, Shelley; Lewis, David; Devic, Slobodan

    2018-01-01

    To evaluate the response of the newest XR-QA2 GafChromic™ film model in terms of postexposure signal growth and energy response in comparison with the older XR-QA (Version 2) model. Pieces of film were irradiated to air kerma in air values up to 12 cGy with several beam qualities (5.3-8.25 mm Al) commonly used for CT scanning. Film response was scored in terms of net reflectance from scanned film images at various points in time postirradiation ranging from 1 to 7 days and 5 months postexposure. To reconstruct the measurement signal changes with postirradiation delay, we irradiated one film piece and then scanned it at different point times starting from 2" min and up to 3 days postexposure. For all beam qualities and dose range investigated, it appears that the XR-QA2 film signal completely saturated after 15 h. Compared to 15 h postirradiation scanning time, the observed variation in net reflectance were 3%, 2%, and 1% for film scanned 2" min, 20 min, and 3 h after exposure, respectively, which is well within the measurement uncertainty of the XR-QA2 based reference radiochromic film dosimetry system. A comparison between the XR-QA (Version 2) and the XR-QA2 film response after several months (relative to their responses after 24 h) show differences in up to 8% and 1% for each film model respectively. The replacement of cesium bromide in the older XR-QA (Version 2) film model with bismuth oxide in the newer XR-QA2 film, while keeping the same single sensitive layer structure, lead to a significantly more stable postexposure response. © 2017 American Association of Physicists in Medicine.

  2. A structural basis for antigen presentation by the MHC class Ib molecule, Qa-1b.

    PubMed

    Zeng, Li; Sullivan, Lucy C; Vivian, Julian P; Walpole, Nicholas G; Harpur, Christopher M; Rossjohn, Jamie; Clements, Craig S; Brooks, Andrew G

    2012-01-01

    The primary function of the monomorphic MHC class Ib molecule Qa-1(b) is to present peptides derived from the leader sequences of other MHC class I molecules for recognition by the CD94-NKG2 receptors expressed by NK and T cells. Whereas the mode of peptide presentation by its ortholog HLA-E, and subsequent recognition by CD94-NKG2A, is known, the molecular basis of Qa-1(b) function is unclear. We have assessed the interaction between Qa-1(b) and CD94-NKG2A and shown that they interact with an affinity of 17 μM. Furthermore, we have determined the structure of Qa-1(b) bound to the leader sequence peptide, Qdm (AMAPRTLLL), to a resolution of 1.9 Å and compared it with that of HLA-E. The crystal structure provided a basis for understanding the restricted peptide repertoire of Qa-1(b). Whereas the Qa-1(b-AMAPRTLLL) complex was similar to that of HLA-E, significant sequence and structural differences were observed between the respective Ag-binding clefts. However, the conformation of the Qdm peptide bound by Qa-1(b) was very similar to that of peptide bound to HLA-E. Although a number of conserved innate receptors can recognize heterologous ligands from other species, the structural differences between Qa-1(b) and HLA-E manifested in CD94-NKG2A ligand recognition being species specific despite similarities in peptide sequence and conformation. Collectively, our data illustrate the structural homology between Qa-1(b) and HLA-E and provide a structural basis for understanding peptide repertoire selection and the specificity of the interaction of Qa-1(b) with CD94-NKG2 receptors.

  3. TU-AB-201-02: An Automated Treatment Plan Quality Assurance Program for Tandem and Ovoid High Dose-Rate Brachytherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tan, J; Shi, F; Hrycushko, B

    2015-06-15

    Purpose: For tandem and ovoid (T&O) HDR brachytherapy in our clinic, it is required that the planning physicist manually capture ∼10 images during planning, perform a secondary dose calculation and generate a report, combine them into a single PDF document, and upload it to a record- and-verify system to prove to an independent plan checker that the case was planned correctly. Not only does this slow down the already time-consuming clinical workflow, the PDF document also limits the number of parameters that can be checked. To solve these problems, we have developed a web-based automatic quality assurance (QA) program. Methods:more » We set up a QA server accessible through a web- interface. A T&O plan and CT images are exported as DICOMRT files and uploaded to the server. The software checks 13 geometric features, e.g. if the dwell positions are reasonable, and 10 dosimetric features, e.g. secondary dose calculations via TG43 formalism and D2cc to critical structures. A PDF report is automatically generated with errors and potential issues highlighted. It also contains images showing important geometric and dosimetric aspects to prove the plan was created following standard guidelines. Results: The program has been clinically implemented in our clinic. In each of the 58 T&O plans we tested, a 14- page QA report was automatically generated. It took ∼45 sec to export the plan and CT images and ∼30 sec to perform the QA tests and generate the report. In contrast, our manual QA document preparation tooks on average ∼7 minutes under optimal conditions and up to 20 minutes when mistakes were made during the document assembly. Conclusion: We have tested the efficiency and effectiveness of an automated process for treatment plan QA of HDR T&O cases. This software was shown to improve the workflow compared to our conventional manual approach.« less

  4. SU-E-T-472: Improvement of IMRT QA Passing Rate by Correcting Angular Dependence of MatriXX

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chen, Q; Watkins, W; Kim, T

    2015-06-15

    Purpose: Multi-channel planar detector arrays utilized for IMRT-QA, such as the MatriXX, exhibit an incident-beam angular dependent response which can Result in false-positive gamma-based QA results, especially for helical tomotherapy plans which encompass the full range of beam angles. Although MatriXX can use with gantry angle sensor to provide automatically angular correction, this sensor does not work with tomotherapy. The purpose of the study is to reduce IMRT-QA false-positives by correcting for the MatriXX angular dependence. Methods: MatriXX angular dependence was characterized by comparing multiple fixed-angle irradiation measurements with corresponding TPS computed doses. For 81 Tomo-helical IMRT-QA measurements, two differentmore » correction schemes were tested: (1) A Monte-Carlo dose engine was used to compute MatriXX signal based on the angular-response curve. The computed signal was then compared with measurement. (2) Uncorrected computed signal was compared with measurements uniformly scaled to account for the average angular dependence. Three scaling factor (+2%, +2.5%, +3%) were tested. Results: The MatriXX response is 8% less than predicted for a PA beam even when the couch is fully accounted for. Without angular correction, only 67% of the cases pass the >90% points γ<1 (3%, 3mm). After full angular correction, 96% of the cases pass the criteria. Of three scaling factors, +2% gave the highest passing rate (89%), which is still less than the full angular correction method. With a stricter γ(2%,3mm) criteria, the full angular correction method was still able to achieve the 90% passing rate while the scaling method only gives 53% passing rate. Conclusion: Correction for the MatriXX angular dependence reduced the false-positives rate of our IMRT-QA process. It is necessary to correct for the angular dependence to achieve the IMRT passing criteria specified in TG129.« less

  5. Cryogenic Boil-Off Reduction System

    NASA Astrophysics Data System (ADS)

    Plachta, David W.; Guzik, Monica C.

    2014-03-01

    A computational model of the cryogenic boil-off reduction system being developed by NASA as part of the Cryogenic Propellant Storage and Transfer technology maturation project has been applied to a range of propellant storage tanks sizes for high-performing in-space cryogenic propulsion applications. This effort focuses on the scaling of multi-layer insulation (MLI), cryocoolers, broad area cooling shields, radiators, solar arrays, and tanks for liquid hydrogen propellant storage tanks ranging from 2 to 10 m in diameter. Component scaling equations were incorporated into the Cryogenic Analysis Tool, a spreadsheet-based tool used to perform system-level parametric studies. The primary addition to the evolution of this updated tool is the integration of a scaling method for reverse turbo-Brayton cycle cryocoolers, as well as the development and inclusion of Self-Supporting Multi-Layer Insulation. Mass, power, and sizing relationships are traded parametrically to establish the appropriate loiter period beyond which this boil-off reduction system application reduces mass. The projected benefit compares passive thermal control to active thermal control, where active thermal control is evaluated for reduced boil-off with a 90 K shield, zero boil-off with a single heat interception stage at the tank wall, and zero boil-off with a second interception stage at a 90 K shield. Parametric studies show a benefit over passive storage at loiter durations under one month, in addition to showing a benefit for two-stage zero boil-off in terms of reducing power and mass as compared to single stage zero boil-off. Furthermore, active cooling reduces the effect of varied multi-layer insulation performance, which, historically, has been shown to be significant.

  6. Correlational approach to turbulent saturated film boiling

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chu, T.Y.

    A correlation method for saturated film boiling is proposed. The correlation is based on the analogy between film boiling and natural convection. As in the case of natural convection, the turbulent film boiling correlation takes the form of a Nusselt number versus the Raleigh number power law, Nu[sub B] [proportional to] Ra[sub B][sup 1.3]. The proposed correlation shows very good agreement with current data for film boiling of water from vertical surfaces. The general applicability of the correlation is established by comparisons with film boiling data from R-113 and cryogenic fluids. 25 refs., 8 figs.

  7. Evaporation, Boiling and Bubbles

    ERIC Educational Resources Information Center

    Goodwin, Alan

    2012-01-01

    Evaporation and boiling are both terms applied to the change of a liquid to the vapour/gaseous state. This article argues that it is the formation of bubbles of vapour within the liquid that most clearly differentiates boiling from evaporation although only a minority of chemistry textbooks seems to mention bubble formation in this context. The…

  8. Film boiling of mercury droplets

    NASA Technical Reports Server (NTRS)

    Baumeister, K. J.; Schoessow, G. J.; Chmielewski, C. E.

    1975-01-01

    Vaporization times of mercury droplets in Leidenfrost film boiling on a flat horizontal plate are measured in an air atmosphere. Extreme care was used to prevent large amplitude droplet vibrations and surface wetting; therefore, these data can be compared to film boiling theory. Diffusion from the upper surface of the drop appears as a dominant mode of mass transfer from the drop. A closed-form analytical film boiling theory is developed to account for the diffusive evaporation. Reasonable agreement between data and theory is seen.

  9. Quality assurance for online adapted treatment plans: benchmarking and delivery monitoring simulation.

    PubMed

    Li, Taoran; Wu, Qiuwen; Yang, Yun; Rodrigues, Anna; Yin, Fang-Fang; Jackie Wu, Q

    2015-01-01

    An important challenge facing online adaptive radiation therapy is the development of feasible and efficient quality assurance (QA). This project aimed to validate the deliverability of online adapted plans and develop a proof-of-concept online delivery monitoring system for online adaptive radiation therapy QA. The first part of this project benchmarked automatically online adapted prostate treatment plans using traditional portal dosimetry IMRT QA. The portal dosimetry QA results of online adapted plans were compared to original (unadapted) plans as well as randomly selected prostate IMRT plans from our clinic. In the second part, an online delivery monitoring system was designed and validated via a simulated treatment with intentional multileaf collimator (MLC) errors. This system was based on inputs from the dynamic machine information (DMI), which continuously reports actual MLC positions and machine monitor units (MUs) at intervals of 50 ms or less during delivery. Based on the DMI, the system performed two levels of monitoring/verification during the delivery: (1) dynamic monitoring of cumulative fluence errors resulting from leaf position deviations and visualization using fluence error maps (FEMs); and (2) verification of MLC positions against the treatment plan for potential errors in MLC motion and data transfer at each control point. Validation of the online delivery monitoring system was performed by introducing intentional systematic MLC errors (ranging from 0.5 to 2 mm) to the DMI files for both leaf banks. These DMI files were analyzed by the proposed system to evaluate the system's performance in quantifying errors and revealing the source of errors, as well as to understand patterns in the FEMs. In addition, FEMs from 210 actual prostate IMRT beams were analyzed using the proposed system to further validate its ability to catch and identify errors, as well as establish error magnitude baselines for prostate IMRT delivery. Online adapted plans were

  10. Enabling Highly Effective Boiling from Superhydrophobic Surfaces

    NASA Astrophysics Data System (ADS)

    Allred, Taylor P.; Weibel, Justin A.; Garimella, Suresh V.

    2018-04-01

    A variety of industrial applications such as power generation, water distillation, and high-density cooling rely on heat transfer processes involving boiling. Enhancements to the boiling process can improve the energy efficiency and performance across multiple industries. Highly wetting textured surfaces have shown promise in boiling applications since capillary wicking increases the maximum heat flux that can be dissipated. Conversely, highly nonwetting textured (superhydrophobic) surfaces have been largely dismissed for these applications as they have been shown to promote formation of an insulating vapor film that greatly diminishes heat transfer efficiency. The current Letter shows that boiling from a superhydrophobic surface in an initial Wenzel state, in which the surface texture is infiltrated with liquid, results in remarkably low surface superheat with nucleate boiling sustained up to a critical heat flux typical of hydrophilic wetting surfaces, and thus upends this conventional wisdom. Two distinct boiling behaviors are demonstrated on both micro- and nanostructured superhydrophobic surfaces based on the initial wetting state. For an initial surface condition in which vapor occupies the interstices of the surface texture (Cassie-Baxter state), premature film boiling occurs, as has been commonly observed in the literature. However, if the surface texture is infiltrated with liquid (Wenzel state) prior to boiling, drastically improved thermal performance is observed; in this wetting state, the three-phase contact line is pinned during vapor bubble growth, which prevents the development of a vapor film over the surface and maintains efficient nucleate boiling behavior.

  11. SU-F-T-236: Comparison of Two IMRT/VMAT QA Systems Using Gamma Index Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dogan, N; Denissova, S

    2016-06-15

    Purpose: The goal of this study is to assess differences in the Gamma index pass rates when using two commercial QA systems and provide optimum Gamma index parameters for pre-treatment patient specific QA. Methods: Twenty-two VMAT cases that consisted of prostate, lung, head and neck, spine, brain and pancreas, were included in this study. The verification plans have been calculated using AcurosXB(V11) algorithm for different dose grids (1.5mm, 2.5mm, 3mm). The measurements were performed on TrueBeam(Varian) accelerator using both EPID(S1000) portal imager and ArcCheck(SunNuclearCorp) devices. Gamma index criteria variation of 3%/3mm, 2%/3mm, 2%/2mm and threshold (TH) doses of 5% tomore » 50% were used in analysis. Results: The differences in Gamma pass rates between two devices are not statistically significant for 3%/3mm, yielding pass rate higher than 95%. Increase of lower dose TH showed reduced pass rates for both devices. ArcCheck’s more pronounced effect can be attributed to higher contribution of lower dose region spread. As expected, tightening criteria to 2%/2mm (TH: 10%) decreased Gamma pass rates below 95%. Higher EPID (92%) pass rates compared to ArcCheck (86%) probably due to better spatial resolution. Portal Dosimetry results showed lower Gamma pass rates for composite plans compared to individual field pass rates. This may be due to the expansion in the analyzed region which includes pixels not included in the separate field analysis. Decreasing dose grid size from 2.5mm to 1.5mm did not show statistically significant (p<0.05) differences in Gamma pass rates for both QA devices. Conclusion: Overall, both system measurements agree well with calculated dose when using gamma index criteria of 3%/3mm for a variety of VMAT cases. Variability between two systems increases using different dose GRID, TH and tighter gamma criteria and must be carefully assessed prior to clinical use.« less

  12. TH-A-BRC-00: New Task Groups for External Beam QA: An Overview

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    2016-06-15

    AAPM TG-135U1 QA for Robotic Radiosurgery - Sonja Dieterich Since the publication of AAPM TG-135 in 2011, the technology of robotic radiosurgery has rapidly developed. AAPM TG-135U1 will provide recommendations on the clinical practice for using the IRIS collimator, fiducial-less real-time motion tracking, and Monte Carlo based treatment planning. In addition, it will summarize currently available literature about uncertainties. Learning Objectives: Understand the progression of technology since the first TG publication Learn which new QA procedures should be implemented for new technologies Be familiar with updates to clinical practice guidelines AAPM TG-178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance -more » Steven Goetsch Purpose: AAPM Task Group 178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance was formed in August, 2008. The Task Group has 12 medical physicists, two physicians and two consultants. Methods: A round robin dosimetry intercomparison of proposed ionization chambers, electrometer and dosimetry phantoms was conducted over a 15 month period in 2011 and 2012 (Med Phys 42, 11, Nov, 2015). The data obtained at 9 institutions (with ten different Elekta Gamma Knife units) was analyzed by the lead author using several protocols. Results: The most consistent results were obtained using the Elekta ABS 16cm diameter phantom, with the TG-51 protocol modified as recommended by Alfonso et al (Med Phys 35, 11, Nov 2008). A key white paper (Med Phys, in press) sponsored by Elekta Corporation, was used to obtain correction factors for the ionization chambers and phantoms used in this intercomparison. Consistent results were obtained for both Elekta Gamma Knife Model 4C and Gamma Knife Perfexion units as measured with each of two miniature ionization chambers. Conclusion: The full report gives clinical history and background of gamma stereotactic radiosurgery, clinical examples and history, quality assurance recommendations and

  13. TH-A-BRC-01: AAPM TG-135U1 QA for Robotic Radiosurgery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dieterich, S.

    AAPM TG-135U1 QA for Robotic Radiosurgery - Sonja Dieterich Since the publication of AAPM TG-135 in 2011, the technology of robotic radiosurgery has rapidly developed. AAPM TG-135U1 will provide recommendations on the clinical practice for using the IRIS collimator, fiducial-less real-time motion tracking, and Monte Carlo based treatment planning. In addition, it will summarize currently available literature about uncertainties. Learning Objectives: Understand the progression of technology since the first TG publication Learn which new QA procedures should be implemented for new technologies Be familiar with updates to clinical practice guidelines AAPM TG-178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance -more » Steven Goetsch Purpose: AAPM Task Group 178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance was formed in August, 2008. The Task Group has 12 medical physicists, two physicians and two consultants. Methods: A round robin dosimetry intercomparison of proposed ionization chambers, electrometer and dosimetry phantoms was conducted over a 15 month period in 2011 and 2012 (Med Phys 42, 11, Nov, 2015). The data obtained at 9 institutions (with ten different Elekta Gamma Knife units) was analyzed by the lead author using several protocols. Results: The most consistent results were obtained using the Elekta ABS 16cm diameter phantom, with the TG-51 protocol modified as recommended by Alfonso et al (Med Phys 35, 11, Nov 2008). A key white paper (Med Phys, in press) sponsored by Elekta Corporation, was used to obtain correction factors for the ionization chambers and phantoms used in this intercomparison. Consistent results were obtained for both Elekta Gamma Knife Model 4C and Gamma Knife Perfexion units as measured with each of two miniature ionization chambers. Conclusion: The full report gives clinical history and background of gamma stereotactic radiosurgery, clinical examples and history, quality assurance recommendations and

  14. SU-E-T-468: Implementation of the TG-142 QA Process for Seven Linacs with Enhanced Beam Conformance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Woollard, J; Ayan, A; DiCostanzo, D

    2015-06-15

    Purpose: To develop a TG-142 compliant QA process for 7 Varian TrueBeam linear accelerators (linacs) with enhanced beam conformance and dosimetrically matched beam models. To ensure consistent performance of all 7 linacs, the QA process should include a common set of baseline values for use in routine QA on all linacs. Methods: The TG 142 report provides recommended tests, tolerances and frequencies for quality assurance of medical accelerators. Based on the guidance provided in the report, measurement tests were developed to evaluate each of the applicable parameters listed for daily, monthly and annual QA. These tests were then performed onmore » each of our 7 new linacs as they came on line at our institution. Results: The tolerance values specified in TG-142 for each QA test are either absolute tolerances (i.e. ±2mm) or require a comparison to a baseline value. The results of our QA tests were first used to ensure that all 7 linacs were operating within the suggested tolerance values provided in TG −142 for those tests with absolute tolerances and that the performance of the linacs was adequately matched. The QA test results were then used to develop a set of common baseline values for those QA tests that require comparison to a baseline value at routine monthly and annual QA. The procedures and baseline values were incorporated into a spreadsheets for use in monthly and annual QA. Conclusion: We have developed a set of procedures for daily, monthly and annual QA of our linacs that are consistent with the TG-142 report. A common set of baseline values was developed for routine QA tests. The use of this common set of baseline values for comparison at monthly and annual QA will ensure consistent performance of all 7 linacs.« less

  15. A retrospective analysis for patient-specific quality assurance of volumetric-modulated arc therapy plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Li, Guangjun; Wu, Kui; Peng, Guang

    2014-01-01

    Volumetric-modulated arc therapy (VMAT) is now widely used clinically, as it is capable of delivering a highly conformal dose distribution in a short time interval. We retrospectively analyzed patient-specific quality assurance (QA) of VMAT and examined the relationships between the planning parameters and the QA results. A total of 118 clinical VMAT cases underwent pretreatment QA. All plans had 3-dimensional diode array measurements, and 69 also had ion chamber measurements. Dose distribution and isocenter point dose were evaluated by comparing the measurements and the treatment planning system (TPS) calculations. In addition, the relationship between QA results and several planning parameters,more » such as dose level, control points (CPs), monitor units (MUs), average field width, and average leaf travel, were also analyzed. For delivered dose distribution, a gamma analysis passing rate greater than 90% was obtained for all plans and greater than 95% for 100 of 118 plans with the 3%/3-mm criteria. The difference (mean ± standard deviation) between the point doses measured by the ion chamber and those calculated by TPS was 0.9% ± 2.0% for all plans. For all cancer sites, nasopharyngeal carcinoma and gastric cancer have the lowest and highest average passing rates, respectively. From multivariate linear regression analysis, the dose level (p = 0.001) and the average leaf travel (p < 0.001) showed negative correlations with the passing rate, and the average field width (p = 0.003) showed a positive correlation with the passing rate, all indicating a correlation between the passing rate and the plan complexity. No statistically significant correlation was found between MU or CP and the passing rate. Analysis of the results of dosimetric pretreatment measurements as a function of VMAT plan parameters can provide important information to guide the plan parameter setting and optimization in TPS.« less

  16. MO-D-BRB-02: SBRT Treatment Planning and Delivery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yang, Y.

    2016-06-15

    Increased use of SBRT and hypofractionation in radiation oncology practice has posted a number of challenges to medical physicist, ranging from planning, image-guided patient setup and on-treatment monitoring, to quality assurance (QA) and dose delivery. This symposium is designed to provide current knowledge necessary for the safe and efficient implementation of SBRT in various linac platforms, including the emerging digital linacs equipped with high dose rate FFF beams. Issues related to 4D CT, PET and MRI simulations, 3D/4D CBCT guided patient setup, real-time image guidance during SBRT dose delivery using gated/un-gated VMAT/IMRT, and technical advancements in QA of SBRT (inmore » particular, strategies dealing with high dose rate FFF beams) will be addressed. The symposium will help the attendees to gain a comprehensive understanding of the SBRT workflow and facilitate their clinical implementation of the state-of-art imaging and planning techniques. Learning Objectives: Present background knowledge of SBRT, describe essential requirements for safe implementation of SBRT, and discuss issues specific to SBRT treatment planning and QA. Update on the use of multi-dimensional and multi-modality imaging for reliable guidance of SBRT. Discuss treatment planning and QA issues specific to SBRT. Provide a comprehensive overview of emerging digital linacs and summarize the key geometric and dosimetric features of the new generation of linacs for substantially improved SBRT. NIH/NCI; Varian Medical Systems; F. Yin, Duke University has a research agreement with Varian Medical Systems. In addition to research grant, I had a technology license agreement with Varian Medical Systems.« less

  17. Effect of mutations in the qa gene cluster of Neurospora crassa on the enzyme catabolic dehydroquinase.

    PubMed Central

    Jacobson, J W; Hautala, J A; Case, M E; Giles, N H

    1975-01-01

    Catabolic dehydroquinase, which functions in the inducible quinic acid catabolic pathway of Neurospora crassa, has been purified from wild type (74-A) and three mutants in the qa gene cluster. The mutant strains were: 105c, a temperature-sensitive constitutive mutant in the qa-1 regulatory locus; M-16, a qa-3 mutant deficient in quinate dehydrogenase activity; and 237, a leaky qa-2 mutant which possess very low levels of catabolic dehydroquinase activity. The enzymes purified from strains 74-A, 105c, and M-16 are identical with respect to behavior during purification, specific activity, electrophoretic behavior, stability, molecular weight, subunit structure, immunological cross-reactivity, and amino acid content. The mutant enzyme from strain 237 is 1,500-fold less active and appears to have a slightly different amino acid content. It is identical by a number of the other criteria listed above and is presumed to be a mutant at or near the enzyme active site. These data demonstrate that the qa-1 gene product is not involved in the posttranslational expression of enzyme activity. The biochemical identity of catabolic dehydroquinase isolated from strains 105c and M-16 with that from wild type also demonstrates that neither the inducer, quinic acid, nor other enzymes encoded in the qa gene cluster are necessary for the expression of activity. Therefore the combined genetic and biochemical data on the qa system continue to support the hypothesis that the qa-1 regulatory protein acts as a positive initiator of qa enzyme synthesis. Images PMID:126226

  18. PRE-QUALITY ASSURANCE PROJECT PLAN (QAPP) AGREEMENT (PQA) (HANDOUT)

    EPA Science Inventory

    The Land Remediation and Pollution Control Division (LRPCD) QA Manager strives to assist LRPCD researchers in developing functional planning documents for their research projects. As part of the planning process, several pieces of information are needed, including information re...

  19. SU-E-T-460: Impact of the LINAC Repetition Rate On a High-Resolution Liquid Ionization Chamber Array for Patient-Specific QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, S; Driewer, J; Zheng, D

    2015-06-15

    Purpose: The purpose of this study is to investigate the LINAC repetition-rate (dose-rate) dependence of OCTAVIUS 1000SRS liquid ionization chamber (LIC) array for patient specific QA of SRT plans delivered with flattening-filter-free (FFF) beams. Methods: 1) The repetition-rate dependence of 1000SRS was measured in a phantom constructed with 5-cm solid water above and below the array for build-up and backscatter. A 0.3cc calibrated ion chamber was also placed along the central axis 2.3cm below the center chamber of the array for normalizing LINAC output fluctuation. The signals from the center chamber of the array under different repetition rates in themore » range of 400–2400 MU/min for 6xFFF and 10xFFF beams on a Varian TrueBeamSTx LINAC, normalized by the independent chamber readings, were analyzed for the array response dependence on repetition rates. 2) Twelve Step-and-shoot IMRS QA plans (6xFFF and 10xFFF) were delivered to the array under different repetition rates for analysis and comparison. 3) The absolute doses measured by the center chamber were compared to measurements using an independent ionization chamber with the identical setup, taken as the gold standard. 4) The correction factors based on the actual delivery repetition rate were applied to the measurements, and the results were compared again to the gold standard. Results: 1) The 1000SRS array exhibited repetition-rate dependence for FFF beams, up to 5% for 6xFFF and 10% for 10xFFF; 2) The array showed clinically-acceptable repetition-rate dependence for regular flattened beams; 3) This repetition-rate dependence significantly affected the measurement accuracy, thereby affecting IMRS QA results; 4) By applying an empirical repetition-rate correction, the corrected measurements agreed better with the gold standard ion chamber measurements. Conclusion: OCTAVIUS 1000SRS LIC array exhibited considerable repetition-rate dependence for FFF beams, which will affect the accuracy of the absolute QA

  20. On the pulse boiling frequency in thermosyphons

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Liu, J.F.; Wang, J.C.Y.

    1992-02-01

    The unsteady periodic boiling phenomenon, pulse boiling, appearing in the evaporator of thermosyphons has been mentioned and investigated by many researchers. The heat transfer coefficient in evaporators was predicted according to different considerations of flow patterns. For instance, Shiraishi et al. proposed a method based on a combination flow pattern: the nucleate boiling in a liquid pool and the evaporation from a falling condensate film. Liu et al. only considered a pure pulse boiling flow pattern, and Xin et al. focused on the flow pattern of the continuous boiling process without pulse phenomenon. Besides, the forming conditions of pulse boilingmore » were also described differently. Xin et al. also reported that pulse boiling cannot occur in a carbon-steel/water heat pipe; Ma et al., however, observed this phenomenon in a carbon-steel/water thermosyphon. Nearly all researchers mentioned that this phenomenon indeed exists in glass/water thermosyphons. Although the influential factors have been discussed qualitatively, the quantitative analysis has yet to be conducted. This study focuses on the pulse boiling frequency as a criterion for the determination of flow patterns, and attempts are made to predict the frequency both experimentally and theoretically.« less

  1. Noise analysis of nucleate boiling

    NASA Technical Reports Server (NTRS)

    Mcknight, R. D.; Ram, K. S.

    1971-01-01

    The techniques of noise analysis have been utilized to investigate nucleate pool boiling. A simple experimental setup has been developed for obtaining the power spectrum of a nucleate boiling system. These techniques were first used to study single bubbles, and a method of relating the two-dimensional projected size and the local velocity of the bubbles to the auto-correlation functions is presented. This method is much less time consuming than conventional methods of measurement and has no probes to disturb the system. These techniques can be used to determine the contribution of evaporation to total heat flux in nucleate boiling. Also, these techniques can be used to investigate the effect of various parameters upon the frequency response of nucleate boiling. The predominant frequencies of the power spectrum correspond to the frequencies of bubble generation. The effects of heat input, degree of subcooling, and liquid surface tension upon the power spectra of a boiling system are presented. It was found that the degree of subcooling has a more pronounced effect upon bubble size than does heat flux. Also the effect of lowering surface tension can be sufficient to reduce the effect of the degree of subcooling upon the size of the bubbles.

  2. Review of the Constellation Level II Safety, Reliability, and Quality Assurance (SR&QA) Requirements Documents during Participation in the Constellation Level II SR&QA Forum

    NASA Technical Reports Server (NTRS)

    Cameron, Kenneth D.; Gentz, Steven J.; Beil, Robert J.; Minute, Stephen A.; Currie, Nancy J.; Scott, Steven S.; Thomas, Walter B., III; Smiles, Michael D.; Schafer, Charles F.; Null, Cynthia H.; hide

    2009-01-01

    At the request of the Exploration Systems Mission Directorate (ESMD) and the Constellation Program (CxP) Safety, Reliability; and Quality Assurance (SR&QA) Requirements Director, the NASA Engineering and Safety Center (NESC) participated in the Cx SR&QA Requirements forum. The Requirements Forum was held June 24-26; 2008, at GRC's Plum Brook Facility. The forums purpose was to gather all stakeholders into a focused meeting to help complete the process of refining the CxP to refine its Level II SR&QA requirements or defining project-specific requirements tailoring. Element prime contractors had raised specific questions about the wording and intent of many requirements in areas they felt were driving costs without adding commensurate value. NESC was asked to provide an independent and thorough review of requirements that contractors believed were driving Program costs, by active participation in the forum. This document contains information from the forum.

  3. Developing a mailed phantom to implement a local QA program in Egypt radiotherapy centers

    NASA Astrophysics Data System (ADS)

    Soliman, H. A.; Aletreby, M.

    2016-07-01

    In this work, a simple method that differs from the IAEA/WHO Thermoluminescent dosimeters (TLD) postal quality assurance (QA) program is developed. A small perspex; polymethyl methacrylate (PMMA), phantom measured 50 mm × 50 mm × 50 mm is constructed to be used for absorbed dose verification of high-energy photon beams in some major radiotherapy centers in Egypt. The phantom weighted only 140.7 g with two buildup covers weighted 14.8 and 43.19 g for the Cobalt-60 and the 6-MV X-ray beams, respectively. This phantom is aimed for use in the future's external audit/QA services in Egypt for the first time. TLD-700 chips are used for testing and investigating a convenient and national dosimetry QA program. Although the used methodology is comparable to previously introduced but new system; it has smaller size, less weight, and different more available material. Comparison with the previous similar designs is introduced. Theoretical calculations were done by the commercial Eclipse treatment planning system, implementing the pencil beam convolution algorithm to verify the accuracy of the experimental calculation of the dose conversion factor of water to the perspex phantom. The new constructed small phantom and methodology was applied in 10 participating radiotherapy centers. The absorbed dose was verified under the reference conditions for both 60Co and 6-MV high-energy photon beams. The checked beams were within the 5% limit except for four photon beams. There was an agreement of 0.2% between our experimental data and those previously published confirming the validity of the applied method in verifying radiotherapy absorbed dose.

  4. SU-E-T-345: Effect of DLG and MLC Transmission Value Set in the Treatment Planning System (TPS) On Dosimetric Accuracy of True Beam Hypofractionated SRT/SBRT and 2Gy/fx Prostate Rapid Arc Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wu, X; Wang, Y

    Purpose: Due to limited commissioning time, we previously only released our True beam non-FFF mode for prostate treatment. Clinical demand now pushes us to release the non-FFF mode for SRT/SBRT treatment. When re-planning on True beam previously treated SRT/SBRT cases on iX machine we found the patient specific QA pass rate was worse than iX’s, though the 2Gy/fx prostate Result had been as good. We hypothesize that in TPS the True beam DLG and MLC transmission values, of those measured during commissioning could not yet provide accurate SRS/SBRT dosimetry. Hence this work is to investigate how the TPS DLG andmore » transmission value affects Rapid Arc plans’ dosimetric accuracy. Methods: We increased DLG and transmission value of True beam in TPS such that their percentage differences against the measured matched those of iX’s. We re-calculated 2 SRT, 1 SBRT and 2 prostate plans, performed patient specific QA on these new plans and compared the results to the previous. Results: With DLG and transmission value set respectively 40 and 8% higher than the measured, the patient specific QA pass rate (at 3%/3mm) improved from 95.0 to 97.6% vs previous iX’s 97.8% in the case of SRT. In the case of SBRT, the pass rate improved from 75.2 to 93.9% vs previous iX’s 92.5%. In the case of prostate, the pass rate improved from 99.3 to 100%. The maximum dose difference in plans before and after adjusting DLG and transmission was approximately 1% of the prescription dose among all plans. Conclusion: The impact of adjusting DLG and transmission value on dosimetry might be the same among all Rapid Arc plans regardless hypofractionated or not. The large variation observed in patient specific QA pass rate might be due to the data analysis method in the QA software being more sensitive to hypofractionated plans.« less

  5. Quality assurance for online adapted treatment plans: Benchmarking and delivery monitoring simulation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Li, Taoran, E-mail: taoran.li.duke@gmail.com; Wu, Qiuwen; Yang, Yun

    Purpose: An important challenge facing online adaptive radiation therapy is the development of feasible and efficient quality assurance (QA). This project aimed to validate the deliverability of online adapted plans and develop a proof-of-concept online delivery monitoring system for online adaptive radiation therapy QA. Methods: The first part of this project benchmarked automatically online adapted prostate treatment plans using traditional portal dosimetry IMRT QA. The portal dosimetry QA results of online adapted plans were compared to original (unadapted) plans as well as randomly selected prostate IMRT plans from our clinic. In the second part, an online delivery monitoring system wasmore » designed and validated via a simulated treatment with intentional multileaf collimator (MLC) errors. This system was based on inputs from the dynamic machine information (DMI), which continuously reports actual MLC positions and machine monitor units (MUs) at intervals of 50 ms or less during delivery. Based on the DMI, the system performed two levels of monitoring/verification during the delivery: (1) dynamic monitoring of cumulative fluence errors resulting from leaf position deviations and visualization using fluence error maps (FEMs); and (2) verification of MLC positions against the treatment plan for potential errors in MLC motion and data transfer at each control point. Validation of the online delivery monitoring system was performed by introducing intentional systematic MLC errors (ranging from 0.5 to 2 mm) to the DMI files for both leaf banks. These DMI files were analyzed by the proposed system to evaluate the system’s performance in quantifying errors and revealing the source of errors, as well as to understand patterns in the FEMs. In addition, FEMs from 210 actual prostate IMRT beams were analyzed using the proposed system to further validate its ability to catch and identify errors, as well as establish error magnitude baselines for prostate IMRT

  6. Quality Assurance Project Plan Development Tool

    EPA Pesticide Factsheets

    This tool contains information designed to assist in developing a Quality Assurance (QA) Project Plan that meets EPA requirements for projects that involve surface or groundwater monitoring and/or the collection and analysis of water samples.

  7. Fundamental Boiling and RP-1 Freezing Experiments

    NASA Technical Reports Server (NTRS)

    Goode, Brian

    2002-01-01

    The prestart thermal conditioning of the hardware in LOX (liquid oxygen) systems involve heat transfer between LOX and metal where boiling plays a large role. Information is easily found on nucleate boiling, maximum heat flux, minimum heat flux and film boiling for common fluids like water. After looking at these standard correlations it was felt more data was needed for the cool down side transition boiling for the LN2 and LOX. In particular interest is the film boiling values, the temperature at which transition begins and the slope as peak heat flux is approached. The ultimate goal is an array of boiling heat transfer coefficient as a function of surface temperature which can be used in the chilldown model of the feed system, engine and bleed system for X-34. The first experiment consisted of an actual MC-1 LOX Impeller which had been machined backwards, that was instrumented with 17 surface thermocouples and submerged in liquid nitrogen. The thermocouples were installed on metal thicknesses varying from the thin inducer to the thick hub.

  8. Boiling Experiment Facility for Heat Transfer Studies in Microgravity

    NASA Technical Reports Server (NTRS)

    Delombard, Richard; McQuillen, John; Chao, David

    2008-01-01

    Pool boiling in microgravity is an area of both scientific and practical interest. By conducting tests in microgravity, it is possible to assess the effect of buoyancy on the overall boiling process and assess the relative magnitude of effects with regards to other "forces" and phenomena such as Marangoni forces, liquid momentum forces, and microlayer evaporation. The Boiling eXperiment Facility is now being built for the Microgravity Science Glovebox that will use normal perfluorohexane as a test fluid to extend the range of test conditions to include longer test durations and less liquid subcooling. Two experiments, the Microheater Array Boiling Experiment and the Nucleate Pool Boiling eXperiment will use the Boiling eXperiment Facility. The objectives of these studies are to determine the differences in local boiling heat transfer mechanisms in microgravity and normal gravity from nucleate boiling, through critical heat flux and into the transition boiling regime and to examine the bubble nucleation, growth, departure and coalescence processes. Custom-designed heaters will be utilized to achieve these objectives.

  9. SU-F-BRE-08: Feasibility of 3D Printed Patient Specific Phantoms for IMRT/IGRT QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ehler, E; Higgins, P; Dusenbery, K

    Purpose: Test the feasibility of 3D printed, per-patient phantoms for IMRT QA to analyze the treatment delivery quality within the patient geometry. Methods: Using the head and neck region of an anthropomorphic phantom as a substitute for an actual patient, a soft-tissue equivalent model was constructed with the use of a 3D printer. A nine-field IMRT plan was constructed and dose verification measurements were performed for the 3D printed phantom. During the delivery of the IMRT QA on to the 3D printed phantom, the same patient positioning indexing system was used on the phantom and image guidance (cone beam CT)more » was used to localize the phantom, serving as a test of the IGRT system as well. The 3D printed phantom was designed to accommodate four radiochromic film planes (two axial, one coronal and one sagittal) and an ionization chamber measurement. As a frame of comparison, the IMRT QA was also performed on traditional phantoms. Dosimetric tolerance levels such as 3mm / 3% Gamma Index as well as 3% and 5% dose difference were considered. All detector systems were calibrated against a NIST traceable ionization chamber. Results: Comparison of results 3D printed patient phantom with the standard IMRT QA systems showed similar passing rates for the 3D printed phantom and the standard phantoms. However, the locations of the failing regions did not necessarily correlate. The 3D printed phantom was localized within 1 mm and 1° using on-board cone beam CT. Conclusion: A custom phantom was created using a 3D printer. It was determined that the use of patient specific phantoms to perform dosimetric verification and estimate the dose in the patient is feasible. In addition, end-to-end testing on a per-patient basis was possible with the 3D printed phantom. Further refinement of the phantom construction process is needed for routine clinical use.« less

  10. SU-G-BRB-07: Developing a QA Procedure for Gated VMAT SABR Treatments Using 10 MV Beam in Flattening-Filter Free Mode

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chitsazzadeh, S; Wells, D; Mestrovic, A

    2016-06-15

    Purpose: To develop a QA procedure for gated VMAT stereotactic ablative radiotherapy (SABR) treatments. Methods: An interface was constructed to attach the translational stage of a Quasar respiratory motion phantom to a pinpoint ion chamber insert and move the ion chamber inside an ArcCheck diode array. The Quasar phantom controller used a patient specific breathing pattern to translate the ion chamber in a superior-inferior direction inside the ArcCheck. An amplitude-based RPM tracking system was specified to turn the beam on during the exhale phase of the breathing pattern. SABR plans were developed using Eclipse for liver PTVs ranging in sizemore » from 3-12 cm in diameter using a 2-arc VMAT technique. Dose was measured in the middle of the penumbra region, where the high dose gradient allowed for sensitive detection of any inaccuracies in gated dose delivery. The overall fidelity of the dose distribution was confirmed using ArcCheck. The sensitivity of the gating QA procedure was investigated with respect to the following four parameters: PTV size, duration of exhale, baseline drift, and gating window size. Results: The difference between the measured dose to a point in the penumbra and the Eclipse calculated dose was under 2% for small residual motions. The QA procedure was independent of PTV size and duration of exhale. Baseline drift and gating window size, however, significantly affected the penumbral dose measurement, with differences of up to 30% compared to Eclipse. Conclusion: This study described a highly sensitive QA procedure for gated VMAT SABR treatments. The QA outcome was dependent on the gating window size and baseline drift. Analysis of additional patient breathing patterns will be required to determine a clinically relevant gating window size and an appropriate tolerance level for this procedure.« less

  11. Pool and flow boiling in variable and microgravity

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.

    1994-01-01

    As is well known, boiling is an effective mode of heat transfer in that high heat flux levels are possible with relatively small temperature differences. Its optimal application requires that the process be adequately understood. A measure of the understanding of any physical event lies in the ability to predict its behavior in terms of the relevant parameters. Despite many years of research the predictability of boiling is currently possible only for quite specialized circumstances, e.g., the critical heat flux and film boiling for the pool boiling case, and then only with special geometries. Variable gravity down to microgravity provides the opportunity to test this understanding, but possibly more important, by changing the dimensional and time scales involved permits more detailed observations of elements involved in the boiling process, and perhaps discloses phenomena heretofore unknown. The focus here is on nucleate boiling although, as will be demonstrated below, under but certain circumstances in microgravity it can take place concurrently with the dryout process. In the presence of earth gravity or forced convection effects, the latter process is usually referred to as film boiling. However, no vapor film as such forms with pool boiling in microgravity, only dryout. Initial results are presented here for pool boiling in microgravity, and were made possible at such an early date by the availability of the Get-Away-Specials (GAS). Also presented here are some results of ground testing of a flow loop for the study of low velocity boiling, eventually to take place also in microgravity. In the interim, variable buoyancy normal to the heater surface is achieved by rotation of the entire loop relative to earth gravity. Of course, this is at the expense of varying the buoyancy parallel to the heater surface. Two questions which must be resolved early in the study of flow boiling in microgravity are (1) the lower limits of liquid flow velocity where buoyancy

  12. A retrospective analysis for patient-specific quality assurance of volumetric-modulated arc therapy plans.

    PubMed

    Li, Guangjun; Wu, Kui; Peng, Guang; Zhang, Yingjie; Bai, Sen

    2014-01-01

    Volumetric-modulated arc therapy (VMAT) is now widely used clinically, as it is capable of delivering a highly conformal dose distribution in a short time interval. We retrospectively analyzed patient-specific quality assurance (QA) of VMAT and examined the relationships between the planning parameters and the QA results. A total of 118 clinical VMAT cases underwent pretreatment QA. All plans had 3-dimensional diode array measurements, and 69 also had ion chamber measurements. Dose distribution and isocenter point dose were evaluated by comparing the measurements and the treatment planning system (TPS) calculations. In addition, the relationship between QA results and several planning parameters, such as dose level, control points (CPs), monitor units (MUs), average field width, and average leaf travel, were also analyzed. For delivered dose distribution, a gamma analysis passing rate greater than 90% was obtained for all plans and greater than 95% for 100 of 118 plans with the 3%/3-mm criteria. The difference (mean ± standard deviation) between the point doses measured by the ion chamber and those calculated by TPS was 0.9% ± 2.0% for all plans. For all cancer sites, nasopharyngeal carcinoma and gastric cancer have the lowest and highest average passing rates, respectively. From multivariate linear regression analysis, the dose level (p = 0.001) and the average leaf travel (p < 0.001) showed negative correlations with the passing rate, and the average field width (p = 0.003) showed a positive correlation with the passing rate, all indicating a correlation between the passing rate and the plan complexity. No statistically significant correlation was found between MU or CP and the passing rate. Analysis of the results of dosimetric pretreatment measurements as a function of VMAT plan parameters can provide important information to guide the plan parameter setting and optimization in TPS. Copyright © 2014 American Association of Medical Dosimetrists. Published by

  13. Characterizing preferential groundwater discharge through boils using temperature

    NASA Astrophysics Data System (ADS)

    Vandenbohede, A.; de Louw, P. G. B.; Doornenbal, P. J.

    2014-03-01

    In The Netherlands, preferential groundwater discharge trough boils is a key process in the salinization of deep polders. Previous work showed that boils also influence the temperature in the subsurface and of surface water. This paper elaborates on this process combining field observations with numerical modeling. As is the case for salinity, a distinct anomaly in the subsurface and surface water temperature can be attributed to boils. Lines of equal temperature are distorted towards the boil, which can be considered as an upconing of the temperature profile by analogy of the upconing of a fresh-saltwater interface. The zone of this distortion is limited to the immediate vicinity of the boil, being about 5 m in the aquitard which holds the boil's conduit, or maximum a few dozens of meters in the underlying aquifer. In the aquitard, heat transport is conduction dominated whereas this is convection dominated in the aquifer. The temperature anomaly differs from the salinity anomaly by the smaller radius of influence and faster time to reach a new steady-state of the former. Boils discharge water with a temperature equal to the mean groundwater temperature. This influences the yearly and diurnal variation of ditch water temperature in the immediate vicinity of the boil importantly but also the temperature in the downstream direction. Temporary nature of the boil (e.g. stability of the conduit, discharge rate), uncertainty on the 3D construction of the conduit and heterogeneity of the subsoil make it unlikely that temperature measurements can be interpreted further than a qualitative level.

  14. Structural changes of malt proteins during boiling.

    PubMed

    Jin, Bei; Li, Lin; Liu, Guo-Qin; Li, Bing; Zhu, Yu-Kui; Liao, Liao-Ning

    2009-03-09

    Changes in the physicochemical properties and structure of proteins derived from two malt varieties (Baudin and Guangmai) during wort boiling were investigated by differential scanning calorimetry, SDS-PAGE, two-dimensional electrophoresis, gel filtration chromatography and circular dichroism spectroscopy. The results showed that both protein content and amino acid composition changed only slightly during boiling, and that boiling might cause a gradual unfolding of protein structures, as indicated by the decrease in surface hydrophobicity and free sulfhydryl content and enthalpy value, as well as reduced alpha-helix contents and markedly increased random coil contents. It was also found that major component of both worts was a boiling-resistant protein with a molecular mass of 40 kDa, and that according to the two-dimensional electrophoresis and SE-HPLC analyses, a small amount of soluble aggregates might be formed via hydrophobic interactions. It was thus concluded that changes of protein structure caused by boiling that might influence beer quality are largely independent of malt variety.

  15. Physical and biological pretreatment quality assurance of the head and neck cancer plan with the volumetric modulated arc therapy

    NASA Astrophysics Data System (ADS)

    Park, So-Hyun; Lee, Dong-Soo; Lee, Yun-Hee; Lee, Seu-Ran; Kim, Min-Ju; Suh, Tae-Suk

    2015-09-01

    The aim of this work is to demonstrate both the physical and the biological quality assurance (QA) aspects as pretreatment QA of the head and neck (H&N) cancer plan for the volumetric modulated arc therapy (VMAT). Ten H&N plans were studied. The COMPASS® dosimetry analysis system and the tumor control probability (TCP) and the normal tissue complication probability (NTCP) calculation free program were used as the respective measurement and calculation tools. The reliability of these tools was verified by a benchmark study in accordance with the TG-166 report. For the physical component of QA, the gamma passing rates and the false negative cases between the calculated and the measured data were evaluated. The biological component of QA was performed based on the equivalent uniform dose (EUD), TCP and NTCP values. The evaluation was performed for the planning target volumes (PTVs) and the organs at risks (OARs), including the eyes, the lens, the parotid glands, the esophagus, the spinal cord, and the brainstem. All cases had gamma passing rates above 95% at an acceptance tolerance level with the 3%/3 mm criteria. In addition, the false negative instances were presented for the PTVs and OARs. The gamma passing rates exhibited a weak correlation with false negative cases. For the biological QA, the physical dose errors affect the EUD and the TCP for the PTVs, but no linear correlation existed between them. The EUD and NTCP for the OARs were shown the random differences that could not be attributed to the dose errors from the physical QA. The differences in the EUD and NTCP between the calculated and the measured results were mainly demonstrated for the parotid glands. This study describes the importance and the necessity of improved QA to accompany both the physical and the biological aspects for accurate radiation treatment.

  16. DeepQA: improving the estimation of single protein model quality with deep belief networks.

    PubMed

    Cao, Renzhi; Bhattacharya, Debswapna; Hou, Jie; Cheng, Jianlin

    2016-12-05

    Protein quality assessment (QA) useful for ranking and selecting protein models has long been viewed as one of the major challenges for protein tertiary structure prediction. Especially, estimating the quality of a single protein model, which is important for selecting a few good models out of a large model pool consisting of mostly low-quality models, is still a largely unsolved problem. We introduce a novel single-model quality assessment method DeepQA based on deep belief network that utilizes a number of selected features describing the quality of a model from different perspectives, such as energy, physio-chemical characteristics, and structural information. The deep belief network is trained on several large datasets consisting of models from the Critical Assessment of Protein Structure Prediction (CASP) experiments, several publicly available datasets, and models generated by our in-house ab initio method. Our experiments demonstrate that deep belief network has better performance compared to Support Vector Machines and Neural Networks on the protein model quality assessment problem, and our method DeepQA achieves the state-of-the-art performance on CASP11 dataset. It also outperformed two well-established methods in selecting good outlier models from a large set of models of mostly low quality generated by ab initio modeling methods. DeepQA is a useful deep learning tool for protein single model quality assessment and protein structure prediction. The source code, executable, document and training/test datasets of DeepQA for Linux is freely available to non-commercial users at http://cactus.rnet.missouri.edu/DeepQA/ .

  17. A review on boiling heat transfer enhancement with nanofluids

    PubMed Central

    2011-01-01

    There has been increasing interest of late in nanofluid boiling and its use in heat transfer enhancement. This article covers recent advances in the last decade by researchers in both pool boiling and convective boiling applications, with nanofluids as the working fluid. The available data in the literature is reviewed in terms of enhancements, and degradations in the nucleate boiling heat transfer and critical heat flux. Conflicting data have been presented in the literature on the effect that nanofluids have on the boiling heat-transfer coefficient; however, almost all researchers have noted an enhancement in the critical heat flux during nanofluid boiling. Several researchers have observed nanoparticle deposition at the heater surface, which they have related back to the critical heat flux enhancement. PMID:21711794

  18. Transition boiling heat transfer and the film transition regime

    NASA Technical Reports Server (NTRS)

    Ramilison, J. M.; Lienhard, J. H.

    1987-01-01

    The Berenson (1960) flat-plate transition-boiling experiment has been recreated with a reduced thermal resistance in the heater, and an improved access to those portions of the transition boiling regime that have a steep negative slope. Tests have been made in Freon-113, acetone, benzene, and n-pentane boiling on horizontal flat copper heaters that have been mirror-polished, 'roughened', or teflon-coated. The resulting data reproduce and clarify certain features observed by Berenson: the modest surface finish dependence of boiling burnout, and the influence of surface chemistry on both the minimum heat flux and the mode of transition boiling, for example. A rational scheme of correlation yields a prediction of the heat flux in what Witte and Lienhard (1982) previously identified as the 'film-transition boiling' region. It is also shown how to calculate the heat flux at the boundary between the pure-film, and the film-transition, boiling regimes, as a function of the advancing contact angle.

  19. Web Implementation of Quality Assurance (QA) for X-ray Units in Balkanic Medical Institutions.

    PubMed

    Urošević, Vlade; Ristić, Olga; Milošević, Danijela; Košutić, Duško

    2015-08-01

    Diagnostic radiology is the major contributor to the total dose of the population from all artificial sources. In order to reduce radiation exposure and optimize diagnostic x-ray image quality, it is necessary to increase the quality and efficiency of quality assurance (QA) and audit programs. This work presents a web application providing completely new QA solutions for x-ray modalities and facilities. The software gives complete online information (using European standards) with which the corresponding institutions and individuals can evaluate and control a facility's Radiation Safety and QA program. The software enables storage of all data in one place and sharing the same information (data), regardless of whether the measured data is used by an individual user or by an authorized institution. The software overcomes the distance and time separation of institutions and individuals who take part in QA. Upgrading the software will enable assessment of the medical exposure level to ionizing radiation.

  20. TU-FG-201-01: 18-Month Clinical Experience of a Linac Daily Quality Assurance (QA) Solution Using Only EPID and OBI

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cai, B; Sun, B; Yaddanapudi, S

    Purpose: To describe the clinical use of a Linear Accelerator (Linac) DailyQA system with only EPID and OBI. To assess the reliability over an 18-month period and improve the robustness of this system based on QA failure analysis. Methods: A DailyQA solution utilizing an in-house designed phantom, combined EPID and OBI image acquisitions, and a web-based data analysis and reporting system was commissioned and used in our clinic to measure geometric, dosimetry and imaging components of a Varian Truebeam Linac. During an 18-month period (335 working days), the Daily QA results, including the output constancy, beam flatness and symmetry, uniformity,more » TPR20/10, MV and KV imaging quality, were collected and analyzed. For output constancy measurement, an independent monthly QA system with an ionization chamber (IC) and annual/incidental TG51 measurements with ADCL IC were performed and cross-compared to Daily QA system. Thorough analyses were performed on the recorded QA failures to evaluate the machine performance, optimize the data analysis algorithm, adjust the tolerance setting and improve the training procedure to prevent future failures. Results: A clinical workflow including beam delivery, data analysis, QA report generation and physics approval was established and optimized to suit daily clinical operation. The output tests over the 335 working day period cross-correlated with the monthly QA system within 1.3% and TG51 results within 1%. QA passed with one attempt on 236 days out of 335 days. Based on the QA failures analysis, the Gamma criteria is revised from (1%, 1mm) to (2%, 1mm) considering both QA accuracy and efficiency. Data analysis algorithm is improved to handle multiple entries for a repeating test. Conclusion: We described our 18-month clinical experience on a novel DailyQA system using only EPID and OBI. The long term data presented demonstrated the system is suitable and reliable for Linac daily QA.« less

  1. A Strategy to Establish a Quality Assurance/Quality Control Plan for the Application of Biosensors for the Detection of E. coli in Water.

    PubMed

    Hesari, Nikou; Kıratlı Yılmazçoban, Nursel; Elzein, Mohamad; Alum, Absar; Abbaszadegan, Morteza

    2017-01-03

    Rapid bacterial detection using biosensors is a novel approach for microbiological testing applications. Validation of such methods is an obstacle in the adoption of new bio-sensing technologies for water testing. Therefore, establishing a quality assurance and quality control (QA/QC) plan is essential to demonstrate accuracy and reliability of the biosensor method for the detection of E. coli in drinking water samples. In this study, different reagents and assay conditions including temperatures, holding time, E. coli strains and concentrations, dissolving agents, salinity and pH effects, quality of substrates of various suppliers of 4-methylumbelliferyl glucuronide (MUG), and environmental water samples were included in the QA/QC plan and used in the assay optimization and documentation. Furthermore, the procedural QA/QC for the monitoring of drinking water samples was established to validate the performance of the biosensor platform for the detection of E. coli using a culture-based standard technique. Implementing the developed QA/QC plan, the same level of precision and accuracy was achieved using both the standard and the biosensor methods. The established procedural QA/QC for the biosensor will provide a reliable tool for a near real-time monitoring of E. coli in drinking water samples to both industry and regulatory authorities.

  2. A Strategy to Establish a Quality Assurance/Quality Control Plan for the Application of Biosensors for the Detection of E. coli in Water

    PubMed Central

    Hesari, Nikou; Kıratlı Yılmazçoban, Nursel; Elzein, Mohamad; Alum, Absar; Abbaszadegan, Morteza

    2017-01-01

    Rapid bacterial detection using biosensors is a novel approach for microbiological testing applications. Validation of such methods is an obstacle in the adoption of new bio-sensing technologies for water testing. Therefore, establishing a quality assurance and quality control (QA/QC) plan is essential to demonstrate accuracy and reliability of the biosensor method for the detection of E. coli in drinking water samples. In this study, different reagents and assay conditions including temperatures, holding time, E. coli strains and concentrations, dissolving agents, salinity and pH effects, quality of substrates of various suppliers of 4-methylumbelliferyl glucuronide (MUG), and environmental water samples were included in the QA/QC plan and used in the assay optimization and documentation. Furthermore, the procedural QA/QC for the monitoring of drinking water samples was established to validate the performance of the biosensor platform for the detection of E. coli using a culture-based standard technique. Implementing the developed QA/QC plan, the same level of precision and accuracy was achieved using both the standard and the biosensor methods. The established procedural QA/QC for the biosensor will provide a reliable tool for a near real-time monitoring of E. coli in drinking water samples to both industry and regulatory authorities. PMID:28054956

  3. 75 FR 2920 - In the Matter of the Designation of al-Qa'ida in the Arabian Peninsula (AQAP), Also Known as al...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-01-19

    ... Known as Tanzim Qa'idat al-Jihad fi Jazirat al- Arab, Also Known as al-Qa'ida Organization in the... Tanzim Qa'idat al- Jihad fi Jazirat al-Arab, also known as al-Qa'ida Organization in the Arabian...

  4. Water boiling inside carbon nanotubes: toward efficient drug release.

    PubMed

    Chaban, Vitaly V; Prezhdo, Oleg V

    2011-07-26

    We show using molecular dynamics simulation that spatial confinement of water inside carbon nanotubes (CNTs) substantially increases its boiling temperature and that a small temperature growth above the boiling point dramatically raises the inside pressure. Capillary theory successfully predicts the boiling point elevation down to 2 nm, below which large deviations between the theory and atomistic simulation take place. Water behaves qualitatively different inside narrow CNTs, exhibiting transition into an unusual phase, where pressure is gas-like and grows linearly with temperature, while the diffusion constant is temperature-independent. Precise control over boiling by CNT diameter, together with the rapid growth of inside pressure above the boiling point, suggests a novel drug delivery protocol. Polar drug molecules are packaged inside CNTs; the latter are delivered into living tissues and heated by laser. Solvent boiling facilitates drug release.

  5. Size Effect of the 2-D Bodies on the Geothermal Gradient and Q-A Plot

    NASA Astrophysics Data System (ADS)

    Thakur, M.; Blackwell, D. D.

    2009-12-01

    Using numerical models we have investigated some of the criticisms on the Q-A plot of related to the effect of size of the body on the slope and reduced heat flow. The effects of horizontal conduction depend on the relative difference of radioactivity between the body and the country rock (assuming constant thermal conductivity). Horizontal heat transfer due to different 2-D bodies was numerically studied in order to quantify resulting temperature differences at the Moho and errors on the predication of Qr (reduced heat flow). Using the two end member distributions of radioactivity, the step model (thickness 10km) and exponential model, different 2-D models of horizontal scale (width) ranging from 10 -500 km were investigated. Increasing the horizontal size of the body tends to move observations closer towards the 1-D solution. A temperature difference of 50 oC is produced (for the step model) at Moho between models of width 10 km versus 500 km. In other words the 1-D solution effectively provides large scale averaging in terms of heat flow and temperature field in the lithosphere. For bodies’ ≤ 100 km wide the geotherms at shallower levels are affected, but at depth they converge and are 50 oC lower than that of the infinite plate model temperature. In case of 2-D bodies surface heat flow is decreased due to horizontal transfer of heat, which will shift the Q-A point vertically downward on the Q-A plot. The smaller the size of the body, the more will be the deviation from the 1-D solution and the more will be the movement of Q-A point downwards on a Q-A plot. On the Q-A plot, a limited points of bodies of different sizes with different radioactivity contrast (for the step and exponential model), exactly reproduce the reduced heat flow Qr. Thus the size of the body can affect the slope on a Q-A plot but Qr is not changed. Therefore, Qr ~ 32 mWm-2 obtained from the global terrain average Q-A plot represents the best estimate of stable continental mantle heat

  6. Boiling on Microconfigured Composite Surfaces Enhanced

    NASA Technical Reports Server (NTRS)

    Chao, David F.

    2000-01-01

    Boiling heat transfer is one of the key technologies for the two-phase active thermal-control system used on space platforms, as well as for the dynamic power systems aboard the International Space Station. Because it is an effective heat transfer mode, boiling is integral to many space applications, such as heat exchangers and other cooling devices. Nucleate boiling near the critical heat flux (CHF) can transport very large thermal loads with a much smaller device and much lower pumping power than for single-phase heat exchangers. However, boiling performance sharply deteriorates in a reduced-gravity environment, and operation in the CHF regime is somewhat perilous because of the risk of burnout to the device surface. New materials called microconfigured metal-graphite composites can enhance boiling. The photomicrograph shows the microconfiguration (x3000) of the copper-graphite (Cu-Gr) surface as viewed by scanning electronic microscope. The graphite fiber tips appear as plateaus with rugged surfaces embedded in the copper matrix. It has been experimentally demonstrated that this type of material manifests excellent boiling heat transfer performance characteristics and an increased CHF. Nonisothermal surfaces were less sensitive to variations of wall superheat in the CHF regime. Because of the great difference in conductivity between the copper base and the graphite fiber, the composite surfaces have a nonisothermal surface characteristic and, therefore, will have a much larger "safe" operating region in the CHF regime. In addition, the thermocapillary forces induced by the temperature differences between the fiber tips and the metal matrix play an important role in bubble detachment, and may not be adversely affected in a reduced-gravity environment. All these factors indicate that microconfigured composites may improve the reliability and economy (dominant factors in all space applications) of various thermal components found on spacecraft during future

  7. EPA Guidance for Geospatially Related Quality Assurance Project Plans

    EPA Pesticide Factsheets

    This March 2003 document discusses EPA's Quality Assurance (QA) Project Plan as a tool for project managers and planners to document the type and quality of data and information needed for making environmental decisions

  8. SU-E-T-453: A Novel Daily QA System for Robotic Image Guided Radiosurgery with Variable Aperture Collimator

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, L; Nelson, B

    Purpose: A novel end-to-end system using a CCD camera and a scintillator based phantom that is capable of measuring the beam-by-beam delivery accuracy of Robotic Radiosurgery has been developed and reported in our previous work. This work investigates its application to end-to-end type daily QA for Robotic Radiosurgery (Cyberknife) with Variable Aperture Collimator (Iris). Methods: The phantom was first scanned with a CT scanner at 0.625 slice thickness and exported to the Cyberknife Muliplan (v4.6) treatment planning system. An isocentric treatment plan was created consisting of ten beams of 25 Monitor Units each using Iris apertures of 7.5, 10, 15,more » 20, and 25 mm. The plan was delivered six times in two days on the Cyberknife G4 system with fiducial tracking on the four metal fiducials embedded in phantom with re-positioning between the measurements. The beam vectors (X, Y, Z) are measured and compared with the plan from the machine delivery file (XML file). The Iris apertures (FWHM) were measured from the beam flux map and compared with the commissioning data. Results: The average beam positioning accuracies of the six deliveries are 0.71 ± 0.40 mm, 0.72 ± 0.44 mm, 0.74 ± 0.42 mm, 0.70 ± 0.40 mm, 0.79 ± 0.44 mm and 0.69 ± 0.41 mm respectively. Radiation beam width (FWHM) variations are within ±0.05 mm, and they agree with the commissioning data within 0.22 mm. The delivery time for the plan is about 7 minutes and the results are given instantly. Conclusion: The experimental results agree with stated sub-millimeter delivery accuracy of Cyberknife system. Beam FWHM variations comply with the 0.2 mm accuracy of the Iris collimator at SAD. The XRV-100 system has proven to be a powerful tool in performing end-to-end type tests for Robotic Image Guided Radiosurgery Daily QA.« less

  9. BOILING HEAT TRANSFER IN ZERO GRAVITY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zara, E.A.

    1964-01-01

    The preliminary results of a research program to determine the effects of zero and near zero gravity on boiling heat transfer are presented. Zero gravity conditions were obtained on the ASD KC-135 zero gravity test aircraft, capable of providing 30-seconds of zero gravity. Results of the program to date indicate that nucleate (bubble) boiling heat transfer rates are not greatly affected by the absence of gravity forces. However, radical pressure increases were observed that will dictate special design considerations to space vehicle systems utilizing pool boiling processes, such as cryogenic or other fluid storage vessels where thermal input to themore » fluid is used for vessel pressurization. (auth)« less

  10. PCB Analysis Plan for Tank Archive Samples

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NGUYEN, D.M.

    2001-03-22

    This analysis plan specifies laboratory analysis, quality assurance/quality control (QA/QC), and data reporting requirements for analyzing polychlorinated biphenyls (PCB) concentrations in archive samples. Tank waste archive samples that are planned for PCB analysis are identified in Nguyen 2001. The tanks and samples are summarized in Table 1-1. The analytical data will be used to establish a PCB baseline inventory in Hanford tanks.

  11. Enhancements of Nucleate Boiling Under Microgravity Conditions

    NASA Technical Reports Server (NTRS)

    Zhang, Nengli; Chao, David F.; Yang, W. J.

    2000-01-01

    This paper presents two means for enhancing nucleate boiling and critical heat flux under microgravity conditions: using micro-configured metal-graphite composites as the boiling surface and dilute aqueous solutions of long-chain alcohols as the working fluid. In the former, thermocapillary force induced by temperature difference between the graphite-fiber tips and the metal matrix plays an important role in bubble detachment. Thus boiling-heat transfer performance does not deteriorate in a reduced-gravity environment. In the latter cases, the surface tension-temperature gradient of the long-chain alcohol solutions turns positive as the temperature exceeds a certain value. Consequently, the Marangoni effect does not impede, but rather aids in bubble departure from the heating surface. This feature is most favorable in microgravity. As a result, the bubble size of departure is substantially reduced at higher frequencies. Based on the existing experimental data, and a two-tier theoretical model, correlation formulas are derived for nucleate boiling on the copper-graphite and aluminum-graphite composite surfaces, in both the isolated and coalesced bubble regimes. In addition, performance equations for nucleate boiling and critical heat flux in dilute aqueous solutions of long-chain alcohols are obtained.

  12. Sci—Fri PM: Dosimetry—06: Commissioning of a 3D patient specific QA system for hypofractionated prostate treatments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rivest, R; Venkataraman, S; McCurdy, B

    The objective of this work is to commission the 6MV-SRS beam model in COMPASS (v2.1, IBA-Dosimetry) and validate its use for patient specific QA of hypofractionated prostate treatments. The COMPASS system consists of a 2D ion chamber array (MatriXX{sup Evolution}), an independent gantry angle sensor and associated software. The system can either directly calculate or reconstruct (using measured detector responses) a 3D dose distribution on the patient CT dataset for plan verification. Beam models are developed and commissioned in the same manner as a beam model is commissioned in a standard treatment planning system. Model validation was initially performed bymore » comparing both COMPASS calculations and reconstructions to measured open field beam data. Next, 10 hypofractionated prostate RapidArc plans were delivered to both the COMPASS system and a phantom with ion chamber and film inserted. COMPASS dose distributions calculated and reconstructed on the phantom CT dataset were compared to the chamber and film measurements. The mean (± standard deviation) difference between COMPASS reconstructed dose and ion chamber measurement was 1.4 ± 1.0%. The maximum discrepancy was 2.6%. Corresponding values for COMPASS calculation were 0.9 ± 0.9% and 2.6%, respectively. The average gamma agreement index (3%/3mm) for COMPAS reconstruction and film was 96.7% and 95.3% when using 70% and 20% dose thresholds, respectively. The corresponding values for COMPASS calculation were 97.1% and 97.1%, respectively. Based on our results, COMPASS can be used for the patient specific QA of hypofractionated prostate treatments delivered with the 6MV-SRS beam.« less

  13. Boils and Carbuncles

    MedlinePlus

    ... body. The spreading infection, commonly known as blood poisoning (sepsis), can lead to infections deep within your body, such as your heart (endocarditis) and bone (osteomyelitis). Prevention It's not always possible to prevent boils, especially ...

  14. SRT and SBRT: Current practices for QA dosimetry and 3D

    NASA Astrophysics Data System (ADS)

    Benedict, S. H.; Cai, J.; Libby, B.; Lovelock, M.; Schlesinger, D.; Sheng, K.; Yang, W.

    2010-11-01

    The major feature that separates stereotactic radiation therapy (cranial SRT) and stereotactic body radiation therapy (SBRT) from conventional radiation treatment is the delivery of large doses in a few fractions which results in a high biological effective dose (BED). In order to minimize the normal tissue toxicity, quality assurance of the conformation of high doses to the target and rapid fall off doses away from the target is critical. The practice of SRT and SBRT therefore requires a high-level of confidence in the accuracy of the entire treatment delivery process. In SRT and SBRT confidence in this accuracy is accomplished by the integration of modern imaging, simulation, treatment planning and delivery technologies into all phases of the treatment process; from treatment simulation and planning and continuing throughout beam delivery. In this report some of the findings of Task group 101 of the AAPM will be presented which outlines the best-practice guidelines for SBRT. The task group report includes a review of the literature to identify reported clinical findings and expected outcomes for this treatment modality. Information in this task group is provided for establishing an SBRT program, including protocols, equipment, resources, and QA procedures.

  15. Conversion of direct process high-boiling residue to monosilanes

    DOEpatents

    Brinson, Jonathan Ashley; Crum, Bruce Robert; Jarvis, Jr., Robert Frank

    2000-01-01

    A process for the production of monosilanes from the high-boiling residue resulting from the reaction of hydrogen chloride with silicon metalloid in a process typically referred to as the "direct process." The process comprises contacting a high-boiling residue resulting from the reaction of hydrogen chloride and silicon metalloid, with hydrogen gas in the presence of a catalytic amount of aluminum trichloride effective in promoting conversion of the high-boiling residue to monosilanes. The present process results in conversion of the high-boiling residue to monosilanes. At least a portion of the aluminum trichloride catalyst required for conduct of the process may be formed in situ during conduct of the direct process and isolation of the high-boiling residue.

  16. A fundamental study of nucleate pool boiling under microgravity

    NASA Technical Reports Server (NTRS)

    Ervin, Jamie S.; Merte, Herman, Jr.

    1991-01-01

    An experimental study of incipient boiling in short-term microgravity and with a/g = +/- 1 for pool boiling was performed. Calibrated thin gold films sputtered on a smoothly polished quartz surface were used simultaneously for thermal resistance measurements and heating of the boiling surface. The gold films were used for both transient and quasi-steady heating surface temperature measurements. Two test vessels were constructed for precise measurement and control of fluid temperature and pressure: a laboratory pool boiling vessel for the a/g = +/- experiments and a pool boiling vessel designed for the 131 m free-fall in the NASA Lewis Research Center Microgravity Research Facility for the microgravity tests. Measurements included the heater surface temperature, the pressure near the heating surface, and the bulk liquid temperatures. High speed photography was used in the experiments. With high quality microgravity and the measured initial temperature of the quiescent test fluid, R113, the temperature distribution in the liquid at the moment of boiling inception resulting from an imposed step in heat flux is known with a certainty not possible previously. The types of boiling propagation across the large flat heating surface are categorized; the conditions necessary for their occurrence are described. Explosive boiling propagation with a striking pattern of small scale protuberances over the entire vapor mass periphery not observed previously at low heat flux levels is described. For the heater surface with a/g = -1, a step in the heater surface temperature of short duration was imposed. The resulting liquid temperature distribution at the moment of boiling inception was different from that obtained with a step in heat flux.

  17. 40 CFR 98.74 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Ammonia Manufacturing § 98.74 Monitoring and QA/QC... (c)(8) of this section. (f) [Reserved] (g) If CO2 from ammonia production is used to produce urea at...

  18. 40 CFR 98.74 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Ammonia Manufacturing § 98.74 Monitoring and QA/QC... (c)(8) of this section. (f)[Reserved] (g) If CO2 from ammonia production is used to produce urea at...

  19. 40 CFR 98.74 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Ammonia Manufacturing § 98.74 Monitoring and QA/QC... (c)(8) of this section. (f)[Reserved] (g) If CO2 from ammonia production is used to produce urea at...

  20. 40 CFR 98.74 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Ammonia Manufacturing § 98.74 Monitoring and QA/QC... (c)(8) of this section. (f) [Reserved] (g) If CO2 from ammonia production is used to produce urea at...

  1. SU-E-T-205: Improving Quality Assurance of HDR Brachytherapy: Verifying Agreement Between Planned and Delivered Dose Distributions Using DICOM RTDose and Advanced Film Dosimetry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Palmer, A L; University of Surrey, Guildford, Surrey; Bradley, D A

    Purpose: HDR brachytherapy is undergoing significant development, and quality assurance (QA) checks must keep pace. Current recommendations do not adequately verify delivered against planned dose distributions: This is particularly relevant for new treatment planning system (TPS) calculation algorithms (non TG-43 based), and an era of significant patient-specific plan optimisation. Full system checks are desirable in modern QA recommendations, complementary to device-centric individual tests. We present a QA system incorporating TPS calculation, dose distribution export, HDR unit performance, and dose distribution measurement. Such an approach, more common in external beam radiotherapy, has not previously been reported in the literature for brachytherapy.more » Methods: Our QA method was tested at 24 UK brachytherapy centres. As a novel approach, we used the TPS DICOM RTDose file export to compare planned dose distribution with that measured using Gafchromic EBT3 films placed around clinical brachytherapy treatment applicators. Gamma analysis was used to compare the dose distributions. Dose difference and distance to agreement were determined at prescription Point A. Accurate film dosimetry was achieved using a glass compression plate at scanning to ensure physically-flat films, simultaneous scanning of known dose films with measurement films, and triple-channel dosimetric analysis. Results: The mean gamma pass rate of RTDose compared to film-measured dose distributions was 98.1% at 3%(local), 2 mm criteria. The mean dose difference, measured to planned, at Point A was -0.5% for plastic treatment applicators and -2.4% for metal applicators, due to shielding not accounted for in TPS. The mean distance to agreement was 0.6 mm. Conclusion: It is recommended to develop brachytherapy QA to include full-system verification of agreement between planned and delivered dose distributions. This is a novel approach for HDR brachytherapy QA. A methodology using advanced film

  2. Explosive Boiling at Very Low Heat Fluxes: A Microgravity Phenomenon

    NASA Technical Reports Server (NTRS)

    Hasan, M. M.; Lin, C. S.; Knoll, R. H.; Bentz, M. D.

    1993-01-01

    The paper presents experimental observations of explosive boiling from a large (relative to bubble sizes) flat heating surface at very low heat fluxes in microgravity. The explosive boiling is characterized as either a rapid growth of vapor mass over the entire heating surface due to the flashing of superheated liquid or a violent boiling spread following the appearance of single bubbles on the heating surface. Pool boiling data with saturated Freon 113 was obtained in the microgravity environment of the space shuttle. The unique features of the experimental results are the sustainability of high liquid superheat for long periods and the occurrence of explosive boiling at low heat fluxes (0.2 to 1.2 kW/sq m). For a heat flux of 1.0 kW/sq m a wall superheat of 17.9 degrees C was attained in ten minutes of heating. This was followed by an explosive boiling accompanied with a pressure spike and a violent bulk liquid motion. However, at this heat flux the vapor blanketing the heating surface could not be sustained. Stable nucleate boiling continued following the explosive boiling.

  3. SU-F-T-559: High-Resolution Scintillating Fiber Array for In-Vivo Real-Time SRS and SBRT Patient QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Knewtson, T; Pokhrel, S; University of Tennessee Health Science Center, Memphis, TN

    2016-06-15

    Purpose: A high-resolution scintillating fiber detector was built for in-vivo real-time patient specific quality assurance (QA). The detector is designed for stereotactic body radiotherapy (SBRT) and stereotactic radiosurgery (SRS) to monitor treatment delivery and detect real-time deviations from planned dose to increase patient safety and treatment accuracy. Methods: The detector consists of two high-density scintillating fiber arrays layered to form an X-Y grid which can be attached to the accessory tray of a medical linac for SBRT and cone SRS treatment QA. Fiber arrays consist of 128 scintillating fibers embedded within a precision-machined, high-transmission polymer substrate with 0.8mm pitch. Themore » fibers are coupled on both ends to high-sensitivity photodetectors and the output is recorded through a high-speed analog-to-digital converter to capture the linac pulse sequence as treatment delivery progresses. The detector has a software controlled 360 degree rotational system to capture angular beam projections for high-resolution beam profile reconstruction. Results: The detector was validated using SRS cone sizes from 6mm to 34mm and MLC defined field sizes from 5×5mm2 to 100×100mm2. The detector output response is linear with dose and is dose rate independent. Each field can be reconstructed accurately with a spatial resolution of 0.8mm and the current beam output is displayed every 50msec. Dosimetric errors of 1% with respect to the treatment plan can be identified and clinically significant deviations from the expected treatment can be displayed in real-time to alert the therapists. Conclusion: The high resolution detector is capable of reconstructing beam profiles in real-time with submillimeter resolution and 1% dose resolution. This system has the ability to project in-vivo both spatial and dosimetric errors during SBRT and SRS treatments when only a non-clinically significant fraction of the intended dose was delivered. The device has the

  4. Conceptual design for spacelab pool boiling experiment

    NASA Technical Reports Server (NTRS)

    Lienhard, J. H.; Peck, R. E.

    1978-01-01

    A pool boiling heat transfer experiment to be incorporated with a larger two-phase flow experiment on Spacelab was designed to confirm (or alter) the results of earth-normal gravity experiments which indicate that the hydrodynamic peak and minimum pool boiling heat fluxes vanish at very low gravity. Twelve small sealed test cells containing water, methanol or Freon 113 and cylindrical heaters of various sizes are to be built. Each cell will be subjected to one or more 45 sec tests in which the surface heat flux on the heaters is increased linearly until the surface temperature reaches a limiting value of 500 C. The entire boiling process will be photographed in slow-motion. Boiling curves will be constructed from thermocouple and electric input data, for comparison with the motion picture records. The conduct of the experiment will require no more than a few hours of operator time.

  5. Infrared thermometry study of nanofluid pool boiling phenomena

    PubMed Central

    2011-01-01

    Infrared thermometry was used to obtain first-of-a-kind, time- and space-resolved data for pool boiling phenomena in water-based nanofluids with diamond and silica nanoparticles at low concentration (<0.1 vol.%). In addition to macroscopic parameters like the average heat transfer coefficient and critical heat flux [CHF] value, more fundamental parameters such as the bubble departure diameter and frequency, growth and wait times, and nucleation site density [NSD] were directly measured for a thin, resistively heated, indium-tin-oxide surface deposited onto a sapphire substrate. Consistent with other nanofluid studies, the nanoparticles caused deterioration in the nucleate boiling heat transfer (by as much as 50%) and an increase in the CHF (by as much as 100%). The bubble departure frequency and NSD were found to be lower in nanofluids compared with water for the same wall superheat. Furthermore, it was found that a porous layer of nanoparticles built up on the heater surface during nucleate boiling, which improved surface wettability compared with the water-boiled surfaces. Using the prevalent nucleate boiling models, it was possible to correlate this improved surface wettability to the experimentally observed reductions in the bubble departure frequency, NSD, and ultimately to the deterioration in the nucleate boiling heat transfer and the CHF enhancement. PMID:21711754

  6. PRE-QAPP AGREEMENT (PQA) AND ANALYTICAL METHOD CHECKLISTS (AMCS): TOOLS FOR PLANNING RESEARCH PROJECTS

    EPA Science Inventory

    The Land Remediation and Pollution Control Division (LRPCD) QA Manager strives to assist LRPCD researchers in developing functional planning documents for their research projects. As part of the planning process, several pieces of information are needed, including information re...

  7. Theoretical modeling of CHF for near-saturated pool boiling and flow boiling from short heaters using the interfacial lift-off criterion

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mudawar, I.; Galloway, J.E.; Gersey, C.O.

    Pool boiling and flow boiling were examined for near-saturated bulk conditions in order to determine the critical heat flux (CHF) trigger mechanism for each. Photographic studies of the wall region revealed features common to both situations. At fluxes below CHF, the vapor coalesces into a wavy layer which permits wetting only in wetting fronts, the portions of the liquid-vapor interface which contact the wall as a result of the interfacial waviness. Close examination of the interfacial features revealed the waves are generated from the lower edge of the heater in pool boiling and the heater`s upstream region in flow boiling.more » Wavelengths follow predictions based upon the Kelvin-Helmholtz instability criterion. Critical heat flux in both cases occurs when the pressure force exerted upon the interface due to interfacial curvature, which tends to preserve interfacial contact with the wall prior to CHF, is overcome by the momentum of vapor at the site of the first wetting front, causing the interface to lift away from the wall. It is shown this interfacial lift-off criterion facilitates accurate theoretical modeling of CHF in pool boiling and in flow boiling in both straight and curved channels.« less

  8. A Fundamental Study of Nucleate Pool Boiling Under Microgravity

    NASA Technical Reports Server (NTRS)

    Ervin, Jamie S.; Merte, Herman, Jr.

    1996-01-01

    An experimental study of incipient boiling in short-term microgravity and with a/g = +/- 1 for pool boiling was performed. Calibrated thin gold films sputtered on a smoothly polished quartz surface were used simultaneously for thermal-resistance measurements and heating of the boiling surface. The gold films were used for both transient and quasi-steady heating surface temperature measurements. Two test vessels were constructed for precise measurement and control of fluid temperature and pressure: a laboratory pool boiling vessel for the a/g = +/- 1 experiments and a pool boiling vessel designed for the 131 m free-fall in the NASA Lewis Research Center Microgravity Research Facility for the microgravity tests. Measurements included the heater surface temperature, the pressure near the heating surface, the bulk liquid temperatures. High speed photography (up to 1,000 frames per second) was used in the experiments. With high quality microgravity and the measured initial temperature of the quiescent test fluid, R113, the temperature distribution in the liquid at the moment of boiling inception resulting from an imposed step in heat flux is known with a certainty not possible previously. The types of boiling propagation across the large flat heating surface, some observed here for the first time, are categorized; the conditions necessary for their occurrence are described. Explosive boiling propagation with a striking pattern of small scale protuberances over the entire vapor mass periphery not observed previously at low heat flux levels (on the order of 5 W/cm(exp 2)) is described. For the heater surface with a/g = -1, a step in the heater surface temperature of short duration was imposed. The resulting liquid temperature distribution at the moment of boiling inception was different from that obtained with a step in heat flux.

  9. SU-F-T-386: Analysis of Three QA Methods for Predicting Dose Deviation Pass Percentage for Lung SBRT VMAT Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hardin, M; To, D; Giaddui, T

    2016-06-15

    Purpose: To investigate the significance of using pinpoint ionization chambers (IC) and RadCalc (RC) in determining the quality of lung SBRT VMAT plans with low dose deviation pass percentage (DDPP) as reported by ScandiDos Delta4 (D4). To quantify the relationship between DDPP and point dose deviations determined by IC (ICDD), RadCalc (RCDD), and median dose deviation reported by D4 (D4DD). Methods: Point dose deviations and D4 DDPP were compiled for 45 SBRT VMAT plans. Eighteen patients were treated on Varian Truebeam linear accelerators (linacs); the remaining 27 were treated on Elekta Synergy linacs with Agility collimators. A one-way analysis ofmore » variance (ANOVA) was performed to determine if there were any statistically significant differences between D4DD, ICDD, and RCDD. Tukey’s test was used to determine which pair of means was statistically different from each other. Multiple regression analysis was performed to determine if D4DD, ICDD, or RCDD are statistically significant predictors of DDPP. Results: Median DDPP, D4DD, ICDD, and RCDD were 80.5% (47.6%–99.2%), −0.3% (−2.0%–1.6%), 0.2% (−7.5%–6.3%), and 2.9% (−4.0%–19.7%), respectively. The ANOVA showed a statistically significant difference between D4DD, ICDD, and RCDD for a 95% confidence interval (p < 0.001). Tukey’s test revealed a statistically significant difference between two pairs of groups, RCDD-D4DD and RCDD-ICDD (p < 0.001), but no difference between ICDD-D4DD (p = 0.485). Multiple regression analysis revealed that ICDD (p = 0.04) and D4DD (p = 0.03) are statistically significant predictors of DDPP with an adjusted r{sup 2} of 0.115. Conclusion: This study shows ICDD predicts trends in D4 DDPP; however this trend is highly variable as shown by our low r{sup 2}. This work suggests that ICDD can be used as a method to verify DDPP in delivery of lung SBRT VMAT plans. RCDD may not validate low DDPP discovered in D4 QA for small field SBRT treatments.« less

  10. Heat Transfer Performances of Pool Boiling on Metal-Graphite Composite Surfaces

    NASA Technical Reports Server (NTRS)

    Zhang, Nengli; Chao, David F.; Yang, Wen-Jei

    2000-01-01

    Nucleate boiling, especially near the critical heat flux (CHF), can provide excellent economy along with high efficiency of heat transfer. However, the performance of nucleate boiling may deteriorate in a reduced gravity environment and the nucleate boiling usually has a potentially dangerous characteristic in CHF regime. That is, any slight overload can result in burnout of the boiling surface because the heat transfer will suddenly move into the film-boiling regime. Therefore, enhancement of nucleate boiling heat transfer becomes more important in reduced gravity environments. Enhancing nucleate boiling and critical heat flux can be reached using micro-configured metal-graphite composites as the boiling surface. Thermocapillary force induced by temperature difference between the graphite-fiber tips and the metal matrix, which is independent of gravity, will play an important role in bubble detachment. Thus boiling heat transfer performance does not deteriorate in a reduced-gravity environment. Based on the existing experimental data, and a two-tier theoretical model, correlation formulas are derived for nucleate boiling on the copper-graphite and aluminum-graphite composite surfaces, in both the isolated and coalesced bubble regimes. Experimental studies were performed on nucleate pool boiling of pentane on cooper-graphite (Cu-Gr) and aluminum-graphite (Al-Gr) composite surfaces with various fiber volume concentrations for heat fluxes up to 35 W per square centimeter. It is revealed that a significant enhancement in boiling heat transfer performance on the composite surfaces is achieved, due to the presence of micro-graphite fibers embedded in the matrix. The onset of nucleate boiling (the isolated bubble regime) occurs at wall superheat of about 10 C for the Cu-Gr surface and 15 C for the Al-Gr surface, much lower than their respective pure metal surfaces. Transition from an isolated bubble regime to a coalesced bubble regime in boiling occurs at a superheat of

  11. Development of Flow Boiling and Condensation Experiment on the International Space Station- Normal and Low Gravity Flow Boiling Experiment Development and Test Results

    NASA Technical Reports Server (NTRS)

    Nahra, Henry K.; Hall, Nancy R.; Hasan, Mohammad M.; Wagner, James D.; May, Rochelle L.; Mackey, Jeffrey R.; Kolacz, John S.; Butcher, Robert L.; Frankenfield, Bruce J.; Mudawar, Issam; hide

    2013-01-01

    Flow boiling and condensation have been identified as two key mechanisms for heat transport that are vital for achieving weight and volume reduction as well as performance enhancement in future space systems. Since inertia driven flows are demanding on power usage, lower flows are desirable. However, in microgravity, lower flows are dominated by forces other than inertia (like the capillary force). It is of paramount interest to investigate limits of low flows beyond which the flow is inertial enough to be gravity independent. One of the objectives of the Flow Boiling and Condensation Flight Experiment sets to investigate these limits for flow boiling and condensation. A two-phase flow loop consisting of a Flow Boiling Module and two Condensation Modules has been developed to experimentally study flow boiling condensation heat transfer in the reduced gravity environment provided by the reduced gravity platform. This effort supports the development of a flow boiling and condensation facility for the International Space Station (ISS). The closed loop test facility is designed to deliver the test fluid, FC-72 to the inlet of any one of the test modules at specified thermodynamic and flow conditions. The zero-g-aircraft tests will provide subcooled and saturated flow boiling critical heat flux and flow condensation heat transfer data over wide range of flow velocities. Additionally, these tests will verify the performance of all gravity sensitive components, such as evaporator, condenser and accumulator associated with the two-phase flow loop. We will present in this paper the breadboard development and testing results which consist of detailed performance evaluation of the heater and condenser combination in reduced and normal gravity. We will also present the design of the reduced gravity aircraft rack and the results of the ground flow boiling heat transfer testing performed with the Flow Boiling Module that is designed to investigate flow boiling heat transfer and

  12. Cryogenic Boil-Off Reduction System Testing

    NASA Technical Reports Server (NTRS)

    Plachta, David W.; Johnson, Wesley L.; Feller, Jeffrey R.

    2014-01-01

    Cryogenic propellants such as liquid hydrogen (LH2) and liquid oxygen (LO2) are a part of NASA's future space exploration due to the high specific impulse that can be achieved using engines suitable for moving 10's to 100's of metric tons of payload mass to destinations outside of low earth orbit. However, the low storage temperatures of LH2 and LO2 cause substantial boil-off losses for missions with durations greater than several days. The losses can be greatly reduced by incorporating high performance cryocooler technology to intercept heat load to the propellant tanks and by the integration of self-supporting multi-layer insulation. The active thermal control technology under development is the integration of the reverse turbo- Brayton cycle cryocooler to the propellant tank through a distributed cooling network of tubes coupled to a shield in the tank insulation and to the tank wall itself. Also, the self-supporting insulation technology was utilized under the shield to obtain needed tank applied LH2 performance. These elements were recently tested at NASA Glenn Research Center in a series of three tests, two that reduced LH2 boil-off and one to eliminate LO2 boil-off. This test series was conducted in a vacuum chamber that replicated the vacuum of space and the temperatures of low Earth orbit. The test results show that LH2 boil-off was reduced 60% by the cryocooler system operating at 90K and that robust LO2 zero boil-off storage, including full tank pressure control was achieved.

  13. 40 CFR 98.434 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... Contained in Pre-Charged Equipment or Closed-Cell Foams § 98.434 Monitoring and QA/QC requirements. (a) For... equipment or closed-cell foam in the correct quantities (metric tons) and units (kg per piece of equipment...

  14. 40 CFR 98.434 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... Contained in Pre-Charged Equipment or Closed-Cell Foams § 98.434 Monitoring and QA/QC requirements. (a) For... equipment or closed-cell foam in the correct quantities (metric tons) and units (kg per piece of equipment...

  15. Feasibility study of a brine boiling machine by solar energy

    NASA Astrophysics Data System (ADS)

    Phayom, W.

    2018-06-01

    This study presented the technical and operational feasibility of brine boiling machine by using solar energy instead of firewood or husk for salt production. The solar salt brine boiling machine consisted of a boiling chamber with an enhanced thermal efficiency through use of a solar brine heater. The stainless steel solar salt brine boiling chamber had dimensions of 60 cm x 70 cm x 20 cm. The steel brine heater had dimensions of 70 cm x 80 cm x 20 cm. The tilt angle of both the boiling chamber and brine heater was 20 degrees from horizontal. The brine temperature in the reservoir tank was 42°C with a flow rate of 6.64 L/h discharging into the solar boiling machine. It was found that the thermal efficiency and overall efficiency of the solar salt brine boiling machine were 0.63 and 0.38, respectively at a solar irradiance of 787.6 W/m2. The results shows that the potential of using solar energy for salt production system is feasible.

  16. 40 CFR 98.434 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... Contained in Pre-Charged Equipment or Closed-Cell Foams § 98.434 Monitoring and QA/QC requirements. (a) For... equipment or closed-cell foam in the correct quantities and units. [74 FR 56374, Oct. 30, 2009, as amended...

  17. Liquid metal boiling inception

    NASA Technical Reports Server (NTRS)

    Sabin, C. M.; Poppendiek, H. F.; Mouritzen, G.; Meckel, P. T.; Cloakey, J. E.

    1972-01-01

    An experimental study of the inception of boiling in potassium in forced convection is reported. The boiler consisted of a 0.19-inch inside diameter, niobium-1% zirconium boiler tube approximately six feet long. Heating was accomplished by direct electrical tube wall conduction. Experiments were performed with both all-liquid fill and two-phase fill startup sequences and with a range of flow rates, saturation temperatures, inert gas levels, and fill liquid temperatures. Superheat of the liquid above the equilibrium saturation temperature was observed in all the experiments. Incipient boiling liquid superheat ranged from a few degrees to several hundred. Comparisons of these data with other data and with several analytical treatments are presented.

  18. SU-E-P-05: Is Routine Treatment Planning System Quality Assurance Necessary?

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Alaei, P

    Purpose: To evaluate the variation of dose calculations using a treatment planning system (TPS) over a two year period and assessment of the need for TPS QA on regular intervals. Methods: Two phantoms containing solid water and lung- and bone-equivalent heterogeneities were constructed in two different institutions for the same brand treatment planning system. Multiple plans, consisting of photons and electron beams, including IMRT and VMAT ones, were created and calculated on the phantoms. The accuracy of dose computation in the phantoms was evaluated at the onset by dose measurements within the phantoms. The dose values at up to 24more » points of interest (POI) within the solid water, lung, and bone slabs, as well as mean doses to several regions of interest (ROI), were re-calculated over a two-year period which included two software upgrades. The variations in POI and ROI dose values were analyzed and evaluated. Results: The computed doses vary slightly month-over-month. There are noticeable variations at the times of software upgrade, if the upgrade involves remodeling and/or re-commissioning of the beams. The variations are larger in certain points within the phantom, usually in the buildup region or near interfaces, and are almost non-existent for electron beams. Conclusion: Routine TPS QA is recommended by AAPM and other professional societies, and is often required by accreditation organizations. The frequency and type of QA, though, is subject to debate. The results presented here demonstrate that the frequency of these tests could be at longer intervals than monthly. However, it is essential to perform TPS QA at the time of commissioning and after each software upgrade.« less

  19. 26 CFR 54.4980B-2 - Plans that must comply.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... exception for qualified long-term care services is set forth in paragraph (e) of this Q&A-1, and for medical... all of the coverage provided under the plan is for qualified long-term care services (as defined in... whether substantially all of the coverage provided under the plan is for qualified long-term care services...

  20. CONFOCAL MICROSCOPY SYSTEM PERFORMANCE: QA TESTS, QUANTITATION AND SPECTROSCOPY

    EPA Science Inventory

    Confocal Microscopy System Performance: QA tests, Quantitation and Spectroscopy.

    Robert M. Zucker 1 and Jeremy M. Lerner 2,
    1Reproductive Toxicology Division, National Health and Environmental Effects Research Laboratory, Office of Research Development, U.S. Environmen...

  1. An investigation of transition boiling mechanisms of subcooled water under forced convective conditions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kwang-Won, Lee; Sang-Yong, Lee

    1995-09-01

    A mechanistic model for forced convective transition boiling has been developed to investigate transition boiling mechanisms and to predict transition boiling heat flux realistically. This model is based on a postulated multi-stage boiling process occurring during the passage time of the elongated vapor blanket specified at a critical heat flux (CHF) condition. Between the departure from nucleate boiling (DNB) and the departure from film boiling (DFB) points, the boiling heat transfer is established through three boiling stages, namely, the macrolayer evaporation and dryout governed by nucleate boiling in a thin liquid film and the unstable film boiling characterized by themore » frequent touches of the interface and the heated wall. The total heat transfer rates after the DNB is weighted by the time fractions of each stage, which are defined as the ratio of each stage duration to the vapor blanket passage time. The model predictions are compared with some available experimental transition boiling data. The parametric effects of pressure, mass flux, inlet subcooling on the transition boiling heat transfer are also investigated. From these comparisons, it can be seen that this model can identify the crucial mechanisms of forced convective transition boiling, and that the transition boiling heat fluxes including the maximum heat flux and the minimum film boiling heat flux are well predicted at low qualities/high pressures near 10 bar. In future, this model will be improved in the unstable film boiling stage and generalized for high quality and low pressure situations.« less

  2. SU-E-T-254: Development of a HDR-BT QA Tool for Verification of Source Position with Oncentra Applicator Modeling

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kumazaki, Y; Miyaura, K; Hirai, R

    2015-06-15

    Purpose: To develop a High Dose Rate Brachytherapy (HDR-BT) quality assurance (QA) tool for verification of source position with Oncentra applicator modeling, and to report the results of radiation source positions with this tool. Methods: We developed a HDR-BT QA phantom and automated analysis software for verification of source position with Oncentra applicator modeling for the Fletcher applicator used in the MicroSelectron HDR system. This tool is intended for end-to-end tests that mimic the clinical 3D image-guided brachytherapy (3D-IGBT) workflow. The phantom is a 30x30x3 cm cuboid phantom with radiopaque markers, which are inserted into the phantom to evaluate applicatormore » tips and reference source positions; positions are laterally shifted 10 mm from the applicator axis. The markers are lead-based and scatter radiation to expose the films. Gafchromic RTQA2 films are placed on the applicators. The phantom includes spaces to embed the applicators. The source position is determined as the distance between the exposed source position and center position of two pairs of the first radiopaque markers. We generated a 3D-IGBT plan with applicator modeling. The first source position was 6 mm from the applicator tips, and the second source position was 10 mm from the first source position. Results: All source positions were consistent with the exposed positions within 1 mm for all Fletcher applicators using in-house software. Moreover, the distance between source positions was in good agreement with the reference distance. Applicator offset, determined as the distance from the applicator tips at the first source position in the treatment planning system, was accurate. Conclusion: Source position accuracy of applicator modeling used in 3D-IGBT was acceptable. This phantom and software will be useful as a HDR-BT QA tool for verification of source position with Oncentra applicator modeling.« less

  3. SU-F-T-270: A Technique for Modeling a Diode Array Into the TPS for Lung SBRT Patient Specific QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Curley, C; Leventouri, T; Ouhib, Z

    2016-06-15

    Purpose: To accurately match the treatment planning system (TPS) with the measurement environment, where quality assurance (QA) devices are used to collect data, for lung Stereotactic Body Radiation Therapy (SBRT) patient specific QA. Incorporation of heterogeneities is also studied. Methods: Dual energy computerized tomography (DECT) and single energy computerized tomography (SECT) were used to model phantoms incorporating a 2-D diode array into the TPS. A water-equivalent and a heterogeneous phantom (simulating the thoracic region of a patient) were studied. Monte Carlo and pencil beam planar dose distributions were compared to measured distributions. Composite and individual fields were analyzed for normallymore » incident and planned gantry angle deliveries. γ- analysis was used with criteria 3% 3mm, 2% 2mm, and 1% 1mm. Results: The Monte Carlo calculations for the DECT resulted in improved agreements with the diode array for 46.4% of the fields at 3% 3mm, 85.7% at 2% 2mm, and 92.9% at 1% 1mm.For the SECT, the Monte Carlo calculations gave no agreement for the same γ-analysis criteria. Pencil beam calculations resulted in lower agreements with the diode array in the TPS. The DECT showed improvements for 14.3% of the fields at 3% 3mm and 2% 2mm, and 28.6% at 1% 1mm.In SECT comparisons, 7.1% of the fields at 3% 3mm, 10.7% at 2% 2mm, and 17.9% at 1% 1mm showed improved agreements with the diode array. Conclusion: This study demonstrates that modeling the diode array in the TPS is viable using DECT with Monte Carlo for patient specific lung SBRT QA. As recommended by task groups (e.g. TG 65, TG 101, TG 244) of the American Association of Physicists in Medicine (AAPM), pencil beam algorithms should be avoided in the presence of heterogeneous materials, including a diode array.« less

  4. Transient nucleate pool boiling in microgravity: Some initial results

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.; Lee, H. S.; Ervin, J. S.

    1994-01-01

    Variable gravity provides an opportunity to test the understanding of phenomena which are considered to depend on buoyancy, such as nucleate pool boiling. The active fundamental research in nucleate boiling has sought to determine the mechanisms or physical processes responsible for its high effectiveness, manifested by the high heat flux levels possible with relatively low temperature differences. Earlier research on nucleate pool boiling at high gravity levels under steady conditions demonstrated quantitatively that the heat transfer is degraded as the buoyancy normal to the heater surfaced increases. Correspondingly, it was later shown, qualitatively for short periods of time only, that nucleate boiling heat transfer is enhanced as the buoyancy normal to the heater surface is reduced. It can be deduced that nucleate pool boiling can be sustained as a quasi-steady process provided that some means is available to remove the vapor generated from the immediate vicinity of the heater surface. One of the objectives of the research, the initial results of which are presented here, is to quantify the heat transfer associated with boiling in microgravity. Some quantitative results of nucleate pool boiling in high quality microgravity (a/g approximately 10(exp -5)) of 5s duration, obtained in an evacuated drop tower, are presented here. These experiments were conducted as precursors of longer term space experiments. A transient heating technique is used, in which the heater surface is a transparent gold film sputtered on a qua rtz substrate, simultaneously providing the mean surface temperature from resistance thermometry and viewing of the boiling process both from beneath and across the surface. The measurement of the transient mean heater surface temperature permits the computation, by numerical means, of the transient mean heat transfer coefficient. The preliminary data obtained demonstrates that a quasi-steady boiling process can occur in microgravity if the bulk

  5. Studies on a Q/A selector for the SECRAL electron cyclotron resonance ion source.

    PubMed

    Yang, Y; Sun, L T; Feng, Y C; Fang, X; Lu, W; Zhang, W H; Cao, Y; Zhang, X Z; Zhao, H W

    2014-08-01

    Electron cyclotron resonance ion sources are widely used in heavy ion accelerators in the world because they are capable of producing high current beams of highly charged ions. However, the design of the Q/A selector system for these devices is challenging, because it must have a sufficient ion resolution while controlling the beam emittance growth. Moreover, this system has to be matched for a wide range of ion beam species with different intensities. In this paper, research on the Q/A selector system at the SECRAL (Superconducting Electron Cyclotron Resonance ion source with Advanced design in Lanzhou) platform both in experiment and simulation is presented. Based on this study, a new Q/A selector system has been designed for SECRAL II. The features of the new design including beam simulations are also presented.

  6. Electrically Driven Liquid Film Boiling Experiment

    NASA Technical Reports Server (NTRS)

    Didion, Jeffrey R.

    2016-01-01

    This presentation presents the science background and ground based results that form the basis of the Electrically Driven Liquid Film Boiling Experiment. This is an ISS experiment that is manifested for 2021. Objective: Characterize the effects of gravity on the interaction of electric and flow fields in the presence of phase change specifically pertaining to: a) The effects of microgravity on the electrically generated two-phase flow. b) The effects of microgravity on electrically driven liquid film boiling (includes extreme heat fluxes). Electro-wetting of the boiling section will repel the bubbles away from the heated surface in microgravity environment. Relevance/Impact: Provides phenomenological foundation for the development of electric field based two-phase thermal management systems leveraging EHD, permitting optimization of heat transfer surface area to volume ratios as well as achievement of high heat transfer coefficients thus resulting in system mass and volume savings. EHD replaces buoyancy or flow driven bubble removal from heated surface. Development Approach: Conduct preliminary experiments in low gravity and ground-based facilities to refine technique and obtain preliminary data for model development. ISS environment required to characterize electro-wetting effect on nucleate boiling and CHF in the absence of gravity. Will operate in the FIR - designed for autonomous operation.

  7. SUPERHEATING IN A BOILING WATER REACTOR

    DOEpatents

    Treshow, M.

    1960-05-31

    A boiling-water reactor is described in which the steam developed in the reactor is superheated in the reactor. This is accomplished by providing means for separating the steam from the water and passing the steam over a surface of the fissionable material which is not in contact with the water. Specifically water is boiled on the outside of tubular fuel elements and the steam is superheated on the inside of the fuel elements.

  8. 40 CFR 98.434 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98.434 Section 98.434 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Importers and Exporters of Fluorinated Greenhouse Gases...

  9. Answerers' Motivations and Strategies for Providing Information and Social Support in Social Q&A an Investigation of Health Question Answering

    ERIC Educational Resources Information Center

    Oh, Sanghee

    2010-01-01

    Social Q&A allows people to ask and answer questions for each other and to solve problems in everyday life collaboratively. The purpose of the current study is to understand the motivations and strategies of answerers in social Q&A. Thus, three research questions were investigated: (1) Why do answerers participate and contribute in social Q&A? (2)…

  10. TH-A-BRC-03: AAPM TG218: Measurement Methods and Tolerance Levels for Patient-Specific IMRT Verification QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Miften, M.

    2016-06-15

    AAPM TG-135U1 QA for Robotic Radiosurgery - Sonja Dieterich Since the publication of AAPM TG-135 in 2011, the technology of robotic radiosurgery has rapidly developed. AAPM TG-135U1 will provide recommendations on the clinical practice for using the IRIS collimator, fiducial-less real-time motion tracking, and Monte Carlo based treatment planning. In addition, it will summarize currently available literature about uncertainties. Learning Objectives: Understand the progression of technology since the first TG publication Learn which new QA procedures should be implemented for new technologies Be familiar with updates to clinical practice guidelines AAPM TG-178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance -more » Steven Goetsch Purpose: AAPM Task Group 178 Gamma Stereotactic Radiosurgery Dosimetry and Quality Assurance was formed in August, 2008. The Task Group has 12 medical physicists, two physicians and two consultants. Methods: A round robin dosimetry intercomparison of proposed ionization chambers, electrometer and dosimetry phantoms was conducted over a 15 month period in 2011 and 2012 (Med Phys 42, 11, Nov, 2015). The data obtained at 9 institutions (with ten different Elekta Gamma Knife units) was analyzed by the lead author using several protocols. Results: The most consistent results were obtained using the Elekta ABS 16cm diameter phantom, with the TG-51 protocol modified as recommended by Alfonso et al (Med Phys 35, 11, Nov 2008). A key white paper (Med Phys, in press) sponsored by Elekta Corporation, was used to obtain correction factors for the ionization chambers and phantoms used in this intercomparison. Consistent results were obtained for both Elekta Gamma Knife Model 4C and Gamma Knife Perfexion units as measured with each of two miniature ionization chambers. Conclusion: The full report gives clinical history and background of gamma stereotactic radiosurgery, clinical examples and history, quality assurance recommendations and

  11. Criticality in the slowed-down boiling crisis at zero gravity.

    PubMed

    Charignon, T; Lloveras, P; Chatain, D; Truskinovsky, L; Vives, E; Beysens, D; Nikolayev, V S

    2015-05-01

    Boiling crisis is a transition between nucleate and film boiling. It occurs at a threshold value of the heat flux from the heater called CHF (critical heat flux). Usually, boiling crisis studies are hindered by the high CHF and short transition duration (below 1 ms). Here we report on experiments in hydrogen near its liquid-vapor critical point, in which the CHF is low and the dynamics slow enough to be resolved. As under such conditions the surface tension is very small, the experiments are carried out in the reduced gravity to preserve the conventional bubble geometry. Weightlessness is created artificially in two-phase hydrogen by compensating gravity with magnetic forces. We were able to reveal the fractal structure of the contour of the percolating cluster of the dry areas at the heater that precedes the boiling crisis. We provide a direct statistical analysis of dry spot areas that confirms the boiling crisis at zero gravity as a scale-free phenomenon. It was observed that, in agreement with theoretical predictions, saturated boiling CHF tends to zero (within the precision of our thermal control system) in zero gravity, which suggests that the boiling crisis may be observed at any heat flux provided the experiment lasts long enough.

  12. 40 CFR 98.364 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... by Gas Chromatography (incorporated by reference see § 98.7). All gas composition monitors shall be...-90 (Reapproved 2006) Standard Practice for Analysis of Reformed Gas by Gas Chromatography... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Manure Management § 98.364 Monitoring and QA/QC requirements...

  13. 40 CFR 98.364 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... by Gas Chromatography (incorporated by reference see § 98.7). All gas composition monitors shall be...-90 (Reapproved 2006) Standard Practice for Analysis of Reformed Gas by Gas Chromatography... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Manure Management § 98.364 Monitoring and QA/QC requirements...

  14. 40 CFR 98.164 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Hydrogen Production § 98.164 Monitoring and QA/QC requirements. The GHG emissions data for hydrogen production process units must be quality-assured as specified in... Instrumental Determination of Carbon, Hydrogen, and Nitrogen in Petroleum Products and Lubricants (incorporated...

  15. The initiation of boiling during pressure transients. [water boiling on metal surfaces

    NASA Technical Reports Server (NTRS)

    Weisman, J.; Bussell, G.; Jashnani, I. L.; Hsieh, T.

    1973-01-01

    The initiation of boiling of water on metal surfaces during pressure transients has been investigated. The data were obtained by a new technique in which light beam fluctuations and a pressure signal were simultaneously recorded on a dual beam oscilloscope. The results obtained agreed with those obtained using high speed photography. It was found that, for water temperatures between 90-150 C, the wall superheat required to initiate boiling during a rapid pressure transient was significantly higher than required when the pressure was slowly reduced. This result is explained by assuming that a finite time is necessary for vapor to fill the cavity at which the bubble originates. Experimental measurements of this time are in reasonably good agreement with calculations based on the proposed theory. The theory includes a new procedure for estimating the coefficient of vaporization.

  16. MO-FG-CAMPUS-TeP1-05: Rapid and Efficient 3D Dosimetry for End-To-End Patient-Specific QA of Rotational SBRT Deliveries Using a High-Resolution EPID

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yang, Y M; Han, B; Xing, L

    2016-06-15

    Purpose: EPID-based patient-specific quality assurance provides verification of the planning setup and delivery process that phantomless QA and log-file based virtual dosimetry methods cannot achieve. We present a method for EPID-based QA utilizing spatially-variant EPID response kernels that allows for direct calculation of the entrance fluence and 3D phantom dose. Methods: An EPID dosimetry system was utilized for 3D dose reconstruction in a cylindrical phantom for the purposes of end-to-end QA. Monte Carlo (MC) methods were used to generate pixel-specific point-spread functions (PSFs) characterizing the spatially non-uniform EPID portal response in the presence of phantom scatter. The spatially-variant PSFs weremore » decomposed into spatially-invariant basis PSFs with the symmetric central-axis kernel as the primary basis kernel and off-axis representing orthogonal perturbations in pixel-space. This compact and accurate characterization enables the use of a modified Richardson-Lucy deconvolution algorithm to directly reconstruct entrance fluence from EPID images without iterative scatter subtraction. High-resolution phantom dose kernels were cogenerated in MC with the PSFs enabling direct recalculation of the resulting phantom dose by rapid forward convolution once the entrance fluence was calculated. A Delta4 QA phantom was used to validate the dose reconstructed in this approach. Results: The spatially-invariant representation of the EPID response accurately reproduced the entrance fluence with >99.5% fidelity with a simultaneous reduction of >60% in computational overhead. 3D dose for 10{sub 6} voxels was reconstructed for the entire phantom geometry. A 3D global gamma analysis demonstrated a >95% pass rate at 3%/3mm. Conclusion: Our approach demonstrates the capabilities of an EPID-based end-to-end QA methodology that is more efficient than traditional EPID dosimetry methods. Displacing the point of measurement external to the QA phantom reduces the

  17. 40 CFR 180.1056 - Boiled linseed oil; exemption from requirement of tolerance.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... “boiled linseed oil.” This exemption is limited to use on rice before edible parts form. [46 FR 33270... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Boiled linseed oil; exemption from... From Tolerances § 180.1056 Boiled linseed oil; exemption from requirement of tolerance. Boiled linseed...

  18. 40 CFR 98.164 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Hydrogen Production § 98.164 Monitoring and QA/QC requirements. The GHG emissions data for hydrogen production process units must be quality-assured as specified in..., Hydrogen, and Nitrogen in Petroleum Products and Lubricants (incorporated by reference, see § 98.7). (xi...

  19. SU-F-T-459: ArcCHECK Machine QA : Highly Efficient Quality Assurance Tool for VMAT, SRS & SBRT Linear Accelerator Delivery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mhatre, V; Patwe, P; Dandekar, P

    Purpose: Quality assurance (QA) of complex linear accelerators is critical and highly time consuming. ArcCHECK Machine QA tool is used to test geometric and delivery aspects of linear accelerator. In this study we evaluated the performance of this tool. Methods: Machine QA feature allows user to perform quality assurance tests using ArcCHECK phantom. Following tests were performed 1) Gantry Speed 2) Gantry Rotation 3) Gantry Angle 4)MLC/Collimator QA 5)Beam Profile Flatness & Symmetry. Data was collected on trueBEAM stX machine for 6 MV for a period of one year. The Gantry QA test allows to view errors in gantry angle,more » rotation & assess how accurately the gantry moves around the isocentre. The MLC/Collimator QA tool is used to analyze & locate the differences between leaf bank & jaw position of linac. The flatness & Symmetry test quantifies beam flatness & symmetry in IEC-y & x direction. The Gantry & Flatness/Symmetry test can be performed for static & dynamic delivery. Results: The Gantry speed was 3.9 deg/sec with speed maximum deviation around 0.3 deg/sec. The Gantry Isocentre for arc delivery was 0.9mm & static delivery was 0.4mm. The maximum percent positive & negative difference was found to be 1.9 % & – 0.25 % & maximum distance positive & negative diff was 0.4mm & – 0.3 mm for MLC/Collimator QA. The Flatness for Arc delivery was 1.8 % & Symmetry for Y was 0.8 % & X was 1.8 %. The Flatness for gantry 0°,270°,90° & 180° was 1.75,1.9,1.8 & 1.6% respectively & Symmetry for X & Y was 0.8,0.6% for 0°, 0.6,0.7% for 270°, 0.6,1% for 90° & 0.6,0.7% for 180°. Conclusion: ArcCHECK Machine QA is an useful tool for QA of Modern linear accelerators as it tests both geometric & delivery aspects. This is very important for VMAT, SRS & SBRT treatments.« less

  20. NUCLEAR SUPERHEATER FOR BOILING WATER REACTOR

    DOEpatents

    Holl, R.J.; Klecker, R.W.; Graham, C.B.

    1962-05-15

    A description is given of a boiling water reactor having a superheating region integral with the core. The core consists essentially of an annular boiling region surrounding an inner superheating region. Both regions contain fuel elements and are separated by a cylindrical wall, perforations being provided in the lower portion of the cylindrical wall to permit circulation of a common water moderator between the two regions. The superheater region comprises a plurality of tubular fuel assemblies through which the steam emanating from the boiling region passes to the steam outlet. Each superheater fuel assembly has an outer double-walled cylinder, the double walls being concentrically spaced and connected together at their upper ends but open at the bottom to provide for differential thermal expansion of the inner and outer walls. Gas is entrapped in the annulus between the walls which acts as an insulating space between the fissionable material inside and the moderator outside. (AEC)

  1. Physical quality of Simental Ongole crossbred silverside meat at various boiling times

    NASA Astrophysics Data System (ADS)

    Riyanto, J.; Cahyadi, M.; Guntari, W. S.

    2018-03-01

    This study aims to determine the physical quality of silverside beef meat at various boiling times. Samples that have been used are the back thigh or silverside meat. Treatment of boiling meat included TR (meat without boiled), R15 (boiled 15 minutes), and R30 (boiled for 30 minutes). The experimental design using Completely Randomized Design with 3 replications. Each replication was done in triple physical quality test. Determination of physical quality was performed at the Livestock Industry and Processing Laboratory at Sebelas Maret University Surakarta and the Meat Technology Laboratory at the Faculty of Animal Husbandry of Gadjah Mada University. The result of variance analysis showed that boiling affect cooking loss (P≥0.05) and but did not affect (P≤0,05) pH, water holding capacity and meat tenderness. The conclusions of the study showed that boiling for 15 minutes and 30 minutes decreased the cooking loss of Simental Ongole Crossbred silverside meat. Meat physical quality of pH, water holding capacity and the value of tenderness is not affected by boiling for 15 and 30 minutes.

  2. SU-F-BRE-13: Replacing Pre-Treatment Phantom QA with 3D In-Vivo Portal Dosimetry for IMRT Breast Cancer

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stroom, J; Vieira, S; Greco, C

    Purpose: Pre-treatment QA of individual treatment plans requires costly linac time and physics effort. Starting with IMRT breast treatments, we aim to replace pre-treatment QA with in-vivo portal dosimetry. Methods: Our IMRT breast cancer plans are routinely measured using the ArcCheck device (SunNuclear). 2D-Gamma analysis is performed with 3%/3mm criteria and the percentage of points with gamma<1 (nG1) is calculated within the 50% isodose surface. Following AAPM recommendations, plans with nG1<90% are approved; others need further inspection and might be rejected. For this study, we used invivo portal dosimetry (IPD) to measure the 3D back-projected dose of the first threemore » fractions for IMRT breast plans. Patient setup was online corrected before for all measured fractions. To reduce patient related uncertainties, the three IPD results were averaged and 3D-gamma analysis was applied with abovementioned criteria . For a subset of patients, phantom portal dosimetry (PPD) was also performed on a slab phantom. Results: Forty consecutive breast patients with plans that fitted the EPID were analysed. The average difference between planned and IPD dose in the reference point was −0.7+/−1.6% (1SD). Variation in nG1 between the 3 invivo fractions was about 6% (1SD). The average nG1 for IPD was 89+/−6%, worse than ArcCheck (95+/−3%). This can be explained by patient related factors such as changes in anatomy and/or model deficiencies due to e.g. inhomogeneities. For the 20 cases with PPD, mean nG1 was equal to ArcCheck values, which indicates that the two systems are equally accurate. These data therefore suggest that proper criteria for 3D invivo verification of breast treatments should be nG1>80% instead of nG1>90%, which, for our breast cases, would result in 5% (2/40) further inspections. Conclusion: First-fraction in-vivo portal dosimetry using new gamma-evaluation criteria will replace phantom measurements in our institution, saving resources and

  3. Nucleate Boiling Heat Transfer Studied Under Reduced-Gravity Conditions

    NASA Technical Reports Server (NTRS)

    Chao, David F.; Hasan, Mohammad M.

    2000-01-01

    Boiling is known to be a very efficient mode of heat transfer, and as such, it is employed in component cooling and in various energy-conversion systems. In space, boiling heat transfer may be used in thermal management, fluid handling and control, power systems, and on-orbit storage and supply systems for cryogenic propellants and life-support fluids. Recent interest in the exploration of Mars and other planets and in the concept of in situ resource utilization on the Martian and Lunar surfaces highlights the need to understand how gravity levels varying from the Earth's gravity to microgravity (1g = or > g/g(sub e) = or > 10(exp -6)g) affect boiling heat transfer. Because of the complex nature of the boiling process, no generalized prediction or procedure has been developed to describe the boiling heat transfer coefficient, particularly at reduced gravity levels. Recently, Professor Vijay K. Dhir of the University of California at Los Angeles proposed a novel building-block approach to investigate the boiling phenomena in low-gravity to microgravity environments. This approach experimentally investigates the complete process of bubble inception, growth, and departure for single bubbles formed at a well-defined and controllable nucleation site. Principal investigator Professor Vijay K. Dhir, with support from researchers from the NASA Glenn Research Center at Lewis Field, is performing a series of pool boiling experiments in the low-gravity environments of the KC 135 microgravity aircraft s parabolic flight to investigate the inception, growth, departure, and merger of bubbles from single- and multiple-nucleation sites as a function of the wall superheat and the liquid subcooling. Silicon wafers with single and multiple cavities of known characteristics are being used as test surfaces. Water and PF5060 (an inert liquid) were chosen as test liquids so that the role of surface wettability and the magnitude of the effect of interfacial tension on boiling in reduced

  4. CONTINUOUS ANALYZER UTILIZING BOILING POINT DETERMINATION

    DOEpatents

    Pappas, W.S.

    1963-03-19

    A device is designed for continuously determining the boiling point of a mixture of liquids. The device comprises a distillation chamber for boiling a liquid; outlet conduit means for maintaining the liquid contents of said chamber at a constant level; a reflux condenser mounted above said distillation chamber; means for continuously introducing an incoming liquid sample into said reflux condenser and into intimate contact with vapors refluxing within said condenser; and means for measuring the temperature of the liquid flowing through said distillation chamber. (AEC)

  5. 40 CFR 98.84 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Cement Production § 98.84 Monitoring and QA/QC requirements..., shale, iron oxide, and alumina). Facilities that opt to use the default total organic carbon factor... quantity of each category of raw materials consumed by the facility (e.g., limestone, sand, shale, iron...

  6. 40 CFR 98.84 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Cement Production § 98.84 Monitoring and QA/QC requirements..., shale, iron oxide, and alumina). Facilities that opt to use the default total organic carbon factor... quantity of each category of raw materials consumed by the facility (e.g., limestone, sand, shale, iron...

  7. 40 CFR 98.84 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Cement Production § 98.84 Monitoring and QA/QC requirements..., shale, iron oxide, and alumina). Facilities that opt to use the default total organic carbon factor... quantity of each category of raw materials consumed by the facility (e.g., limestone, sand, shale, iron...

  8. 40 CFR 98.144 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Glass Production § 98.144 Monitoring and QA/QC requirements. (a) You must measure annual amounts of carbonate-based raw materials charged to each continuous glass... calibrated scales or weigh hoppers. Total annual mass charged to glass melting furnaces at the facility shall...

  9. 40 CFR 98.74 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... Relative Molecular Mass of Petroleum Oils from Viscosity Measurements (incorporated by reference, see § 98... Weight) of Hydrocarbons by Thermoelectric Measurement of Vapor Pressure (incorporated by reference, see... measurements according to the monitoring and QA/QC requirements for the Tier 3 methodology in § 98.34(b). (e...

  10. 40 CFR 98.364 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... or operator shall document the procedures used to ensure the accuracy of gas flow rate, gas... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Manure Management § 98.364 Monitoring and QA/QC requirements... fraction of total manure managed in each system component. (c) The CH4 concentration of gas from digesters...

  11. 40 CFR 98.364 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... or operator shall document the procedures used to ensure the accuracy of gas flow rate, gas... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Manure Management § 98.364 Monitoring and QA/QC requirements... fraction of total manure managed in each system component. (c) The CH4 concentration of gas from digesters...

  12. 40 CFR 98.364 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... or operator shall document the procedures used to ensure the accuracy of gas flow rate, gas... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Manure Management § 98.364 Monitoring and QA/QC requirements... fraction of total manure managed in each system component. (c) The CH4 concentration of gas from digesters...

  13. 40 CFR 98.414 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Industrial Greenhouse Gases § 98.414 Monitoring... or better. If the mass in paragraph (a) of this section is measured by weighing containers that...

  14. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Petroleum Products § 98.394 Monitoring and QA/QC requirements. (a) Determination of quantity. (1) The quantity of petroleum products, natural gas liquids... product or natural gas liquid on any day of each calendar month of the reporting year in which the...

  15. Cooling of hot bubbles by surface texture during the boiling crisis

    NASA Astrophysics Data System (ADS)

    Dhillon, Navdeep; Buongiorno, Jacopo; Varanasi, Kripa

    2015-11-01

    We report the existence of maxima in critical heat flux (CHF) enhancement for pool boiling on textured hydrophilic surfaces and reveal the interaction mechanism between bubbles and surface texture that governs the boiling crisis phenomenon. Boiling is a process of fundamental importance in many engineering and industrial applications but the maximum heat flux that can be absorbed by the boiling liquid (or CHF) is limited by the boiling crisis. Enhancing the CHF of industrial boilers by surface texturing can lead to substantial energy savings and reduction in greenhouse gas emissions on a global scale. However, the fundamental mechanisms behind this enhancement are not well understood, with some previous studies indicating that CHF should increase monotonically with increasing texture density. However, using pool boiling experiments on a parametrically designed set of plain and nano-textured micropillar surfaces, we show that there is an optimum intermediate texture density that maximizes CHF and further that the length scale of this texture is of fundamental significance. Using imbibition experiments and high-speed optical and infrared imaging, we reveal the fundamental mechanisms governing the CHF enhancement maxima in boiling crisis. We acknowledge funding from the Chevron corporation.

  16. Microbiological effectiveness of disinfecting water by boiling in rural Guatemala.

    PubMed

    Rosa, Ghislaine; Miller, Laura; Clasen, Thomas

    2010-03-01

    Boiling is the most common means of treating water in the home and the benchmark against which alternative point-of-use water treatment options must be compared. In a 5-week study in rural Guatemala among 45 households who claimed they always or almost always boiled their drinking water, boiling was associated with a 86.2% reduction in geometric mean thermotolerant coliforms (TTC) (N = 206, P < 0.0001). Despite consistent levels of fecal contamination in source water, 71.2% of stored water samples from self-reported boilers met the World Health Organization guidelines for safe drinking water (0 TTC/100 mL), and 10.7% fell within the commonly accepted low-risk category of (1-10 TTC/100 mL). As actually practiced in the study community, boiling significantly improved the microbiological quality of drinking water, though boiled and stored drinking water is not always free of fecal contaminations.

  17. SU-F-P-37: Implementation of An End-To-End QA Test of the Radiation Therapy Imaging, Planning and Delivery Process to Identify and Correct Possible Sources of Deviation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Salinas Aranda, F; Suarez, V; Arbiser, S

    2016-06-15

    Purpose: To implement an end-to-end QA test of the radiation therapy imaging, planning and delivery process, aimed to assess the dosimetric agreement accuracy between planned and delivered treatment, in order to identify and correct possible sources of deviation. To establish an internal standard for machine commissioning acceptance. Methods: A test involving all steps of the radiation therapy: imaging, planning and delivery process was designed. The test includes analysis of point dose and planar dose distributions agreement between TPS calculated and measured dose. An ad hoc 16 cm diameter PMMA phantom was constructed with one central and four peripheral bores thatmore » can accommodate calibrated electron density inserts. Using Varian Eclipse 10.0 and Elekta XiO 4.50 planning systems, IMRT, RapidArc and 3DCRT with hard and dynamic wedges plans were planned on the phantom and tested. An Exradin A1SL chamber is used with a Keithley 35617EBS electrometer for point dose measurements in the phantom. 2D dose distributions were acquired using MapCheck and Varian aS1000 EPID.Gamma analysis was performed for evaluation of 2D dose distribution agreement using MapCheck software and Varian Portal Dosimetry Application.Varian high energy Clinacs Trilogy, 2100C/CD, 2000CR and low energy 6X/EX where tested.TPS-CT# vs. electron density table were checked for CT-scanners used. Results: Calculated point doses were accurate to 0.127% SD: 0.93%, 0.507% SD: 0.82%, 0.246% SD: 1.39% and 0.012% SD: 0.01% for LoX-3DCRT, HiX-3DCRT, IMRT and RapidArc plans respectively. Planar doses pass gamma 3% 3mm in all cases and 2% 2mm for VMAT plans. Conclusion: Implementation of a simple and reliable quality assurance tool was accomplished. The end-to-end proved efficient, showing excellent agreement between planned and delivered dose evidencing strong consistency of the whole process from imaging through planning to delivery. This test can be used as a first step in beam model acceptance for

  18. 40 CFR 98.444 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 22 2012-07-01 2012-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Geologic Sequestration of Carbon Dioxide § 98.444 Monitoring... volume of contents in all containers if you receive CO2 in containers by following the most appropriate...

  19. 40 CFR 98.444 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 21 2014-07-01 2014-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Geologic Sequestration of Carbon Dioxide § 98.444 Monitoring... volume of contents in all containers if you receive CO2 in containers by following the most appropriate...

  20. 40 CFR 98.444 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 40 Protection of Environment 22 2013-07-01 2013-07-01 false Monitoring and QA/QC requirements. 98... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Geologic Sequestration of Carbon Dioxide § 98.444 Monitoring... volume of contents in all containers if you receive CO2 in containers by following the most appropriate...

  1. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Petroleum Products § 98.394 Monitoring and QA/QC requirements. (a) Determination of quantity. (1) The quantity of petroleum products, natural gas liquids, and... each petroleum product or natural gas liquid on any day of each calendar month of the reporting year in...

  2. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Petroleum Products § 98.394 Monitoring and QA/QC requirements. (a) Determination of quantity. (1) The quantity of petroleum products, natural gas liquids, and... each petroleum product or natural gas liquid on any day of each calendar month of the reporting year in...

  3. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Petroleum Products § 98.394 Monitoring and QA/QC requirements. (a) Determination of quantity. (1) The quantity of petroleum products, natural gas liquids, and... or natural gas liquid on any day of each calendar month of the reporting year in which the quantity...

  4. 40 CFR 98.394 - Monitoring and QA/QC requirements.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... (CONTINUED) MANDATORY GREENHOUSE GAS REPORTING Suppliers of Petroleum Products § 98.394 Monitoring and QA/QC requirements. (a) Determination of quantity. (1) The quantity of petroleum products, natural gas liquids, and... each petroleum product or natural gas liquid on any day of each calendar month of the reporting year in...

  5. The boiling point of stratospheric aerosols.

    NASA Technical Reports Server (NTRS)

    Rosen, J. M.

    1971-01-01

    A photoelectric particle counter was used for the measurement of aerosol boiling points. The operational principle involves raising the temperature of the aerosol by vigorously heating a portion of the intake tube. At or above the boiling point, the particles disintegrate rather quickly, and a noticeable effect on the size distribution and concentration is observed. Stratospheric aerosols appear to have the same volatility as a solution of 75% sulfuric acid. Chemical analysis of the aerosols indicates that there are other substances present, but that the sulfate radical is apparently the major constituent.

  6. INTERIM EPA GUIDANCE FOR GEOSPATIAL-RELATED QUALITY ASSURANCE PROJECT PLANS

    EPA Science Inventory

    This guidance supplements EPA Guidance for Quality,Assurance Project Plans (EPA QA/G-5), in that the focus here is on collection and use of geospatial rather than other environmental data (e.g., strictly chemical or biological data), including unique aspects of data storage, retr...

  7. Microbiological Effectiveness of Disinfecting Water by Boiling in Rural Guatemala

    PubMed Central

    Rosa, Ghislaine; Miller, Laura; Clasen, Thomas

    2010-01-01

    Boiling is the most common means of treating water in the home and the benchmark against which alternative point-of-use water treatment options must be compared. In a 5-week study in rural Guatemala among 45 households who claimed they always or almost always boiled their drinking water, boiling was associated with a 86.2% reduction in geometric mean thermotolerant coliforms (TTC) (N = 206, P < 0.0001). Despite consistent levels of fecal contamination in source water, 71.2% of stored water samples from self-reported boilers met the World Health Organization guidelines for safe drinking water (0 TTC/100 mL), and 10.7% fell within the commonly accepted low-risk category of (1–10 TTC/100 mL). As actually practiced in the study community, boiling significantly improved the microbiological quality of drinking water, though boiled and stored drinking water is not always free of fecal contaminations. PMID:20207876

  8. BOILING REACTORS

    DOEpatents

    Untermyer, S.

    1962-04-10

    A boiling reactor having a reactivity which is reduced by an increase in the volume of vaporized coolant therein is described. In this system unvaporized liquid coolant is extracted from the reactor, heat is extracted therefrom, and it is returned to the reactor as sub-cooled liquid coolant. This reduces a portion of the coolant which includes vaporized coolant within the core assembly thereby enhancing the power output of the assembly and rendering the reactor substantially self-regulating. (AEC)

  9. Experimental evidence of the vapor recoil mechanism in the boiling crisis.

    PubMed

    Nikolayev, V S; Chatain, D; Garrabos, Y; Beysens, D

    2006-11-03

    Boiling crisis experiments are carried out in the vicinity of the liquid-gas critical point of H2. A magnetic gravity compensation setup is used to enable nucleate boiling at near critical pressure. The measurements of the critical heat flux that defines the threshold for the boiling crisis are carried out as a function of the distance from the critical point. The obtained power law behavior and the boiling crisis dynamics agree with the predictions of the vapor recoil mechanism and disagree with the classical vapor column mechanism.

  10. MO-PIS-Exhibit Hall-01: Tools for TG-142 Linac Imaging QA I

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Clements, M; Wiesmeyer, M

    2014-06-15

    Partners in Solutions is an exciting new program in which AAPM partners with our vendors to present practical “hands-on” information about the equipment and software systems that we use in our clinics. The therapy topic this year is solutions for TG-142 recommendations for linear accelerator imaging QA. Note that the sessions are being held in a special purpose room built on the Exhibit Hall Floor, to encourage further interaction with the vendors. Automated Imaging QA for TG-142 with RIT Presentation Time: 2:45 – 3:15 PM This presentation will discuss software tools for automated imaging QA and phantom analysis for TG-142.more » All modalities used in radiation oncology will be discussed, including CBCT, planar kV imaging, planar MV imaging, and imaging and treatment coordinate coincidence. Vendor supplied phantoms as well as a variety of third-party phantoms will be shown, along with appropriate analyses, proper phantom setup procedures and scanning settings, and a discussion of image quality metrics. Tools for process automation will be discussed which include: RIT Cognition (machine learning for phantom image identification), RIT Cerberus (automated file system monitoring and searching), and RunQueueC (batch processing of multiple images). In addition to phantom analysis, tools for statistical tracking, trending, and reporting will be discussed. This discussion will include an introduction to statistical process control, a valuable tool in analyzing data and determining appropriate tolerances. An Introduction to TG-142 Imaging QA Using Standard Imaging Products Presentation Time: 3:15 – 3:45 PM Medical Physicists want to understand the logic behind TG-142 Imaging QA. What is often missing is a firm understanding of the connections between the EPID and OBI phantom imaging, the software “algorithms” that calculate the QA metrics, the establishment of baselines, and the analysis and interpretation of the results. The goal of our brief presentation

  11. Q&A: The Basics of California's School Finance System

    ERIC Educational Resources Information Center

    EdSource, 2006

    2006-01-01

    In a state as large and complex as California, education financing can become as complicated as rocket science. This two-page Q&A provides a brief, easy-to-understand explanation of California's school finance system and introduces the issues of its adequacy and equity. A list of resources providing additional information is provided.

  12. Estimation of Eye Lens Dose During Brain Scans Using Gafchromic Xr-QA2 Film in Various Multidetector CT Scanners.

    PubMed

    Akhilesh, Philomina; Kulkarni, Arti R; Jamhale, Shramika H; Sharma, S D; Kumar, Rajesh; Datta, D

    2017-04-25

    The purpose of this study was to estimate eye lens dose during brain scans in 16-, 64-, 128- and 256-slice multidetector computed tomography (CT) scanners in helical acquisition mode and to test the feasibility of using radiochromic film as eye lens dosemeter during CT scanning. Eye lens dose measurements were performed using Gafchromic XR-QA2 film on a polystyrene head phantom designed with outer dimensions equivalent to the head size of a reference Indian man. The response accuracy of XR-QA2 film was validated by using thermoluminescence dosemeters. The eye lens dose measured using XR-QA2 film on head phantom for plain brain scanning in helical mode ranged from 43.8 to 45.8 mGy. The XR-QA2 film measured dose values were in agreement with TLD measured dose values within a maximum variation of 8.9%. The good correlation between the two data sets confirms the viability of using XR-QA2 film for eye lens dosimetry. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  13. A novel DTI-QA tool: Automated metric extraction exploiting the sphericity of an agar filled phantom.

    PubMed

    Chavez, Sofia; Viviano, Joseph; Zamyadi, Mojdeh; Kingsley, Peter B; Kochunov, Peter; Strother, Stephen; Voineskos, Aristotle

    2018-02-01

    To develop a quality assurance (QA) tool (acquisition guidelines and automated processing) for diffusion tensor imaging (DTI) data using a common agar-based phantom used for fMRI QA. The goal is to produce a comprehensive set of automated, sensitive and robust QA metrics. A readily available agar phantom was scanned with and without parallel imaging reconstruction. Other scanning parameters were matched to the human scans. A central slab made up of either a thick slice or an average of a few slices, was extracted and all processing was performed on that image. The proposed QA relies on the creation of two ROIs for processing: (i) a preset central circular region of interest (ccROI) and (ii) a signal mask for all images in the dataset. The ccROI enables computation of average signal for SNR calculations as well as average FA values. The production of the signal masks enables automated measurements of eddy current and B0 inhomogeneity induced distortions by exploiting the sphericity of the phantom. Also, the signal masks allow automated background localization to assess levels of Nyquist ghosting. The proposed DTI-QA was shown to produce eleven metrics which are robust yet sensitive to image quality changes within site and differences across sites. It can be performed in a reasonable amount of scan time (~15min) and the code for automated processing has been made publicly available. A novel DTI-QA tool has been proposed. It has been applied successfully on data from several scanners/platforms. The novelty lies in the exploitation of the sphericity of the phantom for distortion measurements. Other novel contributions are: the computation of an SNR value per gradient direction for the diffusion weighted images (DWIs) and an SNR value per non-DWI, an automated background detection for the Nyquist ghosting measurement and an error metric reflecting the contribution of EPI instability to the eddy current induced shape changes observed for DWIs. Copyright © 2017 Elsevier

  14. Heat Transfer in Boiling Dilute Emulsion with Strong Buoyancy

    NASA Astrophysics Data System (ADS)

    Freeburg, Eric Thomas

    Little attention has been given to the boiling of emulsions compared to that of boiling in pure liquids. The advantages of using emulsions as a heat transfer agent were first discovered in the 1970s and several interesting features have since been studied by few researchers. Early research focuses primarily on pool and flow boiling and looks to determine a mechanism by which the boiling process occurs. This thesis looks at the boiling of dilute emulsions in fluids with strong buoyant forces. The boiling of dilute emulsions presents many favorable characteristics that make it an ideal agent for heat transfer. High heat flux electronics, such as those seen in avionics equipment, produce high heat fluxes of 100 W/cm2 or more, but must be maintained at low temperatures. So far, research on single phase convection and flow boiling in small diameter channels have yet to provide an adequate solution. Emulsions allow the engineer to tailor the solution to the specific problem. The fluid can be customized to retain the high thermal conductivity and specific heat capacity of the continuous phase while enhancing the heat transfer coefficient through boiling of the dispersed phase component. Heat transfer experiments were carried out with FC-72 in water emulsions. FC-72 has a saturation temperature of 56 °C, far below that of water. The parameters were varied as follows: 0% ≤ epsilon ≤ 1% and 1.82 x 1012 ≤ RaH ≤ 4.42 x 1012. Surface temperatures along the heated surface reached temperature that were 20 °C in excess of the dispersed phase saturation temperature. An increase of ˜20% was seen in the average Nusselt numbers at the highest Rayleigh numbers. Holography was used to obtain images of individual and multiple FC-72 droplets in the boundary layer next to the heated surface. The droplet diameters ranged from 0.5 mm to 1.3 mm. The Magnus effect was observed when larger individual droplets were injected into the boundary layer, causing the droplets to be pushed

  15. SU-F-T-344: Commissioning Constant Dose Rate VMAT in the Raystation Treatment Planning System for a Varian Clinac IX

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pursley, J; Gueorguiev, G; Prichard, H

    Purpose: To demonstrate the commissioning of constant dose rate volumetric modulated arc therapy (VMAT) in the Raystation treatment planning system for a Varian Clinac iX with Exact couch. Methods: Constant dose rate (CDR) VMAT is an option in the Raystation treatment planning system, enabling VMAT delivery on Varian linacs without a RapidArc upgrade. Raystation 4.7 was used to commission CDR-VMAT for a Varian Clinac iX. Raystation arc model parameters were selected to match machine deliverability characteristics. A Varian Exact couch model was added to Raystation 4.7 and commissioned for use in VMAT optimization. CDR-VMAT commissioning checks were performed on themore » linac, including patient-specific QA measurements for 10 test patients using both the ArcCHECK from Sun Nuclear Corporation and COMPASS from IBA Dosimetry. Multi-criteria optimization (MCO) in Raystation was used for CDR-VMAT planning. Results: Raystation 4.7 generated clinically acceptable and deliverable CDR-VMAT plans for the Varian Clinac. VMAT plans were optimized including a model of the Exact couch with both rails in the out positions. CDR-VMAT plans generated with MCO in Raystation were dosimetrically comparable to Raystation MCO-generated IMRT plans. Patient-specific QA measurements with the ArcCHECK on the couch showed good agreement with the treatment planning system prediction. Patient-specific, structure-specific, multi-statistical parameter 3D QA measurements with gantry-mounted COMPASS also showed good agreement. Conclusion: Constant dose rate VMAT was successfully modeled in Raystation 4.7 for a Varian Clinac iX, and Raystation’s multicriteria optimization generated constant dose rate VMAT plans which were deliverable and dosimetrically comparable to IMRT plans.« less

  16. Numerical and Experimental Study of Mechanisms Involved in Boiling Histotripsy.

    PubMed

    Pahk, Ki Joo; Gélat, Pierre; Sinden, David; Dhar, Dipok Kumar; Saffari, Nader

    2017-12-01

    The aim of boiling histotripsy is to mechanically fractionate tissue as an alternative to thermal ablation for therapeutic applications. In general, the shape of a lesion produced by boiling histotripsy is tadpole like, consisting of a head and a tail. Although many studies have demonstrated the efficacy of boiling histotripsy for fractionating solid tumors, the exact mechanisms underpinning this phenomenon are not yet well understood, particularly the interaction of a boiling vapor bubble with incoming incident shockwaves. To investigate the mechanisms involved in boiling histotripsy, a high-speed camera with a passive cavitation detection system was used to observe the dynamics of bubbles produced in optically transparent tissue-mimicking gel phantoms exposed to the field of a 2.0-MHz high-intensity focused ultrasound (HIFU) transducer. We observed that boiling bubbles were generated in a localized heated region and cavitation clouds were subsequently induced ahead of the expanding bubble. This process was repeated with HIFU pulses and eventually resulted in a tadpole-shaped lesion. A simplified numerical model describing the scattering of the incident ultrasound wave by a vapor bubble was developed to help interpret the experimental observations. Together with the numerical results, these observations suggest that the overall size of a lesion induced by boiling histotripsy is dependent on the sizes of (i) the heated region at the HIFU focus and (ii) the backscattered acoustic field by the original vapor bubble. Copyright © 2017 World Federation for Ultrasound in Medicine and Biology. Published by Elsevier Inc. All rights reserved.

  17. Acoustic emission feedback control for control of boiling in a microwave oven

    DOEpatents

    White, Terry L.

    1991-01-01

    An acoustic emission based feedback system for controlling the boiling level of a liquid medium in a microwave oven is provided. The acoustic emissions from the medium correlated with surface boiling is used to generate a feedback control signal proportional to the level of boiling of the medium. This signal is applied to a power controller to automatically and continuoulsly vary the power applied to the oven to control the boiling at a selected level.

  18. Single-bubble boiling under Earth's and low gravity

    NASA Astrophysics Data System (ADS)

    Khusid, Boris; Elele, Ezinwa; Lei, Qian; Tang, John; Shen, Yueyang

    2017-11-01

    Miniaturization of electronic systems in terrestrial and space applications is challenged by a dramatic increase in the power dissipation per unit volume with the occurrence of localized hot spots where the heat flux is much higher than the average. Cooling by forced gas or liquid flow appears insufficient to remove high local heat fluxes. Boiling that involves evaporation of liquid in a hot spot and condensation of vapor in a cold region can remove a significantly larger amount of heat through the latent heat of vaporization than force-flow cooling can carry out. Traditional methods for enhancing boiling heat transfer in terrestrial and space applications focus on removal of bubbles from the heating surface. In contrast, we unexpectedly observed a new boiling regime of water under Earth's gravity and low gravity in which a bubble was pinned on a small heater up to 270°C and delivered a heat flux up to 1.2 MW/m2 that was as high as the critical heat flux in the classical boiling regime on Earth .Low gravity measurements conducted in parabolic flights in NASA Boeing 727. The heat flux in flight and Earth's experiments was found to rise linearly with increasing the heater temperature. We will discuss physical mechanisms underlying heat transfer in single-bubble boiling. The work supported by NASA Grants NNX12AM26G and NNX09AK06G.

  19. Boiling-induced formation of colloidal gold in black smoker hydrothermal fluids

    USGS Publications Warehouse

    Gartman, Amy; Hannington, Mark; Jamieson, John W.; Peterkin, Ben; Garbe-Schönberg, Dieter; Findlay, Alyssa J; Fuchs, Sebastian; Kwasnitschka, Tom

    2017-01-01

    Gold colloids occur in black smoker fluids from the Niua South hydrothermal vent field, Lau Basin (South Pacific Ocean), confirming the long-standing hypothesis that gold may undergo colloidal transport in hydrothermal fluids. Six black smoker vents, varying in temperature from 250 °C to 325 °C, were sampled; the 325 °C vent was boiling at the time of sampling and the 250 °C fluids were diffusely venting. Native gold particles ranging from <50 nm to 2 µm were identified in 4 of the fluid samples and were also observed to precipitate on the sampler during collection from the boiling vent. Total gold concentrations (dissolved and particulate) in the fluid samples range from 1.6 to 5.4 nM in the high-temperature, focused flow vents. Although the gold concentrations in the focused flow fluids are relatively high, they are lower than potential solubilities prior to boiling and indicate that precipitation was boiling induced, with sulfide lost upon boiling to exsolution and metal sulfide formation. Gold concentrations reach 26.7 nM in the 250 °C diffuse flow sample, and abundant native gold particles were also found in the fluids and associated sulfide chimney and are interpreted to be a product of colloid accumulation and growth following initial precipitation upon boiling. These results indicate that colloid-driven precipitation as a result of boiling, the persistence of colloids after boiling, and the accumulation of colloids in diffuse flow fluids are important mechanisms for the enrichment of gold in seafloor hydrothermal systems.

  20. Steady State Film Boiling Heat Transfer Simulated With Trace V4.160

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Audrius Jasiulevicius; Rafael Macian-Juan

    2006-07-01

    This paper presents the results of the assessment and analysis of TRACE v4.160 heat transfer predictions in the post-CHF (critical heat flux) region and discusses the possibilities to improve the TRACE v4.160 code predictions in the film boiling heat transfer when applying different film boiling correlations. For this purpose, the TRACE v4.160-calculated film boiling heat flux and the resulting maximum inner wall temperatures during film boiling in single tubes were compared with experimental data obtained at the Royal Institute of Technology (KTH) in Stockholm, Sweden. The experimental database included measurements for pressures ranging from 30 to 200 bar and coolantmore » mass fluxes from 500 to 3000 kg/m{sup 2}s. It was found that TRACE v4.160 does not produce correct predictions of the film boiling heat flux, and consequently of the maximum inner wall temperature in the test section, under the wide range of conditions documented in the KTH experiments. In particular, it was found that the standard TRACE v4.160 under-predicts the film boiling heat transfer coefficient at low pressure-low mass flux and high pressure-high mass flux conditions. For most of the rest of the investigated range of parameters, TRACE v4.160 over-predicts the film boiling heat transfer coefficient, which can lead to non-conservative predictions in applications to nuclear power plant analyses. Since no satisfactory agreement with the experimental database was obtained with the standard TRACE v4.160 film boiling heat transfer correlations, we have added seven film boiling correlations to TRACE v4.160 in order to investigate the possibility to improve the code predictions for the conditions similar to the KTH tests. The film boiling correlations were selected among the most commonly used film boiling correlations found in the open literature, namely Groeneveld 5.7, Bishop (2 correlations), Tong, Konkov, Miropolskii and Groeneveld-Delorme correlations. The only correlation among the

  1. 20. RW Meyer Sugar Mill: 18761889. Boiling House Interior, 1878. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    20. RW Meyer Sugar Mill: 1876-1889. Boiling House Interior, 1878. View: Remains of south wall. The molasses storage pits are below the floor in the foreground. The remaining piece of floor indicates the form of the entire floor. The sorghum pan and boiling range flue slope from left to right (east to west) and permitted batches of cane juice to flow through the boiling pan by gravity. The beams, joists, truss work are built of northwest pine. The sides and floor boards are built of redwood. The boiling range flue is built of fire-brick, masonry, and portland cement. The corrugated roof appears to be a later addition, not contemporary with mill operation. - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  2. SU-G-206-01: A Fully Automated CT Tool to Facilitate Phantom Image QA for Quantitative Imaging in Clinical Trials

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wahi-Anwar, M; Lo, P; Kim, H

    Purpose: The use of Quantitative Imaging (QI) methods in Clinical Trials requires both verification of adherence to a specified protocol and an assessment of scanner performance under that protocol, which are currently accomplished manually. This work introduces automated phantom identification and image QA measure extraction towards a fully-automated CT phantom QA system to perform these functions and facilitate the use of Quantitative Imaging methods in clinical trials. Methods: This study used a retrospective cohort of CT phantom scans from existing clinical trial protocols - totaling 84 phantoms, across 3 phantom types using various scanners and protocols. The QA system identifiesmore » the input phantom scan through an ensemble of threshold-based classifiers. Each classifier - corresponding to a phantom type - contains a template slice, which is compared to the input scan on a slice-by-slice basis, resulting in slice-wise similarity metric values for each slice compared. Pre-trained thresholds (established from a training set of phantom images matching the template type) are used to filter the similarity distribution, and the slice with the most optimal local mean similarity, with local neighboring slices meeting the threshold requirement, is chosen as the classifier’s matched slice (if it existed). The classifier with the matched slice possessing the most optimal local mean similarity is then chosen as the ensemble’s best matching slice. If the best matching slice exists, image QA algorithm and ROIs corresponding to the matching classifier extracted the image QA measures. Results: Automated phantom identification performed with 84.5% accuracy and 88.8% sensitivity on 84 phantoms. Automated image quality measurements (following standard protocol) on identified water phantoms (n=35) matched user QA decisions with 100% accuracy. Conclusion: We provide a fullyautomated CT phantom QA system consistent with manual QA performance. Further work will include

  3. Alternative splicing of class Ib major histocompatibility complex transcripts in vivo leads to the expression of soluble Qa-2 molecules in murine blood.

    PubMed Central

    Tabaczewski, P; Shirwan, H; Lewis, K; Stroynowski, I

    1994-01-01

    Class Ib Qa-2 molecules are expressed in tissue culture cells as approximately 40-kDa membrane-bound, glycophosphatidylinositol-linked antigens and as approximately 39-kDa soluble polypeptides. Recently, alternative splicing events which delete exon 5 from a portion of Qa-2 transcripts were demonstrated to give rise to truncated secreted Qa-2 molecules in transfected cell lines. To determine whether this mechanism operates in vivo and to find out whether Qa-2 can be detected in soluble form in circulation, murine blood samples were analyzed. Critical to these experiments was preparation of an anti-peptide antiserum against an epitope encoded by a junction of exon 4 and exon 6. We find that supernatants of splenocytes cultured in vitro as well as serum or plasma contain two forms of soluble Qa-2 molecules. One form corresponds to a secreted molecule translated from transcripts from which exon 5 has been deleted; the other is derived from membrane-bound antigens or their precursors. The levels of both soluble forms of Qa-2 are inducible upon stimulation of the immune system, suggesting an immunoregulatory role for these molecules or for the mechanism leading to the reduction of cell-associated Qa-2 antigens in vivo. Images PMID:8127900

  4. Early Onset of Nucleate Boiling on Gas-covered Biphilic Surfaces.

    PubMed

    Shen, Biao; Yamada, Masayuki; Hidaka, Sumitomo; Liu, Jiewei; Shiomi, Junichiro; Amberg, Gustav; Do-Quang, Minh; Kohno, Masamichi; Takahashi, Koji; Takata, Yasuyuki

    2017-05-17

    For phase-change cooling schemes for electronics, quick activation of nucleate boiling helps safeguard the electronics components from thermal shocks associated with undesired surface superheating at boiling incipience, which is of great importance to the long-term system stability and reliability. Previous experimental studies show that bubble nucleation can occur surprisingly early on mixed-wettability surfaces. In this paper, we report unambiguous evidence that such unusual bubble generation at extremely low temperatures-even below the boiling point-is induced by a significant presence of incondensable gas retained by the hydrophobic surface, which exhibits exceptional stability even surviving extensive boiling deaeration. By means of high-speed imaging, it is revealed that the consequently gassy boiling leads to unique bubble behaviour that stands in sharp contrast with that of pure vapour bubbles. Such findings agree qualitatively well with numerical simulations based on a diffuse-interface method. Moreover, the simulations further demonstrate strong thermocapillary flows accompanying growing bubbles with considerable gas contents, which is associated with heat transfer enhancement on the biphilic surface in the low-superheat region.

  5. EVALUATION OF CONFOCAL MICROSCOPY SYSTEM PERFORMANCE: PRETTY PICTURES OR CONFOCAL QA

    EPA Science Inventory

    Evaluation of confocal microscopy system performance: Pretty pictures or confocal QA?

    Robert M. Zucker

    Reproductive Toxicology Division, National Health and Environmental Effects Research Laboratory, U.S. Environmental Protection Agency, Research Triangle Park, N...

  6. Boiling local heat transfer enhancement in minichannels using nanofluids

    PubMed Central

    2013-01-01

    This paper reports an experimental study on nanofluid convective boiling heat transfer in parallel rectangular minichannels of 800 μm hydraulic diameter. Experiments are conducted with pure water and silver nanoparticles suspended in water base fluid. Two small volume fractions of silver nanoparticles suspended in water are tested: 0.000237% and 0.000475%. The experimental results show that the local heat transfer coefficient, local heat flux, and local wall temperature are affected by silver nanoparticle concentration in water base fluid. In addition, different correlations established for boiling flow heat transfer in minichannels or macrochannels are evaluated. It is found that the correlation of Kandlikar and Balasubramanian is the closest to the water boiling heat transfer results. The boiling local heat transfer enhancement by adding silver nanoparticles in base fluid is not uniform along the channel flow. Better performances and highest effect of nanoparticle concentration on the heat transfer are obtained at the minichannels entrance. PMID:23506445

  7. 46 CFR 154.705 - Cargo boil-off as fuel: General.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... STANDARDS FOR SELF-PROPELLED VESSELS CARRYING BULK LIQUEFIED GASES Design, Construction and Equipment Cargo Pressure and Temperature Control § 154.705 Cargo boil-off as fuel: General. (a) Each cargo boil-off fuel...

  8. 46 CFR 154.705 - Cargo boil-off as fuel: General.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... STANDARDS FOR SELF-PROPELLED VESSELS CARRYING BULK LIQUEFIED GASES Design, Construction and Equipment Cargo Pressure and Temperature Control § 154.705 Cargo boil-off as fuel: General. (a) Each cargo boil-off fuel...

  9. 46 CFR 154.705 - Cargo boil-off as fuel: General.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... STANDARDS FOR SELF-PROPELLED VESSELS CARRYING BULK LIQUEFIED GASES Design, Construction and Equipment Cargo Pressure and Temperature Control § 154.705 Cargo boil-off as fuel: General. (a) Each cargo boil-off fuel...

  10. 46 CFR 154.705 - Cargo boil-off as fuel: General.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... STANDARDS FOR SELF-PROPELLED VESSELS CARRYING BULK LIQUEFIED GASES Design, Construction and Equipment Cargo Pressure and Temperature Control § 154.705 Cargo boil-off as fuel: General. (a) Each cargo boil-off fuel...

  11. 46 CFR 154.705 - Cargo boil-off as fuel: General.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... STANDARDS FOR SELF-PROPELLED VESSELS CARRYING BULK LIQUEFIED GASES Design, Construction and Equipment Cargo Pressure and Temperature Control § 154.705 Cargo boil-off as fuel: General. (a) Each cargo boil-off fuel...

  12. SU-F-T-300: Impact of Electron Density Modeling of ArcCHECK Cylindricaldiode Array On 3DVH Patient Specific QA Software Tool Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Patwe, P; Mhatre, V; Dandekar, P

    Purpose: 3DVH software is a patient specific quality assurance tool which estimates the 3D dose to the patient specific geometry with the help of Planned Dose Perturbation algorithm. The purpose of this study is to evaluate the impact of HU value of ArcCHECK phantom entered in Eclipse TPS on 3D dose & DVH QA analysis. Methods: Manufacturer of ArcCHECK phantom provides CT data set of phantom & recommends considering it as a homogeneous phantom with electron density (1.19 gm/cc or 282 HU) close to PMMA. We performed this study on Eclipse TPS (V13, VMS) & trueBEAM STx VMS Linac &more » ArcCHECK phantom (SNC). Plans were generated for 6MV photon beam, 20cm×20cm field size at isocentre & SPD (Source to phantom distance) of 86.7 cm to deliver 100cGy at isocentre. 3DVH software requires patients DICOM data generated by TPS & plan delivered on ArcCHECK phantom. Plans were generated in TPS by assigning different HU values to phantom. We analyzed gamma index & the dose profile for all plans along vertical down direction of beam’s central axis for Entry, Exit & Isocentre dose. Results: The global gamma passing rate (2% & 2mm) for manufacturer recommended HU value 282 was 96.3%. Detector entry, Isocentre & detector exit Doses were 1.9048 (1.9270), 1.00(1.0199) & 0.5078(0.527) Gy for TPS (Measured) respectively.The global gamma passing rate for electron density 1.1302 gm/cc was 98.6%. Detector entry, Isocentre & detector exit Doses were 1.8714 (1.8873), 1.00(0.9988) & 0.5211(0.516) Gy for TPS (Measured) respectively. Conclusion: Electron density value assigned by manufacturer does not hold true for every user. Proper modeling of electron density of ArcCHECK in TPS is essential to avoid systematic error in dose calculation of patient specific QA.« less

  13. Film Boiling Heat Transfer Properties of Liquid Hydrogen in Natural Convection

    NASA Astrophysics Data System (ADS)

    Horie, Y.; Shirai, Y.; Shiotsu, M.; Matsuzawa, T.; Yoneda, K.; Shigeta, H.; Tatsumoto, H.; Hata, K.; Naruo, Y.; Kobayashi, H.; Inatani, Y.

    Film boiling heat transfer properties of LH2 for various pressures and subcooling conditions were measured by applying electric current to give an exponential heat input to a PtCo wire with a diameter of 1.2 mm submerged in LH2. The heated wire was set to be horizontal to the ground. The heat transfer coefficient in the film boiling region was higher for higher pressure and higher subcooling. The experimental results are compared with the equation of pool film boiling heat transfer. It is confirmed that the pool film boiling heat transfer coefficients in LH2 can be expressed by this equation.

  14. Spinal pedicle screw planning using deformable atlas registration.

    PubMed

    Goerres, J; Uneri, A; De Silva, T; Ketcha, M; Reaungamornrat, S; Jacobson, M; Vogt, S; Kleinszig, G; Osgood, G; Wolinsky, J-P; Siewerdsen, J H

    2017-04-07

    Spinal screw placement is a challenging task due to small bone corridors and high risk of neurological or vascular complications, benefiting from precision guidance/navigation and quality assurance (QA). Implicit to both guidance and QA is the definition of a surgical plan-i.e. the desired trajectories and device selection for target vertebrae-conventionally requiring time-consuming manual annotations by a skilled surgeon. We propose automation of such planning by deriving the pedicle trajectory and device selection from a patient's preoperative CT or MRI. An atlas of vertebrae surfaces was created to provide the underlying basis for automatic planning-in this work, comprising 40 exemplary vertebrae at three levels of the spine (T7, T8, and L3). The atlas was enriched with ideal trajectory annotations for 60 pedicles in total. To define trajectories for a given patient, sparse deformation fields from the atlas surfaces to the input (CT or MR image) are applied on the annotated trajectories. Mean value coordinates are used to interpolate dense deformation fields. The pose of a straight trajectory is optimized by image-based registration to an accumulated volume of the deformed annotations. For evaluation, input deformation fields were created using coherent point drift (CPD) to perform a leave-one-out analysis over the atlas surfaces. CPD registration demonstrated surface error of 0.89  ±  0.10 mm (median  ±  interquartile range) for T7/T8 and 1.29  ±  0.15 mm for L3. At the pedicle center, registered trajectories deviated from the expert reference by 0.56  ±  0.63 mm (T7/T8) and 1.12  ±  0.67 mm (L3). The predicted maximum screw diameter differed by 0.45  ±  0.62 mm (T7/T8), and 1.26  ±  1.19 mm (L3). The automated planning method avoided screw collisions in all cases and demonstrated close agreement overall with expert reference plans, offering a potentially valuable tool in support of

  15. Marangoni Effects in the Boiling of Binary Fluid Mixtures

    NASA Technical Reports Server (NTRS)

    Ahmed, Sayeed; Carey, Van P.; Motil, Brian

    1996-01-01

    Results of very recent experimental studies indicate that during nucleate boiling in some binary mixture, Marangoni effects augment the gravity driven flow of liquid towards the heated surface. With gravity present, it is impossible to separate the two effects. The reduced gravity environment gives an unique opportunity to explore th role of Marangoni effects on the boiling mechanisms free of gravitational body forces that obscure the role of such effects. However, recent experimental results suggest that under reduced gravity conditions, Marangoni effects is the dominant mechanism of vapor-liquid exchange at the surface for some binary mixture. To further explore such effects, experiments have been conducted with water/2-propanol mixtures at three different concentrations under normal gravity with different orientations of the heater surface and under reduce gravity aboard the DC-9 aircraft at NASA Lewis Research Center. The system pressure was sub atmospheric (approx. 8 kP at 1g(n)) and the bulk liquid temperature varied from low subcooling to near saturation. The molar concentrations of 2-propanol tested were 0.015, 0.025, and 0.1. Boiling curves were obtained both for high gravity (approx. 2g(n)) and reduce gravity (approx. 0.01g(n)). For each concentration of 2-propanol, the critical heat flux has been determined in the flight experiments only for reduced gravity conditions. Comparison of boiling curves and CHF obtained under l-g(n) an reduced gravity indicates that boiling mechanism in this mixtures is nearly independent of gravity. The results also indicate that the Marangoni mechanism is strong enough in these mixtures to sustain the boiling under reduced gravity conditions.

  16. Planning, guidance, and quality assurance of pelvic screw placement using deformable image registration

    NASA Astrophysics Data System (ADS)

    Goerres, J.; Uneri, A.; Jacobson, M.; Ramsay, B.; De Silva, T.; Ketcha, M.; Han, R.; Manbachi, A.; Vogt, S.; Kleinszig, G.; Wolinsky, J.-P.; Osgood, G.; Siewerdsen, J. H.

    2017-12-01

    Percutaneous pelvic screw placement is challenging due to narrow bone corridors surrounded by vulnerable structures and difficult visual interpretation of complex anatomical shapes in 2D x-ray projection images. To address these challenges, a system for planning, guidance, and quality assurance (QA) is presented, providing functionality analogous to surgical navigation, but based on robust 3D-2D image registration techniques using fluoroscopy images already acquired in routine workflow. Two novel aspects of the system are investigated: automatic planning of pelvic screw trajectories and the ability to account for deformation of surgical devices (K-wire deflection). Atlas-based registration is used to calculate a patient-specific plan of screw trajectories in preoperative CT. 3D-2D registration aligns the patient to CT within the projective geometry of intraoperative fluoroscopy. Deformable known-component registration (dKC-Reg) localizes the surgical device, and the combination of plan and device location is used to provide guidance and QA. A leave-one-out analysis evaluated the accuracy of automatic planning, and a cadaver experiment compared the accuracy of dKC-Reg to rigid approaches (e.g. optical tracking). Surgical plans conformed within the bone cortex by 3-4 mm for the narrowest corridor (superior pubic ramus) and  >5 mm for the widest corridor (tear drop). The dKC-Reg algorithm localized the K-wire tip within 1.1 mm and 1.4° and was consistently more accurate than rigid-body tracking (errors up to 9 mm). The system was shown to automatically compute reliable screw trajectories and accurately localize deformed surgical devices (K-wires). Such capability could improve guidance and QA in orthopaedic surgery, where workflow is impeded by manual planning, conventional tool trackers add complexity and cost, rigid tool assumptions are often inaccurate, and qualitative interpretation of complex anatomy from 2D projections is prone to trial

  17. Reduced Boil-Off System Sizing

    NASA Technical Reports Server (NTRS)

    Guzik, Monica C.; Plachta, David W.; Feller, Jeffrey R.

    2015-01-01

    NASA is currently developing cryogenic propellant storage and transfer systems for future space exploration and scientific discovery missions by addressing the need to raise the technology readiness level of cryogenic fluid management technologies. Cryogenic propellants are baselined in many propulsion systems due to their inherently high specific impulse; however, their low boiling points can cause substantial boil-off losses over time. Recent efforts such as the Reduced Boil-off Testing and the Active Thermal Control Scaling Study provide important information on the benefit of an active cooling system applied to LH2 propellant storage. Findings show that zero-boil off technologies can reduce overall mass in LH2 storage systems when low Earth orbit loiter periods extend beyond two months. A significant part of this mass reduction is realized by integrating two stages of cooling: a 20 K stage to intercept heat at the tank surface, and a 90 K stage to reduce the heat entering the less efficient 20 K stage. A missing element in previous studies, which is addressed in this paper, is the development of a direct method for sizing the 90 K cooling stage. Such a method requires calculation of the heat entering both the 90 K and 20 K stages as compared to the overall system masses, and is reliant upon the temperature distribution, performance, and unique design characteristics of the system in question. By utilizing the known conductance of a system without active thermal control, the heat being intercepted by a 90 K stage can be calculated to find the resultant lift and mass of each active thermal control stage. Integral to this is the thermal conductance of the cooling straps and the broad area cooling shield, key parts of the 90 K stage. Additionally, a trade study is performed to show the ability of the 90 K cooling stage to reduce the lift on the 20 K cryocooler stage, which is considerably less developed and efficient than 90 K cryocoolers.

  18. Development of concrete QC/QA specifications for highway construction in Kentucky.

    DOT National Transportation Integrated Search

    2001-08-01

    There is a growing trend toward quality-based specifications in highway construction. A large number of quality control/quality assurance (QC/QA) specifications shift the responsibility of day-to-day testing from the state DOH to the contractor. This...

  19. PREPARATION AIDS FOR THE DEVELOPMENT OF CATEGORY II QUALITY ASSURANCE PROJECT PLANS

    EPA Science Inventory

    Data collection activities performed for the Risk Reduction Engineering aboratory (RREL) of the U.S. Environmental Protection Agency are divided into tour categories, depending on the intended use of the data. uality Assurance (QA) Project Plans are written to ensure that project...

  20. PREPARATION AIDS FOR THE DEVELOPMENT OF CATEGORY I QUALITY ASSURANCE PROJECT PLANS

    EPA Science Inventory

    Data collection activities performed for the Risk Reduction Engineering aboratory (RREL) of the U.S. Environmental Protection Agency are divided into tour categories, depending on the intended use of the data. uality Assurance (QA) Project Plans are written to ensure that project...

  1. PREPARATION AIDS FOR THE DEVELOPMENT OF CATEGORY III QUALITY ASSURANCE PROJECT PLANS

    EPA Science Inventory

    Data collection activities performed for the Risk Reduction Engineering Laboratory (RREL) of the U.S. Environmental Protection Agency are divided into four categories, depending on the intended use of the data. uality Assurance (QA) Project Plans are written to ensure that projec...

  2. PREPARATION AIDS FOR THE DEVELOPMENT OF CATEGORY IV QUALITY ASSURANCE PROJECT PLANS

    EPA Science Inventory

    Data collection activities performed for the Risk Reduction Engineering Laboratory (RREL) of the U.S. Environmental Protection Agency are divided into four categories, depending on the intended use of the data. uality Assurance (QA) Project Plans are written to ensure that projec...

  3. Nucleate pool boiling in subcooled liquid under microgravity: Results of TEXUS experimental investigations

    NASA Astrophysics Data System (ADS)

    Zell, M.; Straub, J.; Weinzierl, A.

    1984-12-01

    Experiments on subcooled nucleate pool boiling in microgravity were carried out to separate gravity driven effects on heat transfer within the boiling process. A ballistic trajectory by sounding rocket flight (TEXUS 5 and 10) achieved a gravity level of a/g = 0.0001 for 360 sec. For determination of geometrical effects on heat transport two different experimental configurations (platinum wire and flat plate) were employed. Boiling curves and bubble dynamics recorded by cinematography lead to gravity independent modelling of the boiling phenomena. The results ensure the applicability and high efficiency of nucleate pool boiling for heat exchangers in space laboratories.

  4. Zero Boil-OFF Tank Hardware Setup

    NASA Image and Video Library

    2017-09-19

    iss053e027051 (Sept. 19, 2017) --- Flight Engineer Joe Acaba works in the U.S. Destiny laboratory module setting up hardware for the Zero Boil-Off Tank (ZBOT) experiment. ZBOT uses an experimental fluid to test active heat removal and forced jet mixing as alternative means for controlling tank pressure for volatile fluids. Rocket fuel, spacecraft heating and cooling systems, and sensitive scientific instruments rely on very cold cryogenic fluids. Heat from the environment around cryogenic tanks can cause their pressures to rise, which requires dumping or "boiling off" fluid to release the excess pressure, or actively cooling the tanks in some way.

  5. The Boiling eXperiment Facility (BXF) for the Microgravity Science Glovebox (MSG)

    NASA Technical Reports Server (NTRS)

    McQuillen, John; Chao, David; Vergilii, Frank

    2006-01-01

    Boiling is an effective means of cooling by removing heat from surfaces through vaporization of a working fluid. It is also affected by both the magnitude and direction of gravity. By conducting pool boiling tests in microgravity, the effect of buoyancy n the overall boiling process and the relative magnitude of other phenomena can be assessed. The Boiling eXperiment Facility (BXF) is being built for the Microgravity Science Glovebox. This facility will conduct two pool boiling studies. The first study the Microheater Array Boiling Experiment (MABE) uses two 96 element microheater arrays, 2.7 mm and 7.0 mm in size, to measure localized hear fluxes while operating at a constant temperature. The other experiment, the Nucleate Pool Boiling eXperiment (NPBX) uses a 85 mm diameter heater wafer that has been "seeded" with five individually-controlled nucleation sites to study bubble nucleation, growth, coalescence and departure. The BXF uses normal-perfluorohexane as the test fluid and will operate between pressures of 60 to 244 Pa. and temperatures of 35 to 60 C. Both sets of experimental heaters are highly instrumented. Pressure and bulk fluid temperature measurements will be made with standard rate video. A high speed video system will be used to visualize the boiling process through the bottom of the MABE heater arrays. The BXF is currently scheduled to fly on Utilization Flight-13A.1 to the ISS with facility integration into the MSG and operation during Increment 15

  6. An Investigation of Graduate Scientists' Understandings of Evaporation and Boiling.

    ERIC Educational Resources Information Center

    Goodwin, Alan; Orlik, Yuri

    2000-01-01

    Uses a video presentation of six situations relating to the evaporation and boiling of liquids and the escape of dissolved gases from solution and investigates graduate scientists' understanding of the concepts of boiling and evaporation. (Author/YDS)

  7. New Departure from Nucleate Boiling model relying on first principle energy balance at the boiling surface

    NASA Astrophysics Data System (ADS)

    Demarly, Etienne; Baglietto, Emilio

    2017-11-01

    Predictions of Departure from Nucleate Boiling have been a longstanding challenge when designing heat exchangers such as boilers or nuclear reactors. Many mechanistic models have been postulated over more than 50 years in order to explain this phenomenon but none is able to predict accurately the conditions which trigger the sudden change of heat transfer mode. This work aims at demonstrating the pertinence of a new approach for detecting DNB by leveraging recent experimental insights. The new model proposed departs from all the previous models by making the DNB inception come from an energy balance instability at the heating surface rather than a hydrodynamic instability of the bubbly layer above the surface (Zuber, 1959). The main idea is to modulate the amount of heat flux being exchanged via the nucleate boiling mechanism by the wetted area fraction on the surface, thus allowing a completely automatic trigger of DNB that doesn't require any parameter prescription. This approach is implemented as a surrogate model in MATLAB in order to validate the principles of the model in a simple and controlled geometry. Good agreement is found with the experimental data leveraged from the MIT Flow Boiling at various flow regimes. Consortium for Advanced Simulation of Light Water Reactors (CASL).

  8. Incorrect dosimetric leaf separation in IMRT and VMAT treatment planning: Clinical impact and correlation with pretreatment quality assurance.

    PubMed

    Sjölin, Maria; Edmund, Jens Morgenthaler

    2016-07-01

    Dynamic treatment planning algorithms use a dosimetric leaf separation (DLS) parameter to model the multi-leaf collimator (MLC) characteristics. Here, we quantify the dosimetric impact of an incorrect DLS parameter and investigate whether common pretreatment quality assurance (QA) methods can detect this effect. 16 treatment plans with intensity modulated radiation therapy (IMRT) or volumetric modulated arc therapy (VMAT) technique for multiple treatment sites were calculated with a correct and incorrect setting of the DLS, corresponding to a MLC gap difference of 0.5mm. Pretreatment verification QA was performed with a bi-planar diode array phantom and the electronic portal imaging device (EPID). Measurements were compared to the correct and incorrect planned doses using gamma evaluation with both global (G) and local (L) normalization. Correlation, specificity and sensitivity between the dose volume histogram (DVH) points for the planning target volume (PTV) and the gamma passing rates were calculated. The change in PTV and organs at risk DVH parameters were 0.4-4.1%. Good correlation (>0.83) between the PTVmean dose deviation and measured gamma passing rates was observed. Optimal gamma settings with 3%L/3mm (per beam and composite plan) and 3%G/2mm (composite plan) for the diode array phantom and 2%G/2mm (composite plan) for the EPID system were found. Global normalization and per beam ROC analysis of the diode array phantom showed an area under the curve <0.6. A DLS error can worsen pretreatment QA using gamma analysis with reasonable credibility for the composite plan. A low detectability was demonstrated for a 3%G/3mm per beam gamma setting. Copyright © 2016 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  9. Gravity and Heater Size Effects on Pool Boiling Heat Transfer

    NASA Technical Reports Server (NTRS)

    Kim, Jungho; Raj, Rishi

    2014-01-01

    The current work is based on observations of boiling heat transfer over a continuous range of gravity levels between 0g to 1.8g and varying heater sizes with a fluorinert as the test liquid (FC-72/n-perfluorohexane). Variable gravity pool boiling heat transfer measurements over a wide range of gravity levels were made during parabolic flight campaigns as well as onboard the International Space Station. For large heaters and-or higher gravity conditions, buoyancy dominated boiling and heat transfer results were heater size independent. The power law coefficient for gravity in the heat transfer equation was found to be a function of wall temperature under these conditions. Under low gravity conditions and-or for smaller heaters, surface tension forces dominated and heat transfer results were heater size dependent. A pool boiling regime map differentiating buoyancy and surface tension dominated regimes was developed along with a unified framework that allowed for scaling of pool boiling over a wide range of gravity levels and heater sizes. The scaling laws developed in this study are expected to allow performance quantification of phase change based technologies under variable gravity environments eventually leading to their implementation in space based applications.

  10. SU-E-T-636: ProteusONE Machine QA Procedure and Stabiity Study: Half Year Clinical Operation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Freund, D; Ding, X; Wu, H

    2015-06-15

    Purpose: The objective of this study is to evaluate the stability of ProteusOne, the 1st commercial PBS proton system, throughout the daily QA and monthly over 6 month clinical operation. Method: Daily QA test includes IGRT position/repositioning, output in the middle of SOBP, beam flatness, symmetry, inplane and crossplane dimensions as well as energy range check. Daily range shifter QA consist of output, symmetry and field size checks to make sure its integrity. In 30 mins Daily QA test, all the measurements are performed using the MatriXXPT (IBA dosimetry). The data from these measurement was collected and compare over themore » first 6 month of clinical operation. In addition to the items check in daily QA, the summary also includes the monthly QA gantry star shots, absolute position check using a novel device, XRV-100. Results: Average machine output at the center of the spread out bragg peak was 197.5±.8 cGy and was within 1%of the baseline at 198.4 cGy. Beam flatness was within 1% cross plane with an average of 0.67±0.12% and 2% in-plane with an average of 1.08±0.17% compared to baseline measurements of 0.6 and 1.03, respectively. In all cases the radiation isocenter shift was less than or equal to 1mm. Output for the range shifter was within 2% for each individual measurement and averaged 34.4±.2cGy compare to a baseline reading of 34.5cGy. The average range shifter in and cross plane field size measurements were 19.8±0.5cm and 20.5±0.4cm compared with baseline values of 20.19cm and 20.79cm, respectively. Range shifter field symmetry had an average of less 1% for both in-plane and cross plane measurements. Conclusion: All machine metrics over the past 6 months have proved to be stable. Although, some averages are outside the baseline measurement they are within 1% tolerance and the deviation across all measurements is minimal.« less

  11. A numerical investigation of the effect of surface wettability on the boiling curve.

    PubMed

    Hsu, Hua-Yi; Lin, Ming-Chieh; Popovic, Bridget; Lin, Chii-Ruey; Patankar, Neelesh A

    2017-01-01

    Surface wettability is recognized as playing an important role in pool boiling and the corresponding heat transfer curve. In this work, a systematic study of pool boiling heat transfer on smooth surfaces of varying wettability (contact angle range of 5° - 180°) has been conducted and reported. Based on numerical simulations, boiling curves are calculated and boiling dynamics in each regime are studied using a volume-of-fluid method with contact angle model. The calculated trends in critical heat flux and Leidenfrost point as functions of surface wettability are obtained and compared with prior experimental and theoretical predictions, giving good agreement. For the first time, the effect of contact angle on the complete boiling curve is shown. It is demonstrated that the simulation methodology can be used for studying pool boiling and related dynamics and providing more physical insights.

  12. A numerical investigation of the effect of surface wettability on the boiling curve

    PubMed Central

    Lin, Ming-Chieh; Popovic, Bridget; Lin, Chii-Ruey; Patankar, Neelesh A.

    2017-01-01

    Surface wettability is recognized as playing an important role in pool boiling and the corresponding heat transfer curve. In this work, a systematic study of pool boiling heat transfer on smooth surfaces of varying wettability (contact angle range of 5° − 180°) has been conducted and reported. Based on numerical simulations, boiling curves are calculated and boiling dynamics in each regime are studied using a volume-of-fluid method with contact angle model. The calculated trends in critical heat flux and Leidenfrost point as functions of surface wettability are obtained and compared with prior experimental and theoretical predictions, giving good agreement. For the first time, the effect of contact angle on the complete boiling curve is shown. It is demonstrated that the simulation methodology can be used for studying pool boiling and related dynamics and providing more physical insights. PMID:29125847

  13. A high-fidelity approach towards simulation of pool boiling

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yazdani, Miad; Radcliff, Thomas; Soteriou, Marios

    2016-01-15

    A novel numerical approach is developed to simulate the multiscale problem of pool-boiling phase change. The particular focus is to develop a simulation technique that is capable of predicting the heat transfer and hydrodynamic characteristics of nucleate boiling and the transition to critical heat flux on surfaces of arbitrary shape and roughness distribution addressing a critical need to design enhanced boiling heat transfer surfaces. The macro-scale of the phase change and bubble dynamics is addressed through employing off-the-shelf Computational Fluid Dynamics (CFD) methods for interface tracking and interphase mass and energy transfer. The micro-scale of the microlayer, which forms atmore » early stage of bubble nucleation near the wall, is resolved through asymptotic approximation of the thin-film theory which provides a closed-form solution for the distribution of the micro-layer and its influence on the evaporation process. In addition, the sub-grid surface roughness is represented stochastically through probabilistic density functions and its role in bubble nucleation and growth is then represented based on the thermodynamics of nucleation process. This combination of deterministic CFD, local approximation, and stochastic representation allows the simulation of pool boiling on any surface with known roughness and enhancement characteristics. The numerical model is validated for dynamics and hydrothermal characteristics of a single nucleated bubble on a flat surface against available literature data. In addition, the prediction of pool-boiling heat transfer coefficient is verified against experimental measurements as well as reputable correlations for various roughness distributions and different surface orientations. Finally, the model is employed to demonstrate pool-boiling phenomenon on enhanced structures with reentrance cavities and to explore the effect of enhancement feature design on thermal and hydrodynamic characteristics of these surfaces.« less

  14. A study of forced convection boiling under reduced gravity

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.

    1992-01-01

    This report presents the results of activities conducted over the period 1/2/85-12/31/90, in which the study of forced convection boiling under reduced gravity was initiated. The study seeks to improve the understanding of the basic processes that constitute forced convection boiling by removing the buoyancy effects which may mask other phenomena. Specific objectives may also be expressed in terms of the following questions: (1) what effects, if any, will the removal of body forces to the lowest possible levels have on the forced convection boiling heat transfer processes in well-defined and meaningful circumstances? (this includes those effects and processes associated with the nucleation or onset of boiling during the transient increase in heater surface temperature, as well as the heat transfer and vapor bubble behaviors with established or steady-state conditions); and (2) if such effects are present, what are the boundaries of the relevant parameters such as heat flux, heater surface superheat, fluid velocity, bulk subcooling, and geometric/orientation relationships within which such effects will be produced?

  15. Boiling process modelling peculiarities analysis of the vacuum boiler

    NASA Astrophysics Data System (ADS)

    Slobodina, E. N.; Mikhailov, A. G.

    2017-06-01

    The analysis of the low and medium powered boiler equipment development was carried out, boiler units possible development directions with the purpose of energy efficiency improvement were identified. Engineering studies for the vacuum boilers applying are represented. Vacuum boiler heat-exchange processes where boiling water is the working body are considered. Heat-exchange intensification method under boiling at the maximum heat- transfer coefficient is examined. As a result of the conducted calculation studies, heat-transfer coefficients variation curves depending on the pressure, calculated through the analytical and numerical methodologies were obtained. The conclusion about the possibility of numerical computing method application through RPI ANSYS CFX for the boiling process description in boiler vacuum volume was given.

  16. Using a User-Interactive QA System for Personalized E-Learning

    ERIC Educational Resources Information Center

    Hu, Dawei; Chen, Wei; Zeng, Qingtian; Hao, Tianyong; Min, Feng; Wenyin, Liu

    2008-01-01

    A personalized e-learning framework based on a user-interactive question-answering (QA) system is proposed, in which a user-modeling approach is used to capture personal information of students and a personalized answer extraction algorithm is proposed for personalized automatic answering. In our approach, a topic ontology (or concept hierarchy)…

  17. SU-E-T-662: Quick and Efficient Daily QA for Compact PBS Proton Therapy Machine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Patel, B; Syh, J; Ding, X

    2015-06-15

    Purpose: As proton therapy machines become widespread the need for a quick simple routine daily QA like that for linear accelerators becomes more important. Willis-Knighton has developed an accurate and efficient daily QA that can be performed in 15 minutes. Methods: A holder for a 2D ionization chamber array (MatriXX PT) was created that is indexed to the couch to allow for quick setup, lasers accuracy with respect to beam isocenter, and couch reproducibility. Image position/reposition was performed to check Isocentricity accuracy by placing BBs on the MatriXX. The couch coordinates are compared to that of commissioning. Laser positions weremore » confirmed with the MatriXX isocenter. After IGRT, three beams were separately delivered according to setup. For the first beam, range shifter was inserted and dose at R90, field size, flatness and symmetry in X and Y direction was measured. R90 was used so any minor changes in the range shifter can be detected. For the open beam, dose at center of SOBP, flatness and symmetry in X and Y direction was measured. Field size was measured in ±X and ±Y direction at FWHM. This is measured so any variation in spot size will be detected. For the third beam additional solid water was added and dose at R50 was measured so that any variation in beam energy will be detected. Basic mechanical and safety checks were also performed. Results: Medical physicists were able to complete the daily QA and reduce the time by half to two-third from initial daily QA procedure. All the values measured were within tolerance of that of the baseline which was established from water tank and initial MatriXX measurements. Conclusion: The change in daily QA procedure resulted in quick and easy setup and was able to measure all the basic functionality of the proton therapy PBS.« less

  18. 17. RW Meyer Sugar Mill: 18761889. Boiling House, 1878. View: ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    17. RW Meyer Sugar Mill: 1876-1889. Boiling House, 1878. View: Southwest corner of boiling house. The amimal-powered cane mill is located in the undergrowth in the right foreground, - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  19. Extended hydrodynamic theory of the peak and minimum pool boiling heat fluxes

    NASA Technical Reports Server (NTRS)

    Linehard, J. H.; Dhir, V. K.

    1973-01-01

    The hydrodynamic theory of the extreme pool boiling heat fluxes is expanded to embrace a variety of problems that have not previously been analyzed. These problems include the prediction of the peak heat flux on a variety of finite heaters, the influence of viscosity on the Taylor and Helmoltz instability mechanisms with application to film boiling and to the peak heat flux in viscous liquids, the formalization of the analogy between high-current-density electrolysis and boiling, and the description of boiling in the low-gravity limit. The predictions are verified with a large number of new data.

  20. Enhanced Boiling on Micro-Configured Composite Surfaces Under Microgravity Conditions

    NASA Technical Reports Server (NTRS)

    Zhang, Nengli; Chai, An-Ti

    1999-01-01

    In order to accommodate the growing thermal management needs of future space platforms, several two-phase active thermal control systems (ATCSs) have evolved and were included in the designs of space stations. Compared to the pumped single-phase liquid loops used in the conventional Space Transportation System and Spacelab, ATCSs offer significant benefits that may be realized by adopting a two-phase fluid-loop system. Alternately, dynamic power systems (DPSs), based on the Rankine cycle, seem inevitably to be required to supply the electrical power requirements of expanding space activities. Boiling heat transfer is one of the key technologies for both ATCSs and DPSs. Nucleate boiling near critical heat flux (CHF) can transport very large thermal loads with much smaller device size and much lower pumping power. However, boiling performance deteriorates in a reduced gravity environment and operation in the CHF regime is precarious because any slight overload will cause the heat transfer to suddenly move to the film boiling regime, which in turn, will result in burnout of the heat transfer surfaces. New materials, such as micro-configured metal-graphite composites, can provide a solution for boiling enhancement. It has been shown experimentally that this type of material manifests outstanding boiling heat transfer performance and their CHF is also extended to higher values. Due to the high thermal conductivity of graphite fiber (up to 1,200 W/m-K in the fiber direction), the composite surfaces are non-isothermal during the boiling process. The composite surfaces are believed to have a much wider safe operating region (a more uniform boiling curve in the CHF regime) because non-isothermal surfaces have been found to be less sensitive to variations of wall superheat in the CHF regime. The thermocapillary forces formed by the temperature difference between the fiber tips and the metal matrix play a more important role than the buoyancy in the bubble detachment, for the

  1. Turning bubbles on and off during boiling using charged surfactants

    PubMed Central

    Cho, H. Jeremy; Mizerak, Jordan P.; Wang, Evelyn N.

    2015-01-01

    Boiling—a process that has powered industries since the steam age—is governed by bubble formation. State-of-the-art boiling surfaces often increase bubble nucleation via roughness and/or wettability modification to increase performance. However, without active in situ control of bubbles, temperature or steam generation cannot be adjusted for a given heat input. Here we report the ability to turn bubbles ‘on and off' independent of heat input during boiling both temporally and spatially via molecular manipulation of the boiling surface. As a result, we can rapidly and reversibly alter heat transfer performance up to an order of magnitude. Our experiments show that this active control is achieved by electrostatically adsorbing and desorbing charged surfactants to alter the wettability of the surface, thereby affecting nucleation. This approach can improve performance and flexibility in existing boiling technologies as well as enable emerging or unprecedented energy applications. PMID:26486275

  2. Poster - Thurs Eve-03: Dose verification using a 2D diode array (Mapcheck) for electron beam modeling, QA and patient customized cutouts.

    PubMed

    Ghasroddashti, E; Sawchuk, S

    2008-07-01

    To assess a diode detector array (MapCheck) for commissioning, quality assurance (QA); and patient specific QA for electrons. 2D dose information was captured for various depths at several square fields ranging from 2×2 to 25×25cm 2 , and 9 patient customized cutouts using both Mapcheck and a scanning water phantom. Beam energies of 6, 9, 12, 16 and 20 MeV produced by Varian linacs were used. The water tank, beam energies and fields were also modeled on the Pinnacle planning system obtaining dose information. Mapcheck, water phantom and Pinnacle results were compared. Relative output factors (ROF) acquired with Mapcheck were compared to an in-house algorithm (JeffIrreg). Inter- and intra-observer variability was also investigated Results: Profiles and %DD data for Mapcheck, water tank, and Pinnacle agree well. High-dose, low-dose-gradient comparisons agree to within 1% between Mapcheck and water phantom. Field size comparisons showed mostly sub-millimeter agreement. ROFs for Mapcheck and JeffIrreg agreed within 2.0% (mean=0.9%±0.6%). The current standard for electron commissioning and QA is the scanning water tank which may be inefficient. Our results demonstrate that MapCheck can potentially be an alternative. Also the dose distributions for patient specific electron treatment require verification. This procedure is particularly challenging when the minimum dimension across the central axis of the cutout is smaller than the range of the electrons in question. Mapcheck offers an easy and efficient way of determining patient dose distributions especially compared to using the alternatives, namely, ion chamber and film. © 2008 American Association of Physicists in Medicine.

  3. Zero Boil Off Cryogen Storage for Future Launchers

    NASA Technical Reports Server (NTRS)

    Valentian, D.; Plachta, D.; Kittel, P.; Hastings, L. J.; Salerno, Louis J.; Arnold, James O. (Technical Monitor)

    2001-01-01

    Zero boil off (ZBO) cryogen storage using both cryocoolers and passive insulation technologies will enable long-term exploration missions by allowing designers to optimize tankage without the need for excess cryogen storage to account for boil off. Studies of ZBO (zero boil off) have been on-going in the USA for several years. More recently, a review of the needs of advanced space propulsion took place in Europe. This showed the interest of the European community in cryogenic propulsion for planetary missions as well as the use of liquid hydrogen for large power electric propulsion (manned Mars missions). Although natural boiling could be acceptable for single leg missions, passive insulation techniques yield roughly a I% per month cryogen loss and this would not be cost effective for robotic planetary missions involving storage times greater than one year. To make economic sense, long-term exploration missions require lower tank capacity and longer storage times. Recent advances in cryocooler technology, resulting in vast improvements in both cooler efficiency and reliability, make ZBO is a clear choice for planetary exploration missions. Other, more near term applications of ZBO include boil-off reduction or elimination applied to first and upper stages of future earth-to-orbit (ETO) launchers. This would extend launch windows and reduce infrastructure costs. Successors to vehicles like Ariane 5 could greatly benefit by implementing ZBO. Zero Boil Off will only be successful in ETO launcher applications if it makes economic sense to implement. The energy cost is only a fraction of the total cost of buying liquid cryogen, the rest being transportation and other overhead. Because of this, higher boiling point cryogens will benefit more from on-board liquefaction, thus reducing the infrastructure costs. Since hydrogen requires a liquefier with at least a 17% efficiency just to break even from a cost standpoint, one approach for implementing ZBO in upper stages would

  4. Boiling regimes of impacting drops on a heated substrate under reduced pressure

    NASA Astrophysics Data System (ADS)

    van Limbeek, Michiel A. J.; Hoefnagels, Paul B. J.; Shirota, Minori; Sun, Chao; Lohse, Detlef

    2018-05-01

    We experimentally investigate the boiling behavior of impacting ethanol drops on a heated smooth sapphire substrate at pressures ranging from P =0.13 bar to atmospheric pressure. We employ frustrated total internal reflection imaging to study the wetting dynamics of the contact between the drop and the substrate. The spreading drop can be in full contact (contact boiling), it can partially touch (transition boiling), or the drop can be fully levitated (Leidenfrost boiling). We show that the temperature of the boundary between contact and transition boiling shows at most a weak dependence on the impact velocity, but a significant decrease with decreasing ambient gas pressure. A striking correspondence is found between the temperature of this boundary and the static Leidenfrost temperature for all pressures. We therefore conclude that both phenomena share the same mechanism and are dominated by the dynamics taking place at the contact line. On the other hand, the boundary between transition boiling and Leidenfrost boiling, i.e., the dynamic Leidenfrost temperature, increases for increasing impact velocity for all ambient gas pressures. Moreover, the dynamic Leidenfrost temperature coincides for pressures between P =0.13 and 0.54 bar, whereas for atmospheric pressure the dynamic Leidenfrost temperature is slightly elevated. This indicates that the dynamic Leidenfrost temperature is at most weakly dependent on the enhanced evaporation by the lower saturation temperature of the liquid.

  5. An Analytical Approach for Relating Boiling Points of Monofunctional Organic Compounds to Intermolecular Forces

    ERIC Educational Resources Information Center

    Struyf, Jef

    2011-01-01

    The boiling point of a monofunctional organic compound is expressed as the sum of two parts: a contribution to the boiling point due to the R group and a contribution due to the functional group. The boiling point in absolute temperature of the corresponding RH hydrocarbon is chosen for the contribution to the boiling point of the R group and is a…

  6. WE-AB-206-00: Diagnostic QA/QC Hands-On Workshop

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    The involvement of medical physicists in diagnostic ultrasound imaging service is increasing due to QC and accreditation requirements. The goal of this ultrasound hands-on workshop is to demonstrate quality control (QC) testing in diagnostic ultrasound and to provide updates in ACR ultrasound accreditation requirements. The first half of this workshop will include two presentations reviewing diagnostic ultrasound QA/QC and ACR ultrasound accreditation requirements. The second half of the workshop will include live demonstrations of basic QC tests. An array of ultrasound testing phantoms and ultrasound scanners will be available for attendees to learn diagnostic ultrasound QC in a hands-on environmentmore » with live demonstrations and on-site instructors. The targeted attendees are medical physicists in diagnostic imaging. Learning Objectives: Gain familiarity with common elements of a QA/QC program for diagnostic ultrasound imaging dentify QC tools available for testing diagnostic ultrasound systems and learn how to use these tools Learn ACR ultrasound accreditation requirements Jennifer Walter is an employee of American College of Radiology on Ultrasound Accreditation.« less

  7. Utilizing knowledge from prior plans in the evaluation of quality assurance

    NASA Astrophysics Data System (ADS)

    Stanhope, Carl; Wu, Q. Jackie; Yuan, Lulin; Liu, Jianfei; Hood, Rodney; Yin, Fang-Fang; Adamson, Justus

    2015-06-01

    Increased interest regarding sensitivity of pre-treatment intensity modulated radiotherapy and volumetric modulated arc radiotherapy (VMAT) quality assurance (QA) to delivery errors has led to the development of dose-volume histogram (DVH) based analysis. This paradigm shift necessitates a change in the acceptance criteria and action tolerance for QA. Here we present a knowledge based technique to objectively quantify degradations in DVH for prostate radiotherapy. Using machine learning, organ-at-risk (OAR) DVHs from a population of 198 prior patients’ plans were adapted to a test patient’s anatomy to establish patient-specific DVH ranges. This technique was applied to single arc prostate VMAT plans to evaluate various simulated delivery errors: systematic single leaf offsets, systematic leaf bank offsets, random normally distributed leaf fluctuations, systematic lag in gantry angle of the mutli-leaf collimators (MLCs), fluctuations in dose rate, and delivery of each VMAT arc with a constant rather than variable dose rate. Quantitative Analyses of Normal Tissue Effects in the Clinic suggests V75Gy dose limits of 15% for the rectum and 25% for the bladder, however the knowledge based constraints were more stringent: 8.48   ±   2.65% for the rectum and 4.90   ±   1.98% for the bladder. 19   ±   10 mm single leaf and 1.9   ±   0.7 mm single bank offsets resulted in rectum DVHs worse than 97.7% (2σ) of clinically accepted plans. PTV degradations fell outside of the acceptable range for 0.6   ±   0.3 mm leaf offsets, 0.11   ±   0.06 mm bank offsets, 0.6   ±   1.3 mm of random noise, and 1.0   ±   0.7° of gantry-MLC lag. Utilizing a training set comprised of prior treatment plans, machine learning is used to predict a range of achievable DVHs for the test patient’s anatomy. Consequently, degradations leading to statistical outliers may be identified

  8. Poster — Thur Eve — 32: Stereotactic Body Radiation Therapy for Peripheral Lung Lesion: Treatment Planning and Quality Assurance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wan, Shuying; Oliver, Michael; Wang, Xiaofang

    2014-08-15

    Stereotactic body radiation therapy (SBRT), due to its high precision for target localizing, has become widely used to treat tumours at various locations, including the lungs. Lung SBRT program was started at our institution a year ago. Eighteen patients with peripheral lesions up to 3 cm diameter have been treated with 48 Gy in 4 fractions. Based on four-dimensional computed tomography (4DCT) simulation, internal target volume (ITV) was delineated to encompass the respiratory motion of the lesion. A margin of 5 mm was then added to create the planning target volume (PTV) for setup uncertainties. There was no expansion frommore » gross tumour volume (GTV) to clinical target volume (CTV). Pinnacle 9.6 was used as the primary treatment planning system. Volumetric modulated arc therapy (VMAT) technique, with one or two coplanar arcs, generally worked well. For quality assurance (QA), each plan was exported to Eclipse 10 and dose calculation was repeated. Dose volume histograms (DVHs) of the targets and organs at risk (OARs) were then compared between the two treatment planning systems. Winston-Lutz tests were carried out as routine machine QA. Patient-specific QA included ArcCheck measurement with an insert, where an ionization chamber was placed at the centre to measure dose at the isocenter. For the first several patients, and subsequently for the plans with extremely strong modulation, Gafchromic film dosimetry was also employed. For each patient, a mock setup was scheduled prior to treatments. Daily pre- and post-CBCT were acquired for setup and assessment of intra-fractional motion, respectively.« less

  9. Robotic radiosurgery system patient-specific QA for extracranial treatments using the planar ion chamber array and the cylindrical diode array.

    PubMed

    Lin, Mu-Han; Veltchev, Iavor; Koren, Sion; Ma, Charlie; Li, Jinsgeng

    2015-07-08

    Robotic radiosurgery system has been increasingly employed for extracranial treatments. This work is aimed to study the feasibility of a cylindrical diode array and a planar ion chamber array for patient-specific QA with this robotic radiosurgery system and compare their performance. Fiducial markers were implanted in both systems to enable image-based setup. An in-house program was developed to postprocess the movie file of the measurements and apply the beam-by-beam angular corrections for both systems. The impact of noncoplanar delivery was then assessed by evaluating the angles created by the incident beams with respect to the two detector arrangements and cross-comparing the planned dose distribution to the measured ones with/without the angular corrections. The sensitivity of detecting the translational (1-3 mm) and the rotational (1°-3°) delivery errors were also evaluated for both systems. Six extracranial patient plans (PTV 7-137 cm³) were measured with these two systems and compared with the calculated doses. The plan dose distributions were calculated with ray-tracing and the Monte Carlo (MC) method, respectively. With 0.8 by 0.8 mm² diodes, the output factors measured with the cylindrical diode array agree better with the commissioning data. The maximum angular correction for a given beam is 8.2% for the planar ion chamber array and 2.4% for the cylindrical diode array. The two systems demonstrate a comparable sensitivity of detecting the translational targeting errors, while the cylindrical diode array is more sensitive to the rotational targeting error. The MC method is necessary for dose calculations in the cylindrical diode array phantom because the ray-tracing algorithm fails to handle the high-Z diodes and the acrylic phantom. For all the patient plans, the cylindrical diode array/ planar ion chamber array demonstrate 100% / > 92% (3%/3 mm) and > 96% / ~ 80% (2%/2 mm) passing rates. The feasibility of using both systems for robotic

  10. Enriching consumer health vocabulary through mining a social Q&A site: A similarity-based approach.

    PubMed

    He, Zhe; Chen, Zhiwei; Oh, Sanghee; Hou, Jinghui; Bian, Jiang

    2017-05-01

    The widely known vocabulary gap between health consumers and healthcare professionals hinders information seeking and health dialogue of consumers on end-user health applications. The Open Access and Collaborative Consumer Health Vocabulary (OAC CHV), which contains health-related terms used by lay consumers, has been created to bridge such a gap. Specifically, the OAC CHV facilitates consumers' health information retrieval by enabling consumer-facing health applications to translate between professional language and consumer friendly language. To keep up with the constantly evolving medical knowledge and language use, new terms need to be identified and added to the OAC CHV. User-generated content on social media, including social question and answer (social Q&A) sites, afford us an enormous opportunity in mining consumer health terms. Existing methods of identifying new consumer terms from text typically use ad-hoc lexical syntactic patterns and human review. Our study extends an existing method by extracting n-grams from a social Q&A textual corpus and representing them with a rich set of contextual and syntactic features. Using K-means clustering, our method, simiTerm, was able to identify terms that are both contextually and syntactically similar to the existing OAC CHV terms. We tested our method on social Q&A corpora on two disease domains: diabetes and cancer. Our method outperformed three baseline ranking methods. A post-hoc qualitative evaluation by human experts further validated that our method can effectively identify meaningful new consumer terms on social Q&A. Copyright © 2017 Elsevier Inc. All rights reserved.

  11. Portland cement concrete pavement review of QC/QA data 2000 through 2009.

    DOT National Transportation Integrated Search

    2011-04-01

    This report analyzes the Quality Control/Quality Assurance (QC/QA) data for Portland cement concrete pavement : (PCCP) awarded in the years 2000 through 2009. Analysis of the overall performance of the projects is accomplished by : reviewing the Calc...

  12. Zero Boil-Off System Testing

    NASA Technical Reports Server (NTRS)

    Plachta, David W.; Johnson, Wesley L.; Feller, Jeffrey R.

    2015-01-01

    Cryogenic propellants such as liquid hydrogen (LH2) and liquid oxygen (LO2) are a part of NASA's future space exploration due to their high specific impulse for rocket motors of upper stages suitable for transporting 10s to 100s of metric tons of payload mass to destinations outside of low earth orbit and for their return. However, the low storage temperatures of LH2 and LO2 cause substantial boil-off losses for missions with durations greater than several months. These losses can be eliminated by incorporating high performance cryocooler technology to intercept heat load to the propellant tanks and modulating the cryocooler to control tank pressure. The active thermal control technology being developed by NASA is the reverse turbo-Brayton cycle cryocooler and its integration to the propellant tank through a distributed cooling tubing network coupled to the tank wall. This configuration was recently tested at NASA Glenn Research Center, in a vacuum chamber and cryoshroud that simulated the essential thermal aspects of low Earth orbit, its vacuum and temperature. Testing consisted of three passive tests with the active cryocooler system off, and 7 active tests, with the cryocooler powered up. The test matrix included zero boil-off tests performed at 90 full and 25 full, and several demonstrations at excess cooling capacity and reduced cooling capacity. From this, the tank pressure response with varied cryocooler power inputs was determined. This test series established that the active cooling system integrated with the propellant tank eliminated boil-off and robustly controlled tank pressure.

  13. Pool boiling of water on nano-structured micro wires at sub-atmospheric conditions

    NASA Astrophysics Data System (ADS)

    Arya, Mahendra; Khandekar, Sameer; Pratap, Dheeraj; Ramakrishna, S. Anantha

    2016-09-01

    Past decades have seen active research in enhancement of boiling heat transfer by surface modifications. Favorable surface modifications are expected to enhance boiling efficiency. Several interrelated mechanisms such as capillarity, surface energy alteration, wettability, cavity geometry, wetting transitions, geometrical features of surface morphology, etc., are responsible for change in the boiling behavior of modified surfaces. Not much work is available on pool boiling at low pressures on microscale/nanoscale geometries; low pressure boiling is attractive in many applications wherein low operating temperatures are desired for a particular working fluid. In this background, an experimental setup was designed and developed to investigate the pool boiling performance of water on (a) plain aluminum micro wire (99.999 % pure) and, (b) nano-porous alumina structured aluminum micro wire, both having diameter of 250 µm, under sub-atmospheric pressure. Nano-structuring on the plain wire surface was achieved via anodization. Two samples, A and B of anodized wires, differing by the degree of anodization were tested. The heater length scale (wire diameter) was much smaller than the capillary length scale. Pool boiling characteristics of water were investigated at three different sub-atmospheric pressures of 73, 123 and 199 mbar (corresponding to T sat = 40, 50 and 60 °C). First, the boiling characteristics of plain wire were measured. It was noticed that at sub-atmospheric pressures, boiling heat transfer performance for plain wire was quite low due to the increased bubble sizes and low nucleation site density. Subsequently, boiling performance of nano-structured wires (both Sample A and Sample B) was compared with plain wire and it was noted that boiling heat transfer for the former was considerably enhanced as compared to the plain wire. This enhancement is attributed to increased nucleation site density, change in wettability and possibly due to enhanced pore scale

  14. Length Scale and Gravity Effects on Microgravity Boiling Heat Transfer

    NASA Technical Reports Server (NTRS)

    Kim, Jungho; McQuillen, John; Balombin, Joe

    2002-01-01

    Boiling is a complex phenomenon where hydrodynamics, heat transfer, mass transfer, and interfacial phenomena are tightly interwoven. An understanding of boiling and critical heat flux in microgravity environments is of importance to space based hardware and processes such as heat exchange, cryogenic fuel storage and transportation, electronic cooling, and material processing due to the large amounts of heat that can be removed with relatively little increase in temperature. Although research in this area has been performed in the past four decades, the mechanisms by which heat is removed from surfaces in microgravity are still unclear. In earth gravity, buoyancy is an important parameter that affects boiling heat transfer through the rate at which bubbles are removed from the surface. A simple model describing the bubble departure size based on a quasistatic force balance between buoyancy and surface tension is given by the Fritz [I] relation: Bo(exp 1/2) = 0.0208 theta where Bo is the ratio between buoyancy and surface tension forces. For small, rapidly growing bubbles, inertia associated with the induced liquid motion can also cause bubble departure. In microgravity, the magnitude of effects related to natural convection and buoyancy are small and physical mechanisms normally masked by natural convection in earth gravity such as Marangoni convection can substantially influence the boiling and vapor bubble dynamics. CHF (critical heat transfer) is also substantially affected by microgravity. In 1 g environments, Bo has been used as a correlating parameter for CHF. Zuber's CHF model for an infinite horizontal surface assumes that vapor columns formed by the merger of bubbles become unstable due to a Helmholtz instability blocking the supply of liquid to the surface. The jets are spaced lambda(sub D) apart, where lambda(sub D) = 2pi square root of 3[(sigma)/(g(rho(sub l) - rho(sub v)](exp 1/2) = 2pi square root of 3 L Bo(exp -1/2) = square root of 3 lambda(sub c

  15. QSPR using MOLGEN-QSPR: the challenge of fluoroalkane boiling points.

    PubMed

    Rücker, Christoph; Meringer, Markus; Kerber, Adalbert

    2005-01-01

    By means of the new software MOLGEN-QSPR, a multilinear regression model for the boiling points of lower fluoroalkanes is established. The model is based exclusively on simple descriptors derived directly from molecular structure and nevertheless describes a broader set of data more precisely than previous attempts that used either more demanding (quantum chemical) descriptors or more demanding (nonlinear) statistical methods such as neural networks. The model's internal consistency was confirmed by leave-one-out cross-validation. The model was used to predict all unknown boiling points of fluorobutanes, and the quality of predictions was estimated by means of comparison with boiling point predictions for fluoropentanes.

  16. Maintaining High Quality Data and Consistency Across a Diverse Flux Network: The Ameriflux QA/QC Technical Team

    NASA Astrophysics Data System (ADS)

    Chan, S.; Billesbach, D. P.; Hanson, C. V.; Biraud, S.

    2014-12-01

    The AmeriFlux quality assurance and quality control (QA/QC) technical team conducts short term (<2 weeks) intercomparisons using a portable eddy covariance system (PECS) to maintain high quality data observations and data consistency across the AmeriFlux network (http://ameriflux.lbl.gov/). Site intercomparisons identify discrepancies between the in situ and portable measurements and calculated fluxes. Findings are jointly discussed by the site staff and the QA/QC team to improve in the situ observations. Despite the relatively short duration of an individual site intercomparison, the accumulated record of all site visits (numbering over 100 since 2002) is a unique dataset. The ability to deploy redundant sensors provides a rare opportunity to identify, quantify, and understand uncertainties in eddy covariance and ancillary measurements. We present a few specific case studies from QA/QC site visits to highlight and share new and relevant findings related to eddy covariance instrumentation and operation.

  17. Teaching Structure-Property Relationships: Investigating Molecular Structure and Boiling Point

    ERIC Educational Resources Information Center

    Murphy, Peter M.

    2007-01-01

    A concise, well-organized table of the boiling points of 392 organic compounds has facilitated inquiry-based instruction in multiple scientific principles. Many individual or group learning activities can be derived from the tabulated data of molecular structure and boiling point based on the instructor's education objectives and the students'…

  18. From field notes to data portal - An operational QA/QC framework for tower networks

    NASA Astrophysics Data System (ADS)

    Sturtevant, C.; Hackley, S.; Meehan, T.; Roberti, J. A.; Holling, G.; Bonarrigo, S.

    2016-12-01

    Quality assurance and control (QA/QC) is one of the most important yet challenging aspects of producing research-quality data. This is especially so for environmental sensor networks collecting numerous high-frequency measurement streams at distributed sites. Here, the quality issues are multi-faceted, including sensor malfunctions, unmet theoretical assumptions, and measurement interference from the natural environment. To complicate matters, there are often multiple personnel managing different sites or different steps in the data flow. For large, centrally managed sensor networks such as NEON, the separation of field and processing duties is in the extreme. Tower networks such as Ameriflux, ICOS, and NEON continue to grow in size and sophistication, yet tools for robust, efficient, scalable QA/QC have lagged. Quality control remains a largely manual process relying on visual inspection of the data. In addition, notes of observed measurement interference or visible problems are often recorded on paper without an explicit pathway to data flagging during processing. As such, an increase in network size requires a near-proportional increase in personnel devoted to QA/QC, quickly stressing the human resources available. There is a need for a scalable, operational QA/QC framework that combines the efficiency and standardization of automated tests with the power and flexibility of visual checks, and includes an efficient communication pathway from field personnel to data processors to end users. Here we propose such a framework and an accompanying set of tools in development, including a mobile application template for recording tower maintenance and an R/shiny application for efficiently monitoring and synthesizing data quality issues. This framework seeks to incorporate lessons learned from the Ameriflux community and provide tools to aid continued network advancements.

  19. On mechanism of explosive boiling in nanosecond regime

    NASA Astrophysics Data System (ADS)

    Çelen, Serap

    2016-06-01

    Today laser-based machining is used to manufacture vital parts for biomedical, aviation and aerospace industries. The aim of the paper is to report theoretical, numerical and experimental investigations of explosive boiling under nanosecond pulsed ytterbium fiber laser irradiation. Experiments were performed in an effective peak power density range between 1397 and 1450 MW/cm2 on pure titanium specimens. The threshold laser fluence for phase explosion, the pressure and temperature at the target surface and the velocity of the expulsed material were reported. A narrow transition zone was realized between the normal vaporization and phase explosion fields. The proof of heterogeneous boiling was given with detailed micrographs. A novel thermal model was proposed for laser-induced splashing at high fluences. Packaging factor and scattering arc radius terms were proposed to state the level of the melt ejection process. Results of the present investigation explain the explosive boiling during high-power laser interaction with metal.

  20. An Experimental Study of Boiling in Reduced and Zero Gravity Fields

    NASA Technical Reports Server (NTRS)

    Usiskin, C. M.; Siegel, R.

    1961-01-01

    A pool boiling apparatus was mounted on a counterweighted platform which could be dropped a distance of nine feet. By varying the size of the counterweight, the effective gravity field on the equipment was adjusted between zero and unity. A study of boiling burnout in water indicated that a variation in the critical heat flux according to the one quarter power of gravity was reasonable. A consideration of the transient burnout process was necessary in order to properly interpret the data. A photographic study of nucleate boiling showed how the velocity of freely rising vapor bubbles decreased as gravity was reduced. The bubble diameters at the time of breakoff from the heated surface were found to vary inversely as gravity to the 1/3.5 power. Motion pictures were taken to illustrate both nucleate and film boiling in the low gravity range.

  1. A Study of Nucleate Boiling with Forced Convection in Microgravity

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.

    1999-01-01

    The ultimate objective of basic studies of flow boiling in microgravity is to improve the understanding of the processes involved, as manifested by the ability to predict its behavior. This is not yet the case for boiling heat transfer even in earth gravity, despite the considerable research activity over the past 30 years. The elements that constitute the nucleate boiling process - nucleation, growth, motion, and collapse of the vapor bubbles (if the bulk liquid is subcooled) - are common to both pool and flow boiling. It is well known that the imposition of bulk liquid motion affects the vapor bubble behavior relative to pool boiling, but does not appear to significantly influence the heat transfer. Indeed, it has been recommended in the past that empirical correlations or experimental data of pool boiling be used for design purposes with forced convection nucleate boiling. It is anticipated that such will most certainly not be possible for boiling in microgravity, based on observations made with pool boiling in microgravity. In earth gravity buoyancy will act to remove the vapor bubbles from the vicinity of the heater surface regardless of how much the imposed bulk velocity is reduced, depending, of course, on the geometry of the system. Vapor bubbles have been observed to dramatically increase in size in pool boiling in microgravity, and the heat flux at which dryout took place was reduced considerably below what is generally termed the critical heat flux (CHF) in earth gravity, depending on the bulk liquid subcooling. However, at heat flux levels below dryout, the nucleate pool boiling process was enhanced considerably over that in earth gravity, in spite of the large vapor bubbles formed in microgravity and perhaps as a consequence. These large vapor bubbles tended to remain in the vicinity of the heater surface, and the enhanced heat transfer appeared to be associated with the presence of what variously has been referred to as a liquid microlayer between the

  2. QA/QC requirements for physical properties sampling and analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Innis, B.E.

    1993-07-21

    This report presents results of an assessment of the available information concerning US Environmental Protection Agency (EPA) quality assurance/quality control (QA/QC) requirements and guidance applicable to sampling, handling, and analyzing physical parameter samples at Comprehensive Environmental Restoration, Compensation, and Liability Act (CERCLA) investigation sites. Geotechnical testing laboratories measure the following physical properties of soil and sediment samples collected during CERCLA remedial investigations (RI) at the Hanford Site: moisture content, grain size by sieve, grain size by hydrometer, specific gravity, bulk density/porosity, saturated hydraulic conductivity, moisture retention, unsaturated hydraulic conductivity, and permeability of rocks by flowing air. Geotechnical testing laboratories alsomore » measure the following chemical parameters of soil and sediment samples collected during Hanford Site CERCLA RI: calcium carbonate and saturated column leach testing. Physical parameter data are used for (1) characterization of vadose and saturated zone geology and hydrogeology, (2) selection of monitoring well screen sizes, (3) to support modeling and analysis of the vadose and saturated zones, and (4) for engineering design. The objectives of this report are to determine the QA/QC levels accepted in the EPA Region 10 for the sampling, handling, and analysis of soil samples for physical parameters during CERCLA RI.« less

  3. Fundamental Boiling and RP-1 Freezing Experiments

    NASA Technical Reports Server (NTRS)

    Goode, Brian; Turner, Larry D. (Technical Monitor)

    2001-01-01

    This paper describes results from experiments performed to help understand certain aspects of the MC-1 engine prestart thermal conditioning procedure. The procedure was constrained by the fact that the engine must chill long enough to get quality LOX at the LOX pump inlet but must be short enough to prevent freezing of RP-1 in the fuel pump. A chill test of an MC-1 LOX impeller was performed in LN2 to obtain data on film boiling, transition boiling and impeller temperature histories. The transition boiling data was important to the chill time so a subsequent experiment was performed chilling simple steel plates in LOX to obtain similar data for LOX. To address the fuel freezing concern, two experiments were performed. First, fuel was frozen in a tray and its physical characteristics were observed and temperatures of the fuel were measured. The result was physical characteristics as a function of temperature. Second was an attempt to measure the frozen thickness of RP-1 on a cold wall submerged in warm RP-1 and to develop a method for calculating that thickness for other conditions.

  4. A Novel Role of Three Dimensional Graphene Foam to Prevent Heater Failure during Boiling

    PubMed Central

    Ahn, Ho Seon; Kim, Ji Min; Park, Chibeom; Jang, Ji-Wook; Lee, Jae Sung; Kim, Hyungdae; Kaviany, Massoud; Kim, Moo Hwan

    2013-01-01

    We report a novel boiling heat transfer (NBHT) in reduced graphene oxide (RGO) suspended in water (RGO colloid) near critical heat flux (CHF), which is traditionally the dangerous limitation of nucleate boiling heat transfer because of heater failure. When the heat flux reaches the maximum value (CHF) in RGO colloid pool boiling, the wall temperature increases gradually and slowly with an almost constant heat flux, contrary to the rapid wall temperature increase found during water pool boiling. The gained time by NBHT would provide the safer margin of the heat transfer and the amazing impact on the thermal system as the first report of graphene application. In addition, the CHF and boiling heat transfer performance also increase. This novel boiling phenomenon can effectively prevent heater failure because of the role played by the self-assembled three-dimensional foam-like graphene network (SFG). PMID:23743619

  5. BOILING SLURRY REACTOR AND METHOD FO CONTROL

    DOEpatents

    Petrick, M.; Marchaterre, J.F.

    1963-05-01

    The control of a boiling slurry nuclear reactor is described. The reactor consists of a vertical tube having an enlarged portion, a steam drum at the top of the vertical tube, and at least one downcomer connecting the steam drum and the bottom of the vertical tube, the reactor being filled with a slurry of fissionabie material in water of such concentration that the enlarged portion of the vertical tube contains a critical mass. The slurry boils in the vertical tube and circulates upwardly therein and downwardly in the downcomer. To control the reactor by controlling the circulation of the slurry, a gas is introduced into the downcomer. (AEC)

  6. QA MANAGERS' RESPONSIBILITY TO ALLEGATIONS OF SCIENTIFIC MISCONDUCT. WHAT SHOULD THEY DO?

    EPA Science Inventory

    Although it is not the primary responsibility of QA Managers'/Officers' to deal with scientific misconduct, because of the nature of their daily activities they may encounter potential cases. It is important to maintain an ambience of openness and creativity as positive scientifi...

  7. Dysfunction and Decline: Lessons Learned from Inside Al-Qa’ida in Iraq

    DTIC Science & Technology

    2009-03-16

    facilitated AQI’s decline by killing and capturing  key   leadership, disrupting communications and logistics  processes , and giving  the  local  tribes a legitimate...dysfunctional politics will create space for al‐Qa`ida than that al‐Qa`ida will  unilaterally upset  the  political  process .  AQI’s setbacks have created a...jihadists that became AQI—offering little more than its reputation and  brand  name.   The  lessons of AQI’s failure are only partially applicable elsewhere

  8. Nucleate pool boiling in the long duration low gravity environment of the space shuttle

    NASA Technical Reports Server (NTRS)

    Hasan, M. M.; Lin, C. S.; Knoll, R. H.; Bentz, M. D.; Meserole, J. S.

    1993-01-01

    The results are presented of an experimental study of nucleate pool boiling performed in the low gravity environment of the space shuttle. Photographic observations of pool boiling in Freon 113 were obtained during the 'Tank Pressure Control Experiment', flown on the Space Transportation System STS-43 in August 1991. Nucleate boiling data from large (relative to bubble size) flat heating surfaces (0.1046 by 0.0742 m) was obtained at very low heat fluxes (0.22 to 1.19 kw/so m). The system pressure and the bulk liquid subcooling varied in the range of 40 to 60 kPa and 3 to 5 C respectively. Thirty-eight boiling tests, each of 10 min duration for a given heat flux, were conducted. Measurements included the heater power, heater surface temperature, the liquid temperature and the system pressure as functions of heating time. Video data of the first 2 min of heating was recorded for each test. In some tests the video clearly shows the inception of boiling and the growth and departure of bubbles from the surface during the first 2 min of heating. In the absence of video data, the heater temperature variation during heating shows the inception of boiling and stable nucleate boiling. During the stable nucleate boiling, the wall superheat varied between 2.8 to 3.8 C for heat fluxes in the range of 0.95 to 1.19 kw/so m. The wall superheat at the inception of boiling varied between 2 to 13 C.

  9. Nucleate pool boiling in the long duration low gravity environment of the Space Shuttle

    NASA Technical Reports Server (NTRS)

    Hasan, M. M.; Lin, C. S.; Knoll, R. H.; Bentz, M. D.; Meserole, J. S.

    1993-01-01

    The results are presented of an experimental study of nucleate pool boiling performed in the low gravity environment of the space shuttle. Photographic observations of pool boiling in Freon 113 were obtained during the 'Tank Pressure Control Experiment,' flown on the Space Transportation System, STS-43 in August 1991. Nucleate boiling data from large (relative to bubble size) flat heating surfaces (0.1046 by 0.0742 m) was obtained at very low heat fluxes (0.22 to 1.19 kW/sq m). The system pressure and the bulk liquid subcooling varied in the range of 40 to 60 kPa and 3 to 5 C respectively. Thirty-eight boiling tests, each of 10-min duration for a given heat flux, were conducted. Measurements included the heater power, heater surface temperature, the liquid temperature and the system pressure as functions of heating time. Video data of the first 2 min of heating was recorded for each test. In some tests the video clearly shows the inception of boiling and the growth and departure of bubbles from the surface during the first 2 min of heating. In the absence of video data, the heater temperature variation during heating shows the inception of boiling and stable nucleate boiling. During the stable nucleate boiling, the wall superheat varied between 2.8 to 3.8 C for heat fluxes in the range of 0.95 to 1.19 kW/sq m. The wall superheat at the inception of boiling varied between 2 to 13 C.

  10. Nucleate boiling performance on nano/microstructures with different wetting surfaces

    PubMed Central

    2012-01-01

    A study of nucleate boiling phenomena on nano/microstructures is a very basic and useful study with a view to the potential application of modified surfaces as heating surfaces in a number of fields. We present a detailed study of boiling experiments on fabricated nano/microstructured surfaces used as heating surfaces under atmospheric conditions, employing identical nanostructures with two different wettabilities (silicon-oxidized and Teflon-coated). Consequently, enhancements of both boiling heat transfer (BHT) and critical heat flux (CHF) are demonstrated in the nano/microstructures, independent of their wettability. However, the increment of BHT and CHF on each of the different wetting surfaces depended on the wetting characteristics of heating surfaces. The effect of water penetration in the surface structures by capillary phenomena is suggested as a plausible mechanism for the enhanced CHF on the nano/microstructures regardless of the wettability of the surfaces in atmospheric condition. This is supported by comparing bubble shapes generated in actual boiling experiments and dynamic contact angles under atmospheric conditions on Teflon-coated nano/microstructured surfaces. PMID:22559173

  11. Develop a Methodology to Evaluate the Effectiveness of QC/QA Specifications (Phase II)

    DOT National Transportation Integrated Search

    1998-08-01

    The Texas Department of Transportation (TxDOT) has been implementing statistically based quality control/quality assurance (QC/QA) specifications for hot mix asphalt concrete pavements since the early 1990s. These specifications have been continuousl...

  12. 3. RW Meyer Sugar Mill: 18761889. Sorghum pan and boiling ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    3. RW Meyer Sugar Mill: 1876-1889. Sorghum pan and boiling range flue. Manufactured by John Nott & Co., Honolulu, Hawaii, 1878. View: South side of sorghum pan and boiling range flue. In the sorghum pan heat was applied to the cane juice to clarify it, evaporate its water content, and concentrate the sugar crystals. Hot gasses moved through the flue underneath the entire copper bottom of the sorghum pan from the furnace (east) end to the smokestack (west) end of the boiling range. The sorghum pan sides are of redwood. The flue is built of fire-brick, masonry, and portland cement. - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  13. Phase relations and adiabats in boiling seafloor geothermal systems

    USGS Publications Warehouse

    Bischoff, J.L.; Pitzer, Kenneth S.

    1985-01-01

    Observations of large salinity variations and vent temperatures in the range of 380-400??C suggest that boiling or two-phase separation may be occurring in some seafloor geothermal systems. Consideration of flow rates and the relatively small differences in density between vapors and liquids at the supercritical pressures at depth in these systems suggests that boiling is occurring under closed-system conditions. Salinity and temperature of boiling vents can be used to estimate the pressure-temperature point in the subsurface at which liquid seawater first reached the two-phase boundary. Data are reviewed to construct phase diagrams of coexisting brines and vapors in the two-phase region at pressures corresponding to those of the seafloor geothermal systems. A method is developed for calculating the enthalpy and entropy of the coexisting mixtures, and results are used to construct adiabats from the seafloor to the P-T two-phase boundary. Results for seafloor vents discharging at 2300 m below sea level indicate that a 385??C vent is composed of a brine (7% NaCl equivalent) in equilibrium with a vapor (0.1% NaCl). Brine constitutes 45% by weight of the mixture, and the fluid first boiled at approximately 1 km below the seafloor at 415??C, 330 bar. A 400??C vent is primarily vapor (88 wt.%, 0.044% NaCl) with a small amount of brine (26% NaCl) and first boiled at 2.9 km below the seafloor at 500??C, 520 bar. These results show that adiabatic decompression in the two-phase region results in dramatic cooling of the fluid mixture when there is a large fraction of vapor. ?? 1985.

  14. From Field Notes to Data Portal - A Scalable Data QA/QC Framework for Tower Networks: Progress and Preliminary Results

    NASA Astrophysics Data System (ADS)

    Sturtevant, C.; Hackley, S.; Lee, R.; Holling, G.; Bonarrigo, S.

    2017-12-01

    Quality assurance and control (QA/QC) is one of the most important yet challenging aspects of producing research-quality data. Data quality issues are multi-faceted, including sensor malfunctions, unmet theoretical assumptions, and measurement interference from humans or the natural environment. Tower networks such as Ameriflux, ICOS, and NEON continue to grow in size and sophistication, yet tools for robust, efficient, scalable QA/QC have lagged. Quality control remains a largely manual process heavily relying on visual inspection of data. In addition, notes of measurement interference are often recorded on paper without an explicit pathway to data flagging. As such, an increase in network size requires a near-proportional increase in personnel devoted to QA/QC, quickly stressing the human resources available. We present a scalable QA/QC framework in development for NEON that combines the efficiency and standardization of automated checks with the power and flexibility of human review. This framework includes fast-response monitoring of sensor health, a mobile application for electronically recording maintenance activities, traditional point-based automated quality flagging, and continuous monitoring of quality outcomes and longer-term holistic evaluations. This framework maintains the traceability of quality information along the entirety of the data generation pipeline, and explicitly links field reports of measurement interference to quality flagging. Preliminary results show that data quality can be effectively monitored and managed for a multitude of sites with a small group of QA/QC staff. Several components of this framework are open-source, including a R-Shiny application for efficiently monitoring, synthesizing, and investigating data quality issues.

  15. The sudden coalescene model of the boiling crisis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Carrica, P.M.; Clausse, A.

    1995-09-01

    A local two-phase flow integral model of nucleate boiling and crisis is presented. The model is based on average balances on a control volume, yielding to a set of three nonlinear differential equations for the local void fraction, bubble number density and velocity. Boiling crisis as critical heat flux is interpreted as a dynamic transition caused by the coalescence of bubbles near the heater. The theoretical dynamic model is compared with experimental results obtained for linear power ramps in a horizontal plate heater in R-113, showing an excellent qualitative agreement.

  16. On Boiling of Crude Oil under Elevated Pressure

    NASA Astrophysics Data System (ADS)

    Pimenova, Anastasiya V.; Goldobin, Denis S.

    2016-02-01

    We construct a thermodynamic model for theoretical calculation of the boiling process of multicomponent mixtures of hydrocarbons (e.g., crude oil). The model governs kinetics of the mixture composition in the course of the distillation process along with the boiling temperature increase. The model heavily relies on the theory of dilute solutions of gases in liquids. Importantly, our results are applicable for modelling the process under elevated pressure (while the empiric models for oil cracking are not scalable to the case of extreme pressure), such as in an oil field heated by lava intrusions.

  17. Boiling points of halogenated ethanes: an explanatory model implicating weak intermolecular hydrogen-halogen bonding.

    PubMed

    Beauchamp, Guy

    2008-10-23

    This study explores via structural clues the influence of weak intermolecular hydrogen-halogen bonds on the boiling point of halogenated ethanes. The plot of boiling points of 86 halogenated ethanes versus the molar refraction (linked to polarizability) reveals a series of straight lines, each corresponding to one of nine possible arrangements of hydrogen and halogen atoms on the two-carbon skeleton. A multiple linear regression model of the boiling points could be designed based on molar refraction and subgroup structure as independent variables (R(2) = 0.995, standard error of boiling point 4.2 degrees C). The model is discussed in view of the fact that molar refraction can account for approximately 83.0% of the observed variation in boiling point, while 16.5% could be ascribed to weak C-X...H-C intermolecular interactions. The difference in the observed boiling point of molecules having similar molar refraction values but differing in hydrogen-halogen intermolecular bonds can reach as much as 90 degrees C.

  18. Nucleate pool boiling: High gravity to reduced gravity; liquid metals to cryogens

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.

    1988-01-01

    Requirements for the proper functioning of equipment and personnel in reduced gravity associated with space platforms and future space station modules introduce unique problems in temperature control; power generation; energy dissipation; the storage, transfer, control and conditioning of fluids; and liquid-vapor separation. The phase change of boiling is significant in all of these. Although both pool and flow boiling would be involved, research results to date include only pool boiling because buoyancy effects are maximized for this case. The effective application of forced convection boiling heat transfer in the microgravity of space will require a well grounded and cogent understanding of the mechanisms involved. Experimental results are presented for pool boiling from a single geometrical configuration, a flat surface, covering a wide range of body forces from a/g = 20 to 1 to a/g = 0 to -1 for a cryogenic liquid, and from a/g = 20 to 1 for water and a liquid metal. Similarities in behavior are noted for these three fluids at the higher gravity levels, and may reasonably be expected to continue at reduced gravity levels.

  19. Modeling acid-gas generation from boiling chloride brines

    PubMed Central

    2009-01-01

    Background This study investigates the generation of HCl and other acid gases from boiling calcium chloride dominated waters at atmospheric pressure, primarily using numerical modeling. The main focus of this investigation relates to the long-term geologic disposal of nuclear waste at Yucca Mountain, Nevada, where pore waters around waste-emplacement tunnels are expected to undergo boiling and evaporative concentration as a result of the heat released by spent nuclear fuel. Processes that are modeled include boiling of highly concentrated solutions, gas transport, and gas condensation accompanied by the dissociation of acid gases, causing low-pH condensate. Results Simple calculations are first carried out to evaluate condensate pH as a function of HCl gas fugacity and condensed water fraction for a vapor equilibrated with saturated calcium chloride brine at 50-150°C and 1 bar. The distillation of a calcium-chloride-dominated brine is then simulated with a reactive transport model using a brine composition representative of partially evaporated calcium-rich pore waters at Yucca Mountain. Results show a significant increase in boiling temperature from evaporative concentration, as well as low pH in condensates, particularly for dynamic systems where partial condensation takes place, which result in enrichment of HCl in condensates. These results are in qualitative agreement with experimental data from other studies. Conclusion The combination of reactive transport with multicomponent brine chemistry to study evaporation, boiling, and the potential for acid gas generation at the proposed Yucca Mountain repository is seen as an improvement relative to previously applied simpler batch evaporation models. This approach allows the evaluation of thermal, hydrological, and chemical (THC) processes in a coupled manner, and modeling of settings much more relevant to actual field conditions than the distillation experiment considered. The actual and modeled distillation

  20. Stability and potency of raw and boiled shrimp extracts for skin prick test.

    PubMed

    Pariyaprasert, Wipada; Piboonpocanun, Surapon; Jirapongsananuruk, Orathai; Visitsunthorn, Nualanong

    2015-06-01

    The difference of stability between raw and boiled shrimp extracts used in prick tests has never been investigated despite its potential consequences in tests development. The aim of this study was to compare the raw and boiled shrimp extracts of two species; Macrobrachium rosenbergii (freshwater shrimp) and Penaeus monodon (seawater shrimp) held at 4 ?C for different periods of time for their stability and potency in vivo by using the skin prick test (SPT) method. Raw and boiled M. rosenbergii and P. monodon extracts were prepared and stored at 4 ?C for 1, 7, 14 and 30 days. Thirty patients were pricked with raw and boiled shrimp extracts at all storage times, as well as prick to prick skin test (PTP) to fresh raw and boiled shrimps of both species. The mean wheal diameter (MWD) resulting from prick tests for all shrimp extracts was measured and compared. The shrimp extracts of all storage times yielded positive skin test results in the range of 90% - 100%. Raw P. monodon extracts induced larger wheals than boiled extracts at all storage times. There was no significant difference of MWD between raw and boiled M. rosenbergii extracts on day 1, 7, and 14. Significant correlations between MWD of PTP to fresh shrimps and SPT to all shrimp extracts were observed. All shrimp extracts were sterile at all storage times. Raw and boiled M. rosenbergii and P. monodon extracts were stable and sterile at 4 ?C for at most 30 days. SPT with these extracts induced more than 10 mm in shrimp allergy patients and the results were comparable with PTP to fresh shrimps.

  1. Identification of quantitative trait loci associated with boiled seed hardness in soybean

    PubMed Central

    Hirata, Kaori; Masuda, Ryoichi; Tsubokura, Yasutaka; Yasui, Takeshi; Yamada, Tetsuya; Takahashi, Koji; Nagaya, Taiko; Sayama, Takashi; Ishimoto, Masao; Hajika, Makita

    2014-01-01

    Boiled seed hardness is an important factor in the processing of soybean food products such as nimame and natto. Little information is available on the genetic basis for boiled seed hardness, despite the wide variation in this trait. DNA markers linked to the gene controlling this trait should be useful in soybean breeding programs because of the difficulty of its evaluation. In this report, quantitative trait locus (QTL) analysis was performed to reveal the genetic factors associated with boiled seed hardness using a recombinant inbred line population developed from a cross between two Japanese cultivars, ‘Natto-shoryu’ and ‘Hyoukei-kuro 3’, which differ largely in boiled seed hardness, which in ‘Natto-shoryu’ is about twice that of ‘Hyoukei-kuro 3’. Two significantly stable QTLs, qHbs3-1 and qHbs6-1, were identified on chromosomes 3 and 6, for which the ‘Hyoukei-kuro 3’ alleles contribute to decrease boiled seed hardness for both QTLs. qHbs3-1 also showed significant effects in progeny of a residual heterozygous line and in a different segregating population. Given its substantial effect on boiled seed hardness, SSR markers closely linked to qHbs3-1, such as BARCSOYSSR_03_0165 and BARCSOYSSR_03_0185, could be useful for marker-assisted selection in soybean breeding. PMID:25914591

  2. Flow Boiling and Condensation Experiment (FBCE) for the International Space Station

    NASA Technical Reports Server (NTRS)

    Mudawar, Issam; O'Neill, Lucas; Hasan, Mohammad; Nahra, Henry; Hall, Nancy; Balasubramaniam, R.; Mackey, Jeffrey

    2016-01-01

    An effective means to reducing the size and weight of future space vehicles is to replace present mostly single-phase thermal management systems with two-phase counterparts. By capitalizing upon both latent and sensible heat of the coolant rather than sensible heat alone, two-phase thermal management systems can yield orders of magnitude enhancement in flow boiling and condensation heat transfer coefficients. Because the understanding of the influence of microgravity on two-phase flow and heat transfer is quite limited, there is an urgent need for a new experimental microgravity facility to enable investigators to perform long-duration flow boiling and condensation experiments in pursuit of reliable databases, correlations and models. This presentation will discuss recent progress in the development of the Flow Boiling and Condensation Experiment (FBCE) for the International Space Station (ISS) in collaboration between Purdue University and NASA Glenn Research Center. Emphasis will be placed on the design of the flow boiling module and on new flow boiling data that were measured in parabolic flight, along with extensive flow visualization of interfacial features at heat fluxes up to critical heat flux (CHF). Also discussed a theoretical model that will be shown to predict CHF with high accuracy.

  3. Numerical Modeling of Propellant Boil-Off in a Cryogenic Storage Tank

    NASA Technical Reports Server (NTRS)

    Majumdar, A. K.; Steadman, T. E.; Maroney, J. L.; Sass, J. P.; Fesmire, J. E.

    2007-01-01

    A numerical model to predict boil-off of stored propellant in large spherical cryogenic tanks has been developed. Accurate prediction of tank boil-off rates for different thermal insulation systems was the goal of this collaboration effort. The Generalized Fluid System Simulation Program, integrating flow analysis and conjugate heat transfer for solving complex fluid system problems, was used to create the model. Calculation of tank boil-off rate requires simultaneous simulation of heat transfer processes among liquid propellant, vapor ullage space, and tank structure. The reference tank for the boil-off model was the 850,000 gallon liquid hydrogen tank at Launch Complex 39B (LC- 39B) at Kennedy Space Center, which is under study for future infrastructure improvements to support the Constellation program. The methodology employed in the numerical model was validated using a sub-scale model and tank. Experimental test data from a 1/15th scale version of the LC-39B tank using both liquid hydrogen and liquid nitrogen were used to anchor the analytical predictions of the sub-scale model. Favorable correlations between sub-scale model and experimental test data have provided confidence in full-scale tank boil-off predictions. These methods are now being used in the preliminary design for other cases including future launch vehicles

  4. Bubble dynamics, two-phase flow, and boiling heat transfer in a microgravity environment

    NASA Technical Reports Server (NTRS)

    Chung, Jacob N.

    1994-01-01

    The two-phase bubbly flow and boiling heat transfer in microgravity represents a substantial challenge to scientists and engineers and yet there is an urgent need to seek fundamental understanding in this area for future spacecraft design and space missions. At Washington State University, we have successfully designed, built and tested a 2.1 second drop tower with an innovation airbag deceleration system. Microgravity boiling experiments performed in our 0.6 second Drop Tower produced data flow visualizations that agree with published results and also provide some new understanding concerning flow boiling and microgravity bubble behavior. On the analytical and numerical work, the edge effects of finite divergent electrode plates on the forces experienced by bubbles were investigated. Boiling in a concentric cylinder microgravity and an electric field was numerically predicted. We also completed a feasibility study for microgravity boiling in an acoustic field.

  5. Boiling behavior of sodium-potassium alloy in a bench-scale solar receiver

    NASA Astrophysics Data System (ADS)

    Moreno, J. B.; Andraka, C. E.; Moss, T. A.

    During 1989-90, a 75-kW(sub t) sodium reflux pool-boiler solar receiver was successfully demonstrated at Sandia National Laboratories. Significant features of this receiver include the following: (1) boiling sodium as the heat transfer medium, and (2) electric-discharge-machined (EDM) cavities as artificial nucleation sites to stabilize boiling. Since this first demonstration, design of a second-generation pool-boiler receiver that will bring the concept closer to commercialization has begun. For long life, the new receiver uses Haynes Alloy 230. For increased safety factors against film boiling and flooding, it has a refined shape and somewhat larger dimensions. To eliminate the need for trace heating, the receiver will boil the sodium-potassium alloy NaK-78 instead of sodium. To reduce manufacturing costs, it will use one of a number of alternatives to EDM cavities for stabilization of boiling. To control incipient-boiling superheats, especially during hot restarts, it will contain a small amount of inert gas. Before the new receiver design could be finalized, bench-scale tests of some of the proposed changes were necessary. A series of bench-scale pool boilers were built from Haynes Alloy 230 and filled with NaK-78. Various boiling-stabilizer candidates were incorporated into them, including laser-drilled cavities and a number of different sintered-powder-metal coatings. These bench-scale pool boilers have been operated at temperatures up to 750 C, heated by quartz lamps with incident radiant fluxes up to 95 W/sq cm. The effects of various orientations and added gases have been studied. Results of these studies are presented.

  6. SU-E-T-354: Efficient and Enhanced QA Testing of Linear Accelerators Using a Real-Time Beam Monitor

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jung, J; Farrokhkish, M; Norrlinger, B

    2015-06-15

    Purpose: To investigate the feasibility of performing routine QA tests of linear accelerators (Linac) using the Integral Quality Monitoring (IQM) system. The system, consisting of a 1-D sensitivity gradient large area ion-chamber mounted at the collimator, allows automatic collection and analysis of beam data. Methods: The IQM was investigated to perform several QA constancy tests, similar to those recommended by AAPM TG142, of a Linac including: beam output, MLC calibration, beam symmetry, relative dose factor (RDF), dose linearity, output as a function of gantry angle and dose rate. All measurements by the IQM system accompanied a reference measurement using amore » conventional dosimetry system and were performed on an Elekta Infinity Linac with Agility MLC. The MLC calibration check is done using a Picket-Fence type 2×10cm{sup 2} field positioned at different off-axis locations along the chamber gradient. Beam symmetry constancy values are established by signals from an 4×4cm{sup 2} aperture located at various off-axis positions; the sensitivity of the test was determined by the changes in the signals in response to a tilt in the beam. The data for various square field sizes were used to develop a functional relationship with RDF. Results: The IQM tracked the beam output well within 1% of the reference ion-chamber readings. The Picket-Fence type field test detected a 1mm shift error of one MLC bank. The system was able to detect 2.5% or greater beam asymmetry. The IQM results for all other QA tests were found to agree with the reference values to within 0.5%. Conclusion: It was demonstrated that the IQM system can effectively monitor the Linac performance parameters for the purpose of routine QA constancy tests. With minimum user interactions a comprehensive set of tests can be performed efficiently, allowing frequent monitoring of the Linac. The presenting author’s salary is funded by the manufacturer of the QA device. All the other authors have

  7. California's Fiscal Crisis: What Does It Mean for Schools? Q&A

    ERIC Educational Resources Information Center

    EdSource, 2011

    2011-01-01

    This EdSource Q&A provides clear, accurate answers to some of the most common questions about the current state of education funding in California and what the state's fiscal crisis means for schools. School districts and other education stakeholders can share this brief with their communities and cite it in addressing questions about the…

  8. Characteristics of Pool Boiling on Graphite-Copper Composite Surfaces

    NASA Technical Reports Server (NTRS)

    Zhang, Nengli; Chao, David F.; Yang, Wen-Jei

    2002-01-01

    Nucleate pool boiling performance of different liquids on graphite-copper composite (Gr-Cu) surfaces has been experimentally studied and modeled. Both highly wetting fluids, such as freon-113 and pentane, and a moderately wetting fluid (water) were tested on the Gr-Cu surfaces with different graphite-fiber volume fractions to reveal the enhancement effects of the composite surfaces on the nucleate pool boiling. Results of the experiments show that the graphite-fiber volume fraction has an optimum value. The Gr-Cu composite surface with 25 percent graphite-fiber volume (f=0.25) has a maximum enhancement effect on the nucleate boiling heat transfer comparing to the pure copper surface. For the highly wetting fluid, the nucleate boiling heat transfer is generally enhanced on the Gr- Cu composite surfaces by 3 to 6 times shown. In the low heat flux region, the enhancement is over 6 times, but in the high heat flux region, the enhancement is reduced to about 40%. For the moderately wetting fluid (water), stronger enhancement of nucleate boiling heat transfer is achieved on the composite surface. It shown the experimental results in which one observes the nucleate boiling heat transfer enhancement of 5 to 10 times in the low heat flux region and an enhancement of 3 to 5 times in the high heat flux region. Photographs of bubble departure during the initial stage of nucleate boiling indicate that the bubbles detached from the composite surface are much smaller in diameter than those detached from the pure copper surface. Typical photographs are presented.It shows that the bubbles departed from the composite surface have diameters of only O(0.1) mm, while those departed from the pure copper surface have diameters of O(1) mm. It is also found that the bubbles depart from the composite surface at a much higher frequency, thus forming vapor columns. These two phenomena combined with high thermal conductivity of the graphite fiber are considered the mechanisms for such a

  9. Energy dependent calibration of XR-QA2 radiochromic film with monochromatic and polychromatic x-ray beams

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Di Lillo, F.; Mettivier, G., E-mail: mettivier@na.infn.it; Sarno, A.

    2016-01-15

    Purpose: This work investigates the energy response and dose-response curve determinations for XR-QA2 radiochromic film dosimetry system used for synchrotron radiation work and for quality assurance in diagnostic radiology, in the range of effective energies 18–46.5 keV. Methods: Pieces of XR-QA2 films were irradiated, in a plane transverse to the beam axis, with a monochromatic beam of energy in the range 18–40 keV at the ELETTRA synchrotron radiation facility (Trieste, Italy) and with a polychromatic beam from a laboratory x-ray tube operated at 80, 100, and 120 kV. The film calibration curve was expressed as air kerma (measured free-in-air withmore » an ionization chamber) versus the net optical reflectance change (netΔR) derived from the red channel of the RGB scanned film image. Four functional relationships (rational, linear exponential, power, and logarithm) were tested to evaluate the best curve for fitting the calibration data. The adequacy of the various fitting functions was tested by using the uncertainty analysis and by assessing the average of the absolute air kerma error calculated as the difference between calculated and delivered air kerma. The sensitivity of the film was evaluated as the ratio of the change in net reflectance to the corresponding air kerma. Results: The sensitivity of XR-QA2 films increased in the energy range 18–39 keV, with a maximum variation of about 170%, and decreased in the energy range 38–46.5 keV. The present results confirmed and extended previous findings by this and other groups, as regards the dose response of the radiochromic film XR-QA2 to monochromatic and polychromatic x-ray beams, respectively. Conclusions: The XR-QA2 radiochromic film response showed a strong dependence on beam energy for both monochromatic and polychromatic beams in the range of half value layer values from 0.55 to 6.1 mm Al and corresponding effective energies from 18 to 46.5 keV. In this range, the film response varied by 170

  10. Critical Heat Flux in Pool Boiling on Metal-Graphite Composite Surfaces

    NASA Technical Reports Server (NTRS)

    Zhang, Nengli; Yang, Wen-Jei; Chao, David F.; Chao, David F. (Technical Monitor)

    2000-01-01

    A study is conducted on high heat-flux pool boiling of pentane on micro-configured composite surfaces. The boiling surfaces are copper-graphite (Cu-Gr) and aluminum-graphite (Al-Gr) composites with a fiber volume concentration of 50%. The micro-graphite fibers embedded in the matrix contribute to a substantial enhancement in boiling heat-transfer performance. Correlation equations are obtained for both the isolated and coalesced bubble regimes, utilizing a mathematical model based on a metal-graphite, two-tier configuration with the aid of experimental data. A new model to predict the critical heat flux (CHF) on the composites is proposed to explain the fundamental aspects of the boiling phenomena. Three different factors affecting the CHF are considered in the model. Two of them are expected to become the main agents driving vapor volume detachment under microgravity conditions, using the metal-graphite composite surfaces as the heating surface and using liquids with an unusual Marangoni effect as the working fluid.

  11. Incorporating Water Boiling in the Numerical Modelling of Thermal Remediation by Electrical Resistance Heating

    NASA Astrophysics Data System (ADS)

    Molnar, I. L.; Krol, M.; Mumford, K. G.

    2017-12-01

    Developing numerical models for subsurface thermal remediation techniques - such as Electrical Resistive Heating (ERH) - that include multiphase processes such as in-situ water boiling, gas production and recovery has remained a significant challenge. These subsurface gas generation and recovery processes are driven by physical phenomena such as discrete and unstable gas (bubble) flow as well as water-gas phase mass transfer rates during bubble flow. Traditional approaches to multiphase flow modeling soil remain unable to accurately describe these phenomena. However, it has been demonstrated that Macroscopic Invasion Percolation (MIP) can successfully simulate discrete and unstable gas transport1. This has lead to the development of a coupled Electro Thermal-MIP Model2 (ET-MIP) capable of simulating multiple key processes in the thermal remediation and gas recovery process including: electrical heating of soil and groundwater, water flow, geological heterogeneity, heating-induced buoyant flow, water boiling, gas bubble generation and mobilization, contaminant mass transport and removal, and additional mechanisms such as bubble collapse in cooler regions. This study presents the first rigorous validation of a coupled ET-MIP model against two-dimensional water boiling and water/NAPL co-boiling experiments3. Once validated, the model was used to explore the impact of water and co-boiling events and subsequent gas generation and mobilization on ERH's ability to 1) generate, expand and mobilize gas at boiling and NAPL co-boiling temperatures, 2) efficiently strip contaminants from soil during both boiling and co-boiling. In addition, a quantification of the energy losses arising from steam generation during subsurface water boiling was examined with respect to its impact on the efficacy of thermal remediation. While this study specifically targets ERH, the study's focus on examining the fundamental mechanisms driving thermal remediation (e.g., water boiling) renders

  12. Radon depletion in xenon boil-off gas

    NASA Astrophysics Data System (ADS)

    Bruenner, S.; Cichon, D.; Lindemann, S.; Undagoitia, T. Marrodán; Simgen, H.

    2017-03-01

    An important background in detectors using liquid xenon for rare event searches arises from the decays of radon and its daughters. We report for the first time a reduction of ^{222}Rn in the gas phase above a liquid xenon reservoir. We show a reduction factor of ≳ 4 for the ^{222}Rn concentration in boil-off xenon gas compared to the radon enriched liquid phase. A semiconductor-based α -detector and miniaturized proportional counters are used to detect the radon. As the radon depletion in the boil-off gas is understood as a single-stage distillation process, this result establishes the suitability of cryogenic distillation to separate radon from xenon down to the 10^{-15} mol/mol level.

  13. 18. RW Meyer Sugar Mill: 18761889. Boiling House Interior, 1878. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    18. RW Meyer Sugar Mill: 1876-1889. Boiling House Interior, 1878. View: Detail of floor with molasses pits below floor level. The remaining floor boards indicate the structure of the floor covering the entire inside of the boiling house. In the left background the base of the centrifugals are in view. - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  14. Forced convection flow boiling and two-phase flow phenomena in a microchannel

    NASA Astrophysics Data System (ADS)

    Na, Yun Whan

    2008-07-01

    The present study was performed to numerically analyze the evaporation phenomena through the liquid-vapor interface and to investigate bubble dynamics and heat transfer behavior during forced convective flow boiling in a microchannel. Flow instabilities of two-phase flow boiling in a microchannel were studied as well. The main objective of this research is to investigate the fundamental mechanisms of two-phase flow boiling in a microchannel and provide predictive tools to design thermal management systems, for example, microchannel heat sinks. The numerical results obtained from this study were qualitatively and quantitatively compared with experimental results in the open literature. Physical and mathematical models, accounting for evaporating phenomena through the liquid-vapor interface in a microchannel at constant heat flux and constant wall temperature, have been developed, respectively. The heat transfer mechanism is affected by the dominant heat conduction through the thin liquid film and vaporization at the liquid-vapor interface. The thickness of the liquid film and the pressure of the liquid and vapor phases were simultaneously solved by the governing differential equations. The developed semi-analytical evaporation model that takes into account of the interfacial phenomena and surface tension effects was used to obtain solutions numerically using the fourth-order Runge-Kutta method. The effects of heat flux 19 and wall temperature on the liquid film were evaluated. The obtained pressure drops in a microchannel were qualitatively consistent with the experimental results of Qu and Mudawar (2004). Forced convective flow boiling in a single microchannel with different channel heights was studied through a numerical simulation to investigate bubble dynamics, flow patterns, and heat transfer. The momentum and energy equations were solved using the finite volume method while the liquid-vapor interface of a bubble is captured using the VOF (Volume of Fluid

  15. Bubble Dynamics, Two-Phase Flow, and Boiling Heat Transfer in Microgravity

    NASA Technical Reports Server (NTRS)

    Chung, Jacob N.

    1998-01-01

    This report contains two independent sections. Part one is titled "Terrestrial and Microgravity Pool Boiling Heat Transfer and Critical heat flux phenomenon in an acoustic standing wave." Terrestrial and microgravity pool boiling heat transfer experiments were performed in the presence of a standing acoustic wave from a platinum wire resistance heater using degassed FC-72 Fluorinert liquid. The sound wave was created by driving a half wavelength resonator at a frequency of 10.15 kHz. Microgravity conditions were created using the 2.1 second drop tower on the campus of Washington State University. Burnout of the heater wire, often encountered with heat flux controlled systems, was avoided by using a constant temperature controller to regulate the heater wire temperature. The amplitude of the acoustic standing wave was increased from 28 kPa to over 70 kPa and these pressure measurements were made using a hydrophone fabricated with a small piezoelectric ceramic. Cavitation incurred during experiments at higher acoustic amplitudes contributed to the vapor bubble dynamics and heat transfer. The heater wire was positioned at three different locations within the acoustic field: the acoustic node, antinode, and halfway between these locations. Complete boiling curves are presented to show how the applied acoustic field enhanced boiling heat transfer and increased critical heat flux in microgravity and terrestrial environments. Video images provide information on the interaction between the vapor bubbles and the acoustic field. Part two is titled, "Design and qualification of a microscale heater array for use in boiling heat transfer." This part is summarized herein. Boiling heat transfer is an efficient means of heat transfer because a large amount of heat can be removed from a surface using a relatively small temperature difference between the surface and the bulk liquid. However, the mechanisms that govern boiling heat transfer are not well understood. Measurements of

  16. SU-E-T-120: Analytic Dose Verification for Patient-Specific Proton Pencil Beam Scanning Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chang, C; Mah, D

    2015-06-15

    Purpose: To independently verify the QA dose of proton pencil beam scanning (PBS) plans using an analytic dose calculation model. Methods: An independent proton dose calculation engine is created using the same commissioning measurements as those employed to build our commercially available treatment planning system (TPS). Each proton PBS plan is exported from the TPS in DICOM format and calculated by this independent dose engine in a standard 40 x 40 x 40 cm water tank. This three-dimensional dose grid is then compared with the QA dose calculated by the commercial TPS, using standard Gamma criterion. A total of 18more » measured pristine Bragg peaks, ranging from 100 to 226 MeV, are used in the model. Intermediate proton energies are interpolated. Similarly, optical properties of the spots are measured in air over 15 cm upstream and downstream, and fitted to a second-order polynomial. Multiple Coulomb scattering in water is approximated analytically using Preston and Kohler formula for faster calculation. The effect of range shifters on spot size is modeled with generalized Highland formula. Note that the above formulation approximates multiple Coulomb scattering in water and we therefore chose not use the full Moliere/Hanson form. Results: Initial examination of 3 patient-specific prostate PBS plans shows that agreement exists between 3D dose distributions calculated by the TPS and the independent proton PBS dose calculation engine. Both calculated dose distributions are compared with actual measurements at three different depths per beam and good agreements are again observed. Conclusion: Results here showed that 3D dose distributions calculated by this independent proton PBS dose engine are in good agreement with both TPS calculations and actual measurements. This tool can potentially be used to reduce the amount of different measurement depths required for patient-specific proton PBS QA.« less

  17. Steady State Vapor Bubble in Pool Boiling

    PubMed Central

    Zou, An; Chanana, Ashish; Agrawal, Amit; Wayner, Peter C.; Maroo, Shalabh C.

    2016-01-01

    Boiling, a dynamic and multiscale process, has been studied for several decades; however, a comprehensive understanding of the process is still lacking. The bubble ebullition cycle, which occurs over millisecond time-span, makes it extremely challenging to study near-surface interfacial characteristics of a single bubble. Here, we create a steady-state vapor bubble that can remain stable for hours in a pool of sub-cooled water using a femtosecond laser source. The stability of the bubble allows us to measure the contact-angle and perform in-situ imaging of the contact-line region and the microlayer, on hydrophilic and hydrophobic surfaces and in both degassed and regular (with dissolved air) water. The early growth stage of vapor bubble in degassed water shows a completely wetted bubble base with the microlayer, and the bubble does not depart from the surface due to reduced liquid pressure in the microlayer. Using experimental data and numerical simulations, we obtain permissible range of maximum heat transfer coefficient possible in nucleate boiling and the width of the evaporating layer in the contact-line region. This technique of creating and measuring fundamental characteristics of a stable vapor bubble will facilitate rational design of nanostructures for boiling enhancement and advance thermal management in electronics. PMID:26837464

  18. QC/QA differences between hot mix asphalt (HMA) and warm mix asphalt (WMA).

    DOT National Transportation Integrated Search

    2013-01-01

    WMA represents a group of technologies which allow a reduction in temperatures at which asphalt mixtures are produced and placed on the road. ODOT Materials Division has conducted preliminary inquiries into QC/QA testing for WMA. Some respondents ind...

  19. TU-F-CAMPUS-J-05: Fast Volumetric MRI On An MRI-Linac Enables On-Line QA On Dose Deposition in the Patient

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Crijns, S; Glitzner, M; Kontaxis, C

    Purpose: The introduction of the MRI-linac in radiotherapy brings MRI-guided treatment with daily plan adaptions within reach. This paradigm demands on-line QA. With its ability to perform continuous volumetric imaging in an outstanding soft-tissue contrast, the MRI- linac promises to elucidate the dose deposition process during a treatment session. Here we study for a prostate case how dynamic MRI combined with linac machine parameters and a fast dose-engine can be used for on-line dose accumulation. Methods: Prostate imaging was performed in healthy volunteer on a 1.5T MR-scanner (Philips, Best, NL) according to a clinical MR-sim protocol, followed by 10min ofmore » dynamic imaging (FLASH, 4s/volume, FOV 40×40×12cm{sup 3}, voxels 3×3×3mm{sup 3}, TR/TE/α=3.5ms/1.7ms/5°). An experienced radiation oncologist made delineations, considering the prostate CTV. Planning was performed on a two-compartment pseudoCT (air/water density) according to clinical constraints (77Gy in PTV) using a Monte-Carlo (MC) based TPS that accounts for magnetic fields. Delivery of one fraction (2.2Gy) was simulated on an emulator for the Axesse linac (Elekta, Stockholm, SE). Machine parameters (MLC settings, gantry angle, dose rate, etc.) were recorded at 25Hz. These were re-grouped per dynamic volume and fed into the MC-engine to calculate a dose delivered for each of the dynamics. Deformations derived from non-rigid registration of each dynamic against the first allowed dose accumulation on a common reference grid. Results: The DVH parameters on the PTV compared to the optimized plan showed little changes. Local deformations however resulted in local deviations, primarily around the air/rectum interface. This clearly indicates the potential of intra-fraction adaptations based on the accumulated dose. Application in each fraction helps to track the influence of plan adaptations to the eventual dose distribution. Calculation times were about twice the delivery time. Conclusion: The

  20. Estimating surface temperature in forced convection nucleate boiling - A simplified method

    NASA Technical Reports Server (NTRS)

    Hendricks, R. C.; Papell, S. S.

    1977-01-01

    A simplified expression to estimate surface temperatures in forced convection boiling was developed using a liquid nitrogen data base. Using the principal of corresponding states and the Kutateladze relation for maximum pool boiling heat flux, the expression was normalized for use with other fluids. The expression was applied also to neon and water. For the neon data base, the agreement was acceptable with the exclusion of one set suspected to be in the transition boiling regime. For the water data base at reduced pressure greater than 0.05 the agreement is generally good. At lower reduced pressures, the water data scatter and the calculated temperature becomes a function of flow rate.

  1. SU-E-T-616: Plan Quality Assessment of Both Treatment Planning System Dose and Measurement-Based 3D Reconstructed Dose in the Patient

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Olch, A

    2015-06-15

    Purpose: Systematic radiotherapy plan quality assessment promotes quality improvement. Software tools can perform this analysis by applying site-specific structure dose metrics. The next step is to similarly evaluate the quality of the dose delivery. This study defines metrics for acceptable doses to targets and normal organs for a particular treatment site and scores each plan accordingly. The input can be the TPS or the measurement-based 3D patient dose. From this analysis, one can determine whether the delivered dose distribution to the patient receives a score which is comparable to the TPS plan score, otherwise replanning may be indicated. Methods: Elevenmore » neuroblastoma patient plans were exported from Eclipse to the Quality Reports program. A scoring algorithm defined a score for each normal and target structure based on dose-volume parameters. Each plan was scored by this algorithm and the percentage of total possible points was obtained. Each plan also underwent IMRT QA measurements with a Mapcheck2 or ArcCheck. These measurements were input into the 3DVH program to compute the patient 3D dose distribution which was analyzed using the same scoring algorithm as the TPS plan. Results: The mean quality score for the TPS plans was 75.37% (std dev=14.15%) compared to 71.95% (std dev=13.45%) for the 3DVH dose distribution. For 3/11 plans, the 3DVH-based quality score was higher than the TPS score, by between 0.5 to 8.4 percentage points. Eight/11 plans scores decreased based on IMRT QA measurements by 1.2 to 18.6 points. Conclusion: Software was used to determine the degree to which the plan quality score differed between the TPS and measurement-based dose. Although the delivery score was generally in good agreement with the planned dose score, there were some that improved while there was one plan whose delivered dose quality was significantly less than planned. This methodology helps evaluate both planned and delivered dose quality. Sun Nuclear

  2. Hot mix asphalt voids acceptance review of QC/QA data 2000 through 2004.

    DOT National Transportation Integrated Search

    2006-07-01

    This report analyzes the Quality Control/Quality Assurance (QC/QA) data for hot mix asphalt using voids acceptance as : the testing criteria for the years 2000 through 2004. Analysis of the overall quality of the HMA is accomplished by : reviewing th...

  3. Hot mix asphalt voids acceptance review of QC/QA data 2000 through 2010.

    DOT National Transportation Integrated Search

    2011-10-01

    This report analyzes the quality control/quality assurance (QC/QA) data for hot mix asphalt (HMA) using : voids acceptance as the testing criteria awarded in the years 2000 through 2010. Analysis of the overall : performance of the projects is accomp...

  4. SU-G-TeP4-02: A Method for Evaluating the Direct Impact of Failed IMRT QAs On Patient Dose

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Geneser, S; Butkus, M

    Purpose: We developed a method to calculate patient doses corresponding to IMRT QA measurements in order to determine and assess the actual dose delivered for plans with failed (or borderline) IMRT QA. This work demonstrates the feasibility of automatically computing delivered patient dose from portal dosimetry measurements in the Varian TPS system, which would provide a valuable and clinically viable IMRT QA tool for physicists and physicians. Methods: IMRT QA fluences were measured using portal dosimetry, processed using in-house matlab software, and imported back into Eclipse to calculate dose on the planning CT. To validate the proposed workflow, the Eclipsemore » calculated portal dose for a 5-field sliding window prostate boost plan was processed as described above. The resulting dose was compared to the planned dose and found to be within 0.5 Gy. Two IMRT QA results for the prostate boost plan (one that failed and one that passed) were processed and the resulting patient doses were evaluated. Results: The max dose difference between IMRT QA #1 and the original planned and approved dose is 4.5 Gy, while the difference between the planned and IMRT QA #2 dose is 4.0 Gy. The inferior portion of the PTV is slightly underdosed in both plans, and the superior portion is slightly overdosed. The patient dose resulting from IMRT QA #1 and #2 differs by only 0.5 Gy. With this new information, it may be argued that the evaluated plan alteration to obtain passing gamma analysis produced clinically irrelevant differences. Conclusion: Evaluation of the delivered QA dose on the planning CT provides valuable information about the clinical relevance of failed or borderline IMRT QAs. This particular workflow demonstrates the feasibility of pushing the measured IMRT QA portal dosimetry results directly back onto the patient planning CT within the Varian system.« less

  5. TU-G-BRD-06: The Imaging and Radiation Oncology Core Houston (IROC Houston) QA Center International Activities Outside North America

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Followill, D; Kry, S; Molineu, A

    Purpose: To describe the extent of IROC Houston’s (formerly the RPC) QA activities and audit results for radiotherapy institutions outside of North America (NA). Methods: The IROC Houston’s QA program components were designed to audit the radiation dose calculation chain from the NIST traceable reference beam calibration, to inclusion of dosimetry parameters used to calculate tumor doses, to the delivery of the radiation dose. The QA program provided to international institutions includes: 1) remote TLD/OSLD audit of machine output, 2) credentialing for advanced technologies, and 3) review of patient treatment records. IROC Houston uses the same standards and acceptance criteriamore » for all of its audits whether for North American or international sites. Results: IROC Houston’s QA program has reached out to radiotherapy sites in 43 different countries since 2013 through their participation in clinical trials. In the past two years, 2,778 international megavoltage beam outputs were audited with OSLD/TLD. While the average IROC/Inst ratio is near unity for all sites monitored, there are international regions whose results are significantly different from the NA region. In the past 2 years, 477 and 87 IMRT H&N phantoms were irradiated at NA and international sites, respectively. Regardless of the OSLD beam audit results, the overall pass rate (87 percent) for all international sites (no region separation) is equal to the NA sites. Of the 182 international patient charts reviewed, 10.7 percent of the dose calculation points did not meet our acceptance criterion as compared to 13.6 percent for NA sites. The lower pass rate for NA sites results from a much larger brachytherapy component which has been shown to be more error prone. Conclusion: IROC Houston has expanded its QA services worldwide and continues a long history of improving radiotherapy dose delivery in many countries. Funding received for QA audit services from the Korean GOG, DAHANCA, EORTC, ICON

  6. A scenario for a web-based radiation treatment planning structure: A new tool for quality assurance procedure?

    PubMed

    Kouloulias, V E; Ntasis, E; Poortmans, Ph; Maniatis, T A; Nikita, K S

    2003-01-01

    The desire to develop web-based platforms for remote collaboration among physicians and technologists is becoming a great challenge. In this paper we describe a web-based radiotherapy treatment planning (WBRTP) system to facilitate decentralized radiotherapy services by allowing remote treatment planning and quality assurance (QA) of treatment delivery. Significant prerequisites are digital storage of relevant data as well as efficient and reliable telecommunication system between collaborating units. The system of WBRTP includes video conferencing, display of medical images (CT scans, dose distributions etc), replication of selected data from a common database, remote treatment planning, evaluation of treatment technique and follow-up of the treated patients. Moreover the system features real-time remote operations in terms of tele-consulting like target volume delineation performed by a team of experts at different and distant units. An appraisal of its possibilities in quality assurance in radiotherapy is also discussed. As a conclusion, a WBRTP system would not only be a medium for communication between experts in oncology but mainly a tool for improving the QA in radiotherapy.

  7. Test/QA Plan for Verification of Leak Detection and Repair Technologies

    EPA Science Inventory

    The purpose of the leak detection and repair (LDAR) test and quality assurance plan is to specify procedures for a verification test applicable to commercial LDAR technologies. The purpose of the verification test is to evaluate the performance of participating technologies in b...

  8. Poster - 22: Retrospective analysis of portal dosimetry based QA of Prostate VMAT Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Badu, Shyam; Darko, Johnson; Fleck, Andre

    2016-08-15

    Purpose: The purpose of this study is to retrospectively analyze the portal dosimetry based quality assurance of prostate VMAT plans. Methods: Our standard quality assurance of VMAT treatment plans are performed using EPID installed on Varian TrueBeam Linac. In our current study we analyzed 84 prostate pretreatment VMAT plans. All plans consisted of two arcs, 7800cGy in 39 fractions with a 6MV beam. For each of these VMAT plans, the measured fluence for each arc is compared with the reference fluence using gamma index analysis. Results: We have compared the gamma passing rates for three criteria; 3%/3mm, 2%/2mm and 1%/1mm.more » Out of 168 arcs measured, the number below the gamma passing rate 95% using the area, Field+1cm, are 0, 2, and 124 for 3%/3mm, 2%/2mm and 1%/1mm criteria respectively. Corresponding numbers for MLC CIAO are 0, 2, and 139 respectively. The average gamma passing rate for all arcs measured using Field+1cm are 99.9±0.4, 99.6±1.2, and 90.9±6.5 for 3%/3mm, 2%/2mm and 1%/1mm respectively. Similarly if the MLC CIAO area is analyzed, a passing rate of 99.9±0.2, 99.2±1.2 and 87.2±8.5 respectively was observed. The average of the maximum gamma was also found to increase with tighter criteria. Conclusion: Analysis of prostate VMAT quality assurance plans indicate that the gamma passing rate is sensitive to the criteria and the area analyzed.« less

  9. SU-F-T-480: Evaluation of the Role of Varian Machine Performance Check (MPC) in Our Daily QA Routine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Juneja, B; Gao, S; Balter, P

    2016-06-15

    Purpose: (A) To assess the role of Varian MPC in our daily QA routine, and (B) evaluate the accuracy and precision of MPC. Methods: The MPC was performed weekly, for five months, on a Varian TrueBeam for five photon (6x, 10x, 15x, 6xFFF, and 10xFFF) and electron (6e, 9e, 12e, 16e, and 20e) energies. Output results were compared to those determined with an ionization chamber (TN30001, PTW-Freiburg) in plastic and a daily check device (DQA3, Sun Nuclear). Consistency of the Mechanical measurements over five months was analyzed and compared to monthly IsoCal results. Results: The MPC randomly showed large deviationsmore » (3–7%) that disappeared upon reacquisition. The MPC output closely matched monthly ion chamber and DQA3 measurements. The maximum and mean absolute difference between monthly and MPC was 1.18% and 0.28±0.21% for all energies. The maximum and mean absolute difference between DQA3 and MPC was 3.26% and 0.85±0.61%. The results suggest the MPC is comparable to the DQA3 for measuring output. The DQA3 provides wedge output, flatness, symmetry, and energy constancy checks, which are missing from the current implementation of the MPC. However, the MPC provides additional mechanical tests, such as size of the radiation isocenter (0.33±0.02 mm) and its coincidence with MV and kV isocenters (0.17±0.05 and 0.21±0.03 mm). It also provides positional accuracy of individual jaws (maximum σ, 0.33mm), all the MLC leaves (0.08mm), gantry (0.05°) and collimator (0.13°) rotation angles, and couch positioning (0.11mm) accuracy. MPC mechanical tests could replace our current daily on-board imaging QA routine and provide some additional QA not currently performed. Conclusion: MPC has the potential to be a valuable tool that facilitates reliable daily QA including many mechanical tests that are not currently performed. This system can add to our daily QA, but further development would be needed to fully replace our current Daily QA device.« less

  10. Recent Work on Flow Boiling and Condensation in a Single Microchannel

    NASA Astrophysics Data System (ADS)

    Quan, Xiaojun; Wang, Guodong; Cheng, Ping; Wu, Huiying

    2007-06-01

    Recent visualization and measurements results on flow boiling of water and condensation of steam in a single microchannel, carried out at Shanghai Jiaotong University, is summarized in this paper. For flow boiling of water, experiments were conducted in a single microchannel with a trapezoidal cross-section having a hydraulic diameter of 186 μm and a length of 30 mm. A boiling flow pattern map in terms of heat flux versus mass flux, showing the unstable and stable boiling flow regimes in the microchannel, is obtained. For the investigation of condensation, experiments were carried out for steam condensing inside a single microchannel with a length of 60mm having a hydraulic diameter of 87 μm and 120μm respectively. The location of transition from annular flow to plug/slug flow in a microchannel is found to be dependent on both the dimensionless condensation heat transfer rate as well as the Reynolds number of the steam. The frequency for the occurrence of the injection flow is found to increase with the increasing mass flux.

  11. Critical heat flux for free convection boiling in thin rectangular channels

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cheng, Lap Y.; Tichler, P.R.

    A review of the experimental data on free convection boiling critical heat flux (CHF) in vertical rectangular channels reveals three mechanisms of burnout. They are the pool boiling limit, the circulation limit, and the flooding limit associated with a transition in flow regime from churn to annular flow. The dominance of a particular mechanism depends on the dimensions of the channel. Analytical models were developed for each free convection boiling limit. Limited agreement with data is observed. A CHF correlation, which is valid for a wide range of gap sizes, was constructed from the CHFs calculated according to the threemore » mechanisms of burnout. 17 refs., 7 figs.« less

  12. SU-F-P-54: Guidelines to Check Image Registration QA of a Clinical Deformation Registration Software: A Single Institution Preliminary Study

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gill, G; Souri, S; Rea, A

    Purpose: The objective of this study is to verify and analyze the accuracy of a clinical deformable image registration (DIR) software. Methods: To test clinical DIR software qualitatively and quantitatively, we focused on lung radiotherapy and analyzed a single (Lung) patient CT scan. Artificial anatomical changes were applied to account for daily variations during the course of treatment including the planning target volume (PTV) and organs at risk (OAR). The primary CT (pCT) and the structure set (pST) was deformed with commercial tool (ImSimQA-Oncology Systems Limited) and after artificial deformation (dCT and dST) sent to another commercial tool (VelocityAI-Varian Medicalmore » Systems). In Velocity, the deformed CT and structures (dCT and dST) were inversely deformed back to original primary CT (dbpCT and dbpST). We compared the dbpST and pST structure sets using similarity metrics. Furthermore, a binary deformation field vector (BDF) was created and sent to ImSimQA software for comparison with known “ground truth” deformation vector fields (DVF). Results: An image similarity comparison was made by using “ground truth” DVF and “deformed output” BDF with an output of normalized “cross correlation (CC)” and “mutual information (MI)” in ImSimQA software. Results for the lung case were MI=0.66 and CC=0.99. The artificial structure deformation in both pST and dbpST was analyzed using DICE coefficient, mean distance to conformity (MDC) and deformation field error volume histogram (DFEVH) by comparing them before and after inverse deformation. We have noticed inadequate structure match for CTV, ITV and PTV due to close proximity of heart and overall affected by lung expansion. Conclusion: We have seen similarity between pCT and dbpCT but not so well between pST and dbpST, because of inadequate structure deformation in clinical DIR system. This system based quality assurance test will prepare us for adopting the guidelines of upcoming AAPM task

  13. Sensory quality and appropriateness of raw and boiled Jerusalem artichoke tubers (Helianthus tuberosus L.).

    PubMed

    Bach, Vibe; Kidmose, Ulla; Thybo, Anette K; Edelenbos, Merete

    2013-03-30

    The aim of the present study was to investigate the sensory attributes, dry matter and sugar content of five varieties of Jerusalem artichoke tubers and their relation to the appropriateness of the tubers for raw and boiled preparation. Sensory evaluation of raw and boiled Jerusalem artichoke tubers was performed by a trained sensory panel and a semi-trained consumer panel of 49 participants, who also evaluated the appropriateness of the tubers for raw and boiled preparation. The appropriateness of raw Jerusalem artichoke tubers was related to Jerusalem artichoke flavour, green nut flavour, sweetness and colour intensity, whereas the appropriateness of boiled tubers was related to celeriac aroma, sweet aroma, sweetness and colour intensity. In both preparations the variety Dwarf stood out from the others by being the least appropriate tuber. A few sensory attributes can be used as predictors of the appropriateness of Jerusalem artichoke tubers for raw and boiled consumption. Knowledge on the quality of raw and boiled Jerusalem artichoke tubers can be used to inform consumers on the right choice of raw material and thereby increase the consumption of the vegetable. © 2012 Society of Chemical Industry.

  14. Increasing Boiling Heat Transfer using Low Conductivity Materials

    PubMed Central

    Mahamudur Rahman, Md; Pollack, Jordan; McCarthy, Matthew

    2015-01-01

    We report the counterintuitive mechanism of increasing boiling heat transfer by incorporating low-conductivity materials at the interface between the surface and fluid. By embedding an array of non-conductive lines into a high-conductivity substrate, in-plane variations in the local surface temperature are created. During boiling the surface temperature varies spatially across the substrate, alternating between high and low values, and promotes the organization of distinct liquid and vapor flows. By systematically tuning the peak-to-peak wavelength of this spatial temperature variation, a resonance-like effect is seen at a value equal to the capillary length of the fluid. Replacing ~18% of the surface with a non-conductive epoxy results in a greater than 5x increase in heat transfer rate at a given superheat temperature. This drastic and counterintuitive increase is shown to be due to optimized bubble dynamics, where ordered pathways allow for efficient removal of vapor and the return of replenishing liquid. The use of engineered thermal gradients represents a potentially disruptive approach to create high-efficiency and high-heat-flux boiling surfaces which are naturally insensitive to fouling and degradation as compared to other approaches. PMID:26281890

  15. Evaluation of correlations of flow boiling heat transfer of R22 in horizontal channels.

    PubMed

    Zhou, Zhanru; Fang, Xiande; Li, Dingkun

    2013-01-01

    The calculation of two-phase flow boiling heat transfer of R22 in channels is required in a variety of applications, such as chemical process cooling systems, refrigeration, and air conditioning. A number of correlations for flow boiling heat transfer in channels have been proposed. This work evaluates the existing correlations for flow boiling heat transfer coefficient with 1669 experimental data points of flow boiling heat transfer of R22 collected from 18 published papers. The top two correlations for R22 are those of Liu and Winterton (1991) and Fang (2013), with the mean absolute deviation of 32.7% and 32.8%, respectively. More studies should be carried out to develop better ones. Effects of channel dimension and vapor quality on heat transfer are analyzed, and the results provide valuable information for further research in the correlation of two-phase flow boiling heat transfer of R22 in channels.

  16. A novel phantom and procedure providing submillimeter accuracy in daily QA tests of accelerators used for stereotactic radiosurgery*.

    PubMed

    Brezovich, Ivan A; Popple, Richard A; Duan, Jun; Shen, Sui; Wu, Xingen; Benhabib, Sidi; Huang, Mi; Cardan, Rex A

    2016-07-08

    Stereotactic radiosurgery (SRS) places great demands on spatial accuracy. Steel BBs used as markers in quality assurance (QA) phantoms are clearly visible in MV and planar kV images, but artifacts compromise cone-beam CT (CBCT) isocenter localization. The purpose of this work was to develop a QA phantom for measuring with sub-mm accuracy isocenter congruence of planar kV, MV, and CBCT imaging systems and to design a practical QA procedure that includes daily Winston-Lutz (WL) tests and does not require computer aid. The salient feature of the phantom (Universal Alignment Ball (UAB)) is a novel marker for precisely localizing isocenters of CBCT, planar kV, and MV beams. It consists of a 25.4mm diameter sphere of polymethylmetacrylate (PMMA) containing a concentric 6.35mm diameter tungsten carbide ball. The large density difference between PMMA and the polystyrene foam in which the PMMA sphere is embedded yields a sharp image of the sphere for accurate CBCT registration. The tungsten carbide ball serves in finding isocenter in planar kV and MV images and in doing WL tests. With the aid of the UAB, CBCT isocenter was located within 0.10 ± 0.05 mm of its true positon, and MV isocenter was pinpointed in planar images to within 0.06 ± 0.04mm. In clinical morning QA tests extending over an 18 months period the UAB consistently yielded measurements with sub-mm accuracy. The average distance between isocenter defined by orthogonal kV images and CBCT measured 0.16 ± 0.12 mm. In WL tests the central ray of anterior beams defined by a 1.5 × 1.5 cm2 MLC field agreed with CBCT isocenter within 0.03 ± 0.14 mm in the lateral direction and within 0.10 ± 0.19 mm in the longitudinal direction. Lateral MV beams approached CBCT isocenter within 0.00 ± 0.11 mm in the vertical direction and within -0.14 ± 0.15 mm longitudinally. It took therapists about 10 min to do the tests. The novel QA phantom allows pinpointing CBCT and MV isocenter positions to better than 0.2 mm, using

  17. Numerical Investigation of Microgravity Tank Pressure Rise Due to Boiling

    NASA Technical Reports Server (NTRS)

    Hylton, Sonya; Ibrahim, Mounir; Kartuzova, Olga; Kassemi, Mohammad

    2015-01-01

    The ability to control self-pressurization in cryogenic storage tanks is essential for NASAs long-term space exploration missions. Predictions of the tank pressure rise in Space are needed in order to inform the microgravity design and optimization process. Due to the fact that natural convection is very weak in microgravity, heat leaks into the tank can create superheated regions in the liquid. The superheated regions can instigate microgravity boiling, giving rise to pressure spikes during self-pressurization. In this work, a CFD model is developed to predict the magnitude and duration of the microgravity pressure spikes. The model uses the Schrage equation to calculate the mass transfer, with a different accommodation coefficient for evaporation at the interface, condensation at the interface, and boiling in the bulk liquid. The implicit VOF model was used to account for the moving interface, with bounded second order time discretization. Validation of the models predictions was carried out using microgravity data from the Tank Pressure Control Experiment, which flew aboard the Space Shuttle Mission STS-52. Although this experiment was meant to study pressurization and pressure control, it underwent boiling during several tests. The pressure rise predicted by the CFD model compared well with the experimental data. The ZBOT microgravity experiment is scheduled to fly on February 2016 aboard the ISS. The CFD model was also used to perform simulations for setting parametric limits for the Zero-Boil-Off Tank (ZBOT) Experiments Test Matrix in an attempt to avoid boiling in the majority of the test runs that are aimed to study pressure increase rates during self-pressurization. *Supported in part by NASA ISS Physical Sciences Research Program, NASA HQ, USA

  18. Flow-Boiling Critical Heat Flux Experiments Performed in Reduced Gravity

    NASA Technical Reports Server (NTRS)

    Hasan, Mohammad M.; Mudawar, Issam

    2005-01-01

    Poor understanding of flow boiling in microgravity has recently emerged as a key obstacle to the development of many types of power generation and advanced life support systems intended for space exploration. The critical heat flux (CHF) is perhaps the most important thermal design parameter for boiling systems involving both heatflux-controlled devices and intense heat removal. Exceeding the CHF limit can lead to permanent damage, including physical burnout of the heat-dissipating device. The importance of the CHF limit creates an urgent need to develop predictive design tools to ensure both the safe and reliable operation of a two-phase thermal management system under the reduced-gravity (like that on the Moon and Mars) and microgravity environments of space. At present, very limited information is available on flow-boiling heat transfer and the CHF under these conditions.

  19. Monte Carlo-based QA for IMRT of head and neck cancers

    NASA Astrophysics Data System (ADS)

    Tang, F.; Sham, J.; Ma, C.-M.; Li, J.-S.

    2007-06-01

    It is well-known that the presence of large air cavity in a dense medium (or patient) introduces significant electronic disequilibrium when irradiated with megavoltage X-ray field. This condition may worsen by the possible use of tiny beamlets in intensity-modulated radiation therapy (IMRT). Commercial treatment planning systems (TPSs), in particular those based on the pencil-beam method, do not provide accurate dose computation for the lungs and other cavity-laden body sites such as the head and neck. In this paper we present the use of Monte Carlo (MC) technique for dose re-calculation of IMRT of head and neck cancers. In our clinic, a turn-key software system is set up for MC calculation and comparison with TPS-calculated treatment plans as part of the quality assurance (QA) programme for IMRT delivery. A set of 10 off-the-self PCs is employed as the MC calculation engine with treatment plan parameters imported from the TPS via a graphical user interface (GUI) which also provides a platform for launching remote MC simulation and subsequent dose comparison with the TPS. The TPS-segmented intensity maps are used as input for the simulation hence skipping the time-consuming simulation of the multi-leaf collimator (MLC). The primary objective of this approach is to assess the accuracy of the TPS calculations in the presence of air cavities in the head and neck whereas the accuracy of leaf segmentation is verified by fluence measurement using a fluoroscopic camera-based imaging device. This measurement can also validate the correct transfer of intensity maps to the record and verify system. Comparisons between TPS and MC calculations of 6 MV IMRT for typical head and neck treatments review regional consistency in dose distribution except at and around the sinuses where our pencil-beam-based TPS sometimes over-predicts the dose by up to 10%, depending on the size of the cavities. In addition, dose re-buildup of up to 4% is observed at the posterior nasopharyngeal

  20. Science 101: Why Does It Take Longer to Boil Potatoes at High Altitudes?

    ERIC Educational Resources Information Center

    Robertson, Bill

    2017-01-01

    Why Does It Take Longer to Boil Potatoes at High Altitudes? This column provides background science information for elementary teachers. This month's issue looks at why water boils at different temperatures at different altitudes.

  1. 16. RW Meyer Sugar Mill: 18761889. Boiling House Interior, 1878. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    16. RW Meyer Sugar Mill: 1876-1889. Boiling House Interior, 1878. View: Looking from west to east through boiling house. The sorghum pan is on the right. The beams; joists, and trusses are of northwest pine; side boards are of redwood. A foundation line of a loading dock and smokestack are in the foreground. Both end walls have deteriorated completely. - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  2. Effect of soaking, boiling, and steaming on total phenolic contentand antioxidant activities of cool season food legumes.

    PubMed

    Xu, Baojun; Chang, Sam K C

    2008-09-01

    The effects of soaking, boiling and steaming processes on the total phenolic components and antioxidant activity in commonly consumed cool season food legumes (CSFL's), including green pea, yellow pea, chickpea and lentil were investigated. As compared to original unprocessed legumes, all processing steps caused significant (p<0.05) decreases in total phenolic content (TPC), DPPH free radical scavenging activity (DPPH) in all tested CSFL's. All soaking and atmospheric boiling treatments caused significant (p<0.05) decreases in oxygen radical absorbing capacity (ORAC). However, pressure boiling and pressure steaming caused significant (p<0.05) increases in ORAC values. Steaming treatments resulted in a greater retention of TPC, DPPH, and ORAC values in all tested CSFL's as compared to boiling treatments. To obtain cooked legumes with similar palatability and firmness, pressure boiling shortened processing time as compared to atmospheric boiling, resulted in insignificant differences in TPC, DPPH for green and yellow pea. However, TPC and DPPH in cooked lentils differed significantly between atmospheric and pressure boiling. As compared to atmospheric processes, pressure processes significantly increased ORAC values in both boiled and steamed CSFL's. Greater TPC, DPPH and ORAC values were detected in boiling water than that in soaking and steaming water. Boiling also caused more solid loss than steaming. Steam processing exhibited several advantages in retaining the integrity of the legume appearance and texture of the cooked product, shortening process time, and greater retention of antioxidant components and activities. Copyright © 2008 Elsevier Ltd. All rights reserved.

  3. Spinal pedicle screw planning using deformable atlas registration

    NASA Astrophysics Data System (ADS)

    Goerres, J.; Uneri, A.; De Silva, T.; Ketcha, M.; Reaungamornrat, S.; Jacobson, M.; Vogt, S.; Kleinszig, G.; Osgood, G.; Wolinsky, J.-P.; Siewerdsen, J. H.

    2017-04-01

    Spinal screw placement is a challenging task due to small bone corridors and high risk of neurological or vascular complications, benefiting from precision guidance/navigation and quality assurance (QA). Implicit to both guidance and QA is the definition of a surgical plan—i.e. the desired trajectories and device selection for target vertebrae—conventionally requiring time-consuming manual annotations by a skilled surgeon. We propose automation of such planning by deriving the pedicle trajectory and device selection from a patient’s preoperative CT or MRI. An atlas of vertebrae surfaces was created to provide the underlying basis for automatic planning—in this work, comprising 40 exemplary vertebrae at three levels of the spine (T7, T8, and L3). The atlas was enriched with ideal trajectory annotations for 60 pedicles in total. To define trajectories for a given patient, sparse deformation fields from the atlas surfaces to the input (CT or MR image) are applied on the annotated trajectories. Mean value coordinates are used to interpolate dense deformation fields. The pose of a straight trajectory is optimized by image-based registration to an accumulated volume of the deformed annotations. For evaluation, input deformation fields were created using coherent point drift (CPD) to perform a leave-one-out analysis over the atlas surfaces. CPD registration demonstrated surface error of 0.89  ±  0.10 mm (median  ±  interquartile range) for T7/T8 and 1.29  ±  0.15 mm for L3. At the pedicle center, registered trajectories deviated from the expert reference by 0.56  ±  0.63 mm (T7/T8) and 1.12  ±  0.67 mm (L3). The predicted maximum screw diameter differed by 0.45  ±  0.62 mm (T7/T8), and 1.26  ±  1.19 mm (L3). The automated planning method avoided screw collisions in all cases and demonstrated close agreement overall with expert reference plans, offering a potentially valuable tool in support

  4. Enhanced Droplet Control by Transition Boiling

    PubMed Central

    Grounds, Alex; Still, Richard; Takashina, Kei

    2012-01-01

    A droplet of water on a heated surface can levitate over a film of gas produced by its own evaporation in the Leidenfrost effect. When the surface is prepared with ratchet-like saw-teeth topography, these droplets can self-propel and can even climb uphill. However, the extent to which the droplets can be controlled is limited by the physics of the Leidenfrost effect. Here, we show that transition boiling can be induced even at very high surface temperatures and provide additional control over the droplets. Ratchets with acute protrusions enable droplets to climb steeper inclines while ratchets with sub-structures enable their direction of motion to be controlled by varying the temperature of the surface. The droplets' departure from the Leidenfrost regime is assessed by analysing the sound produced by their boiling. We anticipate these techniques will enable the development of more sophisticated methods for controlling small droplets and heat transfer. PMID:23056912

  5. Enhanced Droplet Control by Transition Boiling

    NASA Astrophysics Data System (ADS)

    Grounds, Alex; Still, Richard; Takashina, Kei

    2012-10-01

    A droplet of water on a heated surface can levitate over a film of gas produced by its own evaporation in the Leidenfrost effect. When the surface is prepared with ratchet-like saw-teeth topography, these droplets can self-propel and can even climb uphill. However, the extent to which the droplets can be controlled is limited by the physics of the Leidenfrost effect. Here, we show that transition boiling can be induced even at very high surface temperatures and provide additional control over the droplets. Ratchets with acute protrusions enable droplets to climb steeper inclines while ratchets with sub-structures enable their direction of motion to be controlled by varying the temperature of the surface. The droplets' departure from the Leidenfrost regime is assessed by analysing the sound produced by their boiling. We anticipate these techniques will enable the development of more sophisticated methods for controlling small droplets and heat transfer.

  6. Evaluation of Correlations of Flow Boiling Heat Transfer of R22 in Horizontal Channels

    PubMed Central

    Fang, Xiande; Li, Dingkun

    2013-01-01

    The calculation of two-phase flow boiling heat transfer of R22 in channels is required in a variety of applications, such as chemical process cooling systems, refrigeration, and air conditioning. A number of correlations for flow boiling heat transfer in channels have been proposed. This work evaluates the existing correlations for flow boiling heat transfer coefficient with 1669 experimental data points of flow boiling heat transfer of R22 collected from 18 published papers. The top two correlations for R22 are those of Liu and Winterton (1991) and Fang (2013), with the mean absolute deviation of 32.7% and 32.8%, respectively. More studies should be carried out to develop better ones. Effects of channel dimension and vapor quality on heat transfer are analyzed, and the results provide valuable information for further research in the correlation of two-phase flow boiling heat transfer of R22 in channels. PMID:23956695

  7. 300 Area TEDF NPDES Permit Compliance Monitoring Plan

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Loll, C.M.

    1994-10-13

    This monitoring plan describes the activities and methods that will be employed at the 300 Area Treated Effluent Disposal Facility (TEDF) in order to ensure compliance with the National Discharge Elimination System (NPDES) permit. Included in this document are a brief description of the project, the specifics of the sampling effort, including the physical location and frequency of sampling, the support required for sampling, and the Quality Assurance (QA) protocols to be followed in the sampling procedures.

  8. Quality assurance methodology for Varian RapidArc treatment plans

    PubMed Central

    Cirino, Eileen T.; Xiong, Li; Mower, Herbert W.

    2010-01-01

    With the commercial introduction of the Varian RapidArc, a new modality for treatment planning and delivery, the need has arisen for consistent and efficient techniques for performing patient‐specific quality assurance (QA) tests. In this paper we present our methodology for a RapidArc treatment plan QA procedure. For our measurements we used a 2D diode array (MapCHECK) embedded at 5 cm water equivalent depth in MapPHAN 5 phantom and an Exradin A16 ion chamber placed in six different positions in a cylindrical homogeneous phantom (QUASAR). We also checked the MUs for the RapidArc plans by using independent software (RadCalc). The agreement between Eclipse calculations and MapCHECK/MapPHAN 5 measurements was evaluated using both absolute distance‐to‐agreement (DTA) and gamma index with 10% dose threshold (TH), 3% dose difference (DD), and 3 mm DTA. The average agreement was 94.4% for the DTA approach and 96.3% for the gamma index approach. In high‐dose areas, the discrepancy between calculations and ion chamber measurements using the QUASAR phantom was within 4.5% for prostate cases. For the RadCalc calculations, we used the average SSD along the arc; however, for some patients the agreement for the MUs obtained with RadCalc versus Eclipse was inadequate (discrepancy>5%). In these cases, the plan was divided into partial arc plans so that RadCalc could perform a better estimation of the MUs. The discrepancy was further reduced to within ~4% using this approach. Regardless of the variation in prescribed dose and location of the treated areas, we obtained very good results for all patients studied in this paper. PACS number: 87.55.Qr

  9. Bench-scale screening tests for a boiling sodium-potassium alloy solar receiver

    NASA Astrophysics Data System (ADS)

    Moreno, J. B.; Moss, T. A.

    1993-06-01

    Bench-scale tests were carried out in support of the design of a second-generation 75-kW(sub t) reflux pool-boiler solar receiver. The receiver will be made from Haynes Alloy 230 and will contain the sodium-potassium alloy NaK-78. The bench-scale tests used quartz lamp heated boilers to screen candidate boiling stabilization materials and methods at temperatures up to 750 degree C. Candidates that provided stable boiling were tested for hot-restart behavior. Poor stability was obtained with single 1/4-inch diameter patches of powdered metal hot press sintered onto the wetted side of the heat-input area. Laser-drilled and electric discharge machined cavities in the heated surface also performed poorly. Small additions of xenon, and heated-surface tilt out of the vertical, dramatically improved poor boiling stability; additions of helium or oxygen did not. The most stable boiling was obtained when the entire heat-input area was covered by a powdered-metal coating. The effect of heated-area size was assessed for one coating: at low incident fluxes, when even this coating performed poorly, increasing the heated-area size markedly improved boiling stability. Good hot-restart behavior was not observed with any candidate, although results were significantly better with added xenon in a boiler shortened from 3 to 2 feet. In addition to the screening tests, flash-radiography imaging of metal-vapor bubbles during boiling was attempted. Contrary to the Cole-Rohsenow correlation, these bubble-size estimates did not vary with pressure; instead they were constant, consistent with the only other alkali metal measurements, but about 1/2 their size.

  10. Controlled tissue emulsification produced by high intensity focused ultrasound shock waves and millisecond boiling

    PubMed Central

    Khokhlova, Tatiana D.; Canney, Michael S.; Khokhlova, Vera A.; Sapozhnikov, Oleg A.; Crum, Lawrence A.; Bailey, Michael R.

    2011-01-01

    In high intensity focused ultrasound (HIFU) applications, tissue may be thermally necrosed by heating, emulsified by cavitation, or, as was recently discovered, emulsified using repetitive millisecond boiling caused by shock wave heating. Here, this last approach was further investigated. Experiments were performed in transparent gels and ex vivo bovine heart tissue using 1, 2, and 3 MHz focused transducers and different pulsing schemes in which the pressure, duty factor, and pulse duration were varied. A previously developed derating procedure to determine in situ shock amplitudes and the time-to-boil was refined. Treatments were monitored using B-mode ultrasound. Both inertial cavitation and boiling were observed during exposures, but emulsification occurred only when shocks and boiling were present. Emulsified lesions without thermal denaturation were produced with shock amplitudes sufficient to induce boiling in less than 20 ms, duty factors of less than 0.02, and pulse lengths shorter than 30 ms. Higher duty factors or longer pulses produced varying degrees of thermal denaturation combined with mechanical emulsification. Larger lesions were obtained using lower ultrasound frequencies. The results show that shock wave heating and millisecond boiling is an effective and reliable way to emulsify tissue while monitoring the treatment with ultrasound. PMID:22088025

  11. Transient boiling in two-phase helium natural circulation loops

    NASA Astrophysics Data System (ADS)

    Furci, H.; Baudouy, B.; Four, A.; Meuris, C.

    2014-01-01

    Two-phase helium natural circulation loops are used for cooling large superconducting magnets, as CMS for LHC. During normal operation or in the case of incidents, transients are exerted on the cooling system. Here a cooling system of this type is studied experimentally. Sudden power changes are operated on a vertical-heated-section natural convection loop, simulating a fast increase of heat deposition on magnet cooling pipes. Mass flow rate, heated section wall temperature and pressure drop variations are measured as a function of time, to assess the time behavior concerning the boiling regime according to the values of power injected on the heated section. The boiling curves and critical heat flux (CHF) values have been obtained in steady state. Temperature evolution has been observed in order to explore the operating ranges where heat transfer is deteriorated. Premature film boiling has been observed during transients on the heated section in some power ranges, even at appreciably lower values than the CHF. A way of attenuating these undesired temperature excursions has been identified through the application of high enough initial heating power.

  12. Boiling of an emulsion in a yield stress fluid.

    PubMed

    Guéna, Geoffroy; Wang, Ji; d'Espinose, Jean-Baptiste; Lequeux, François; Talini, Laurence

    2010-11-01

    We report the boiling behavior of pentane emulsified in a yield stress fluid, a colloidal clay (Laponite) suspension. We have observed that a superheated state is easily reached: the emulsion, heated more than 50 °C above the alkane boiling point, does not boil. Superheating is made possible by the suppression of heterogeneous nucleation in pentane, resulting from the emulsification process, a phenomenon evidenced decades ago in studies of the superheating of two phase fluids. We have furthermore studied the growth of isolated bubbles nucleated in the emulsion. The rate of increase of the bubble radius with time depends on both the temperature and emulsion volume fraction but, rather unexpectedly, does not depend on the fluid rheology. We show that the bubbles grow by diffusion of the alkane through the aqueous phase between liquid droplets and bubbles, analogously to an Ostwald ripening process. The peculiarity of the process reported here is that a layer depleted in oil droplets forms around the bubble, layer to which the alkane concentration gradient is confined. We successfully describe our experimental results with a simple transfer model.

  13. Stability of film boiling on inclined plates and spheres

    NASA Astrophysics Data System (ADS)

    Aursand, Eskil; Hammer, Morten; Munkejord, Svend Tollak; Müller, Bernhard; Ytrehus, Tor

    2017-11-01

    In film boiling, a continuous sub-millimeter vapor film forms between a liquid and a heated surface, insulating the two from each other. While quite accurate steady state solutions are readily obtained, the intermediate Reynolds numbers can make transient analysis challenging. The present work is a theoretical study of film boiling instabilities. We study the formation of travelling waves that are a combination of Kelvin-Helmholtz and the Rayleigh-Taylor instabilities. In particular, we study how the nature of this process depends on the Reynolds number, the Bond number, and the inclination of the submerged heated plate. In addition we extend the analysis to the case of a submerged heated sphere. Modelling of the transient dynamics of such films is important for answering practical questions such as how instabilities affect the overall heat transfer, and whether they can lead to complete film boiling collapse (Leidenfrost point). This work has been financed under the MAROFF program. We acknowledge the Research Council of Norway (244076/O80) and The Gas Technology Centre NTNU-SINTEF (GTS) for support.

  14. Large-scale Generation of Patterned Bubble Arrays on Printed Bi-functional Boiling Surfaces

    PubMed Central

    Choi, Chang-Ho; David, Michele; Gao, Zhongwei; Chang, Alvin; Allen, Marshall; Wang, Hailei; Chang, Chih-hung

    2016-01-01

    Bubble nucleation control, growth and departure dynamics is important in understanding boiling phenomena and enhancing nucleate boiling heat transfer performance. We report a novel bi-functional heterogeneous surface structure that is capable of tuning bubble nucleation, growth and departure dynamics. For the fabrication of the surface, hydrophobic polymer dot arrays are first printed on a substrate, followed by hydrophilic ZnO nanostructure deposition via microreactor-assisted nanomaterial deposition (MAND) processing. Wettability contrast between the hydrophobic polymer dot arrays and aqueous ZnO solution allows for the fabrication of heterogeneous surfaces with distinct wettability regions. Heterogeneous surfaces with various configurations were fabricated and their bubble dynamics were examined at elevated heat flux, revealing various nucleate boiling phenomena. In particular, aligned and patterned bubbles with a tunable departure frequency and diameter were demonstrated in a boiling experiment for the first time. Taking advantage of our fabrication method, a 6 inch wafer size heterogeneous surface was prepared. Pool boiling experiments were also performed to demonstrate a heat flux enhancement up to 3X at the same surface superheat using bi-functional surfaces, compared to a bare stainless steel surface. PMID:27034255

  15. Large-scale Generation of Patterned Bubble Arrays on Printed Bi-functional Boiling Surfaces

    NASA Astrophysics Data System (ADS)

    Choi, Chang-Ho; David, Michele; Gao, Zhongwei; Chang, Alvin; Allen, Marshall; Wang, Hailei; Chang, Chih-Hung

    2016-04-01

    Bubble nucleation control, growth and departure dynamics is important in understanding boiling phenomena and enhancing nucleate boiling heat transfer performance. We report a novel bi-functional heterogeneous surface structure that is capable of tuning bubble nucleation, growth and departure dynamics. For the fabrication of the surface, hydrophobic polymer dot arrays are first printed on a substrate, followed by hydrophilic ZnO nanostructure deposition via microreactor-assisted nanomaterial deposition (MAND) processing. Wettability contrast between the hydrophobic polymer dot arrays and aqueous ZnO solution allows for the fabrication of heterogeneous surfaces with distinct wettability regions. Heterogeneous surfaces with various configurations were fabricated and their bubble dynamics were examined at elevated heat flux, revealing various nucleate boiling phenomena. In particular, aligned and patterned bubbles with a tunable departure frequency and diameter were demonstrated in a boiling experiment for the first time. Taking advantage of our fabrication method, a 6 inch wafer size heterogeneous surface was prepared. Pool boiling experiments were also performed to demonstrate a heat flux enhancement up to 3X at the same surface superheat using bi-functional surfaces, compared to a bare stainless steel surface.

  16. SU-F-T-298: The Impact of Modeling the Treatment Couch On Patient Specific VMAT QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gelover, E; Dalhart, A; Hyer, D

    2016-06-15

    Purpose: The aim of this work is to quantify the impact of modeling the treatment couch on the passing rate of ion chamber measurements during VMAT quality assurance. Methods: For the initial characterization, attenuation and surface dose measurements were performed following the guidelines of TG-176 for the Civco Universal couch top using an Elekta VersaHD accelerator at an energy of 6 MV. A simulation CT was performed to aid in the creation of contours for representing the shape and size of the couch top in the treatment planning system (TPS). A uniform value of density for the couch wall wasmore » determined by comparing the ratios of ion chamber measurements made in a 30×30×11 cm3 water phantom with the TPS dose values of a plan with the same geometry. At our institution, patient specific quality assurance is performed using a Sun Nuclear ArcCheck with a multi-plug for chamber measurements, a 0.125cc PTW TN31010 chamber, and a Sun Nuclear 1010 electrometer. Ten VMAT plans were transferred into the phantom geometry created in the TPS with two settings: with and without the couch. The chamber measurements were compared to both treatment plans. Results: A maximum attenuation of 3.6% was observed when the gantry angle was set to 120 and 240 degrees, passing obliquely through the couch. A uniform density of 0.6 g/cm3 for the couch wall was determined in the TPS by comparison with measured data. The VMAT ion chamber measurement/plan ratios systematically improved by 1.79% ±0.53% for all patients when the couch was included in the calculation. Conclusion: The attenuation and surface dose changes produced by the Civco couch can generate observable dose difference in VMAT plans. Including a couch model in the phantom plan used for patient specific VMAT QA can improve the ionization chamber agreement by up to ∼2%.« less

  17. Transient boiling heat transfer in saturated liquid nitrogen and F113 at standard and zero gravity

    NASA Technical Reports Server (NTRS)

    Oker, E.; Merte, H., Jr.

    1973-01-01

    Transient and steady state nucleate boiling in saturated LN2 and F113 at standard and near zero gravity conditions were investigated for the horizontal up, vertical and horizontal down orientations of the heating surface. Two distinct regimes of heat transfer mechanisms were observed during the interval from the step increase of power input to the onset of nucleate boiling: the conduction and convection dominated regimes. The time duration in each regime was considerably shorter with LN2 than with F113, and decreased as heat flux increased, as gravity was reduced, and as the orientation was changed from horizontal up to horizontal down. In transient boiling, boiling initiates at a single point following the step increase in power, and then spreads over the surface. The delay time for the inception of boiling at the first site, and the velocity of spread of boiling varies depending upon the heat flux, orientation, body force, surface roughness and liquid properties, and are a consequence of changes in boundary layer temperature levels associated with changes in natural convection. Following the step increase in power input, surface temperature overshoot and undershoot occur before the steady state boiling temperature level is established.

  18. Assessing the microbiological performance and potential cost of boiling drinking water in urban Zambia.

    PubMed

    Psutka, Rebecca; Peletz, Rachel; Michelo, Sandford; Kelly, Paul; Clasen, Thomas

    2011-07-15

    Boiling is the most common method of disinfecting water in the home and the benchmark against which other point-of-use water treatment is measured. In a six-week study in peri-urban Zambia, we assessed the microbiological effectiveness and potential cost of boiling among 49 households without a water connection who reported "always" or "almost always" boiling their water before drinking it. Source and household drinking water samples were compared weekly for thermotolerant coliforms (TTC), an indicator of fecal contamination. Demographics, costs, and other information were collected through surveys and structured observations. Drinking water samples taken at the household (geometric mean 7.2 TTC/100 mL, 95% CI, 5.4-9.7) were actually worse in microbiological quality than source water (geometric mean 4.0 TTC/100 mL, 95% CI, 3.1-5.1) (p < 0.001), although both are relatively low levels of contamination. Only 60% of drinking water samples were reported to have actually been boiled at the time of collection from the home, suggesting over-reporting and inconsistent compliance. However, these samples were of no higher microbiological quality. Evidence suggests that water quality deteriorated after boiling due to lack of residual protection and unsafe storage and handling. The potential cost of fuel or electricity for boiling was estimated at 5% and 7% of income, respectively. In this setting where microbiological water quality was relatively good at the source, safe-storage practices that minimize recontamination may be more effective in managing the risk of disease from drinking water at a fraction of the cost of boiling.

  19. Multiphysics modeling of two-phase film boiling within porous corrosion deposits

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jin, Miaomiao, E-mail: mmjin@mit.edu; Short, Michael, E-mail: hereiam@mit.edu

    2016-07-01

    Porous corrosion deposits on nuclear fuel cladding, known as CRUD, can cause multiple operational problems in light water reactors (LWRs). CRUD can cause accelerated corrosion of the fuel cladding, increase radiation fields and hence greater exposure risk to plant workers once activated, and induce a downward axial power shift causing an imbalance in core power distribution. In order to facilitate a better understanding of CRUD's effects, such as localized high cladding surface temperatures related to accelerated corrosion rates, we describe an improved, fully-coupled, multiphysics model to simulate heat transfer, chemical reactions and transport, and two-phase fluid flow within these deposits.more » Our new model features a reformed assumption of 2D, two-phase film boiling within the CRUD, correcting earlier models' assumptions of single-phase coolant flow with wick boiling under high heat fluxes. This model helps to better explain observed experimental values of the effective CRUD thermal conductivity. Finally, we propose a more complete set of boiling regimes, or a more detailed mechanism, to explain recent CRUD deposition experiments by suggesting the new concept of double dryout specifically in thick porous media with boiling chimneys. - Highlights: • A two-phase model of CRUD's effects on fuel cladding is developed and improved. • This model eliminates the formerly erroneous assumption of wick boiling. • Higher fuel cladding temperatures are predicted when accounting for two-phase flow. • Double-peaks in thermal conductivity vs. heat flux in experiments are explained. • A “double dryout” mechanism in CRUD is proposed based on the model and experiments.« less

  20. Single-bubble dynamics in pool boiling of one-component fluids.

    PubMed

    Xu, Xinpeng; Qian, Tiezheng

    2014-06-01

    We numerically investigate the pool boiling of one-component fluids with a focus on the effects of surface wettability on the single-bubble dynamics. We employed the dynamic van der Waals theory [Phys. Rev. E 75, 036304 (2007)], a diffuse-interface model for liquid-vapor flows involving liquid-vapor transition in nonuniform temperature fields. We first perform simulations for bubbles on homogeneous surfaces. We find that an increase in either the contact angle or the surface superheating can enhance the bubble spreading over the heating surface and increase the bubble departure diameter as well and therefore facilitate the transition into film boiling. We then examine the dynamics of bubbles on patterned surfaces, which incorporate the advantages of both hydrophobic and hydrophilic surfaces. The central hydrophobic region increases the thermodynamic probability of bubble nucleation while the surrounding hydrophilic region hinders the continuous bubble spreading by pinning the contact line at the hydrophobic-hydrophilic intersection. This leads to a small bubble departure diameter and therefore prevents the transition from nucleate boiling into film boiling. With the bubble nucleation probability increased and the bubble departure facilitated, the efficiency of heat transfer on such patterned surfaces is highly enhanced, as observed experimentally [Int. J. Heat Mass Transfer 57, 733 (2013)]. In addition, the stick-slip motion of contact line on patterned surfaces is demonstrated in one-component fluids, with the effect weakened by surface superheating.

  1. WE-E-BRD-01: HDR Brachytherapy I: Overview of Clinical Application and QA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Libby, B; Showalter, T

    2014-06-15

    With the increased usage of high dose rate (HDR) brachytherapy and the introduction of dedicated image guided brachytherapy suites, it is necessary to review the processes and procedures associated with safely delivering these treatments in the expedited time scales that dedicated treatment suites afford. The speakers will present the clinical aspects of switching from LDR to HDR treatments, including guidelines for patient selection, and the clinical outcomes comparing LDR to HDR. The speakers will also discuss the HDR treatment process itself, because the shortened clinical timeline involved with a streamlined scan/plan/treat workflow can introduce other issues. Safety and QA aspectsmore » involved with the streamlined process, including increased personnel required for parallel tasks, and possible interfering tasks causing delays in patient treatments will also be discussed. Learning Objectives: To understand the clinical aspects of HDR Brachytherapy, including common clinical indications, patient selection, and the evolving evidence in support of this therapeutic modality To review the current prominent clinical trials for HDR brachytherapy To interpret the established guidelines for HDR brachytherapy quality assurance for implementation into practical clinical settings. To introduce the basic requirements for image guided brachytherapy.« less

  2. Insulation of Nitrocellulose Boiling Tubs at Radford Army Ammunition Plant

    DTIC Science & Technology

    1982-03-01

    control system. The amount of steam usea for the on-boil cycle with the single-sensor autocontrol averaged 647 kg/hr (1426 lb/hr) (test 1, table 2...This was a reduc- tion of 210 kg/hr (463 lb/hr) over the manually controlled uninsulated tub. Steam usage with the single sensor autocontrol and...uninsulated tub. At times durin)g the on- boil cycle of tests I and 2, the temperature of the manual sensor was different from the autocontrol sensor indicating

  3. Bubble Dynamics, Two-Phase Flow, and Boiling Heat Transfer in Microgravity

    NASA Technical Reports Server (NTRS)

    Chung, Jacob N.

    1996-01-01

    The objective of the research is to study the feasibility of employing an external force to replace the buoyancy force in order to maintain nucleate boiling in microgravity. We have found that a bulk velocity field, an electric field and an acoustic field could each play the role of the gravity field in microgravity. Nucleate boiling could be maintained by any one of the three external force fields in space.

  4. Microbiological effectiveness and cost of boiling to disinfect drinking water in rural Vietnam.

    PubMed

    Clasen, Thomas F; Thao, Do Hoang; Boisson, Sophie; Shipin, Oleg

    2008-06-15

    Despite certain shortcomings, boiling is still the most common means of treating water in the home and the benchmark against which alternative household-based disinfection and filtration methods must be measured. We assessed the microbiological effectiveness and cost of boiling among a vulnerable population relying on unimproved water sources and commonly practicing boiling as a means of disinfecting water. In a 12 week study among 50 households from a rural community in Vietnam, boiling was associated with a 97% reduction in geometric mean thermotolerant coliforms (TTCs) (p < 0.001). Despite high levels of faecal contamination in source water, 37% of stored water samples from self-reported boilers met the WHO standard for safe drinking water (0 TTC/100 mL), and 38.3% fell within the low risk category (1--10 TTC/100 mL). Nevertheless, 60.5% of stored drinking water samples were positive for TTC, with 22.2% falling into the medium risk category (11--100 TTC/100 mL). The estimated cost of wood used to boil water was US$ 0.272 per month for wood collectors and US$ 1.68 per month for wood purchasers, representing approximately 0.48% to 1.04%, respectively, of the average monthly income of participating households.

  5. 23. RW Meyer Sugar Mill: 18761889. Boiling House Interior, 1878. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    23. RW Meyer Sugar Mill: 1876-1889. Boiling House Interior, 1878. View: North Wall of boiling house. In the original structure the three windows on the right admitted light and air from the outside. A shed occupied the left side of the wall outside (hence no windows). in 1881 the construction of the cooling shed closed in the right three windows. The sorghum is in the foreground. The centrifugals are in the left rear. - R. W. Meyer Sugar Mill, State Route 47, Kualapuu, Maui County, HI

  6. Poster - 10: QA of Ultrasound Images for Prostate Brachytherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Szpala, Stanislaw; Kohli, Kirpal S.

    Purpose: The current QA protocol of ultrasound systems used in prostate brachytherapy (TG128) addresses geometrical verifications, but the scope of evaluation of image quality is limited. We recognized importance of the latter in routine practice, and designed a protocol for QA of the images. Methods: Images of an ultrasound prostate phantom (CIRS053) were collected with BK Flex Focus 400. The images were saved as bmp after adjusting the gain to 50% for consistent results. Mean pixel values and signal to noise ratio were inspected in the representative sections of the phantom, including the mock prostate and the unechoic medium. Constancymore » of these numbers over a one year period was looked at. Results: The typical intensity in the mock prostate region in the transverse images ranged between 95 and 118 (out of 256), and the signal to noise was about 10. The intensity in the urethra region was about 170±40, and the unechoic medium was 2±2. The mean and the signal to noise ratio remained almost unchanged after a year, while the signal in the unechoic medium increased to about 7±4. Similar values were obtained in the sagittal images. Conclusions: The image analysis discussed above allows quick evaluation of constancy of the image quality. This may be also useful in troubleshooting image-quality problems during routine exams, which might not be due to deterioration of the US system, but other reasons, e.g. variations in tissue properties or air being trapped between the probe and the anatomy.« less

  7. SU-E-J-104: Evaluation of Accuracy for Various Deformable Image Registrations with Virtual Deformation QA Software

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Han, S; Kim, K; Kim, M

    Purpose: The accuracy of deformable image registration (DIR) has a significant dosimetric impact in radiation treatment planning. We evaluated accuracy of various DIR algorithms using virtual deformation QA software (ImSimQA, Oncology System Limited, UK). Methods: The reference image (Iref) and volume (Vref) was first generated with IMSIMQA software. We deformed Iref with axial movement of deformation point and Vref depending on the type of deformation that are the deformation1 is to increase the Vref (relaxation) and the deformation 2 is to decrease the Vref (contraction) .The deformed image (Idef) and volume (Vdef) were inversely deformed to Iref and Vref usingmore » DIR algorithms. As a Result, we acquired deformed image (Iid) and volume (Vid). The DIR algorithms were optical flow (HS, IOF) and demons (MD, FD) of the DIRART. The image similarity evaluation between Iref and Iid was calculated by Normalized Mutual Information (NMI) and Normalized Cross Correlation (NCC). The value of Dice Similarity Coefficient (DSC) was used for evaluation of volume similarity. Results: When moving distance of deformation point was 4 mm, the value of NMI was above 1.81 and NCC was above 0.99 in all DIR algorithms. Since the degree of deformation was increased, the degree of image similarity was decreased. When the Vref increased or decreased about 12%, the difference between Vref and Vid was within ±5% regardless of the type of deformation. The value of DSC was above 0.95 in deformation1 except for the MD algorithm. In case of deformation 2, that of DSC was above 0.95 in all DIR algorithms. Conclusion: The Idef and Vdef have not been completely restored to Iref and Vref and the accuracy of DIR algorithms was different depending on the degree of deformation. Hence, the performance of DIR algorithms should be verified for the desired applications.« less

  8. Cryogenic Boil-Off Reduction System Testing

    NASA Technical Reports Server (NTRS)

    Plachta, David W.; Johnson, Wesley L.; Feller, Jeffery

    2014-01-01

    The Cryogenic Boil-Off Reduction System was tested with LH2 and LOX in a vacuum chamber to simulate space vacuum and the temperatures of low Earth orbit. Testing was successful and results validated the scaling study model that predicts active cooling reduces upper stage cryogenic propulsion mass for loiter periods greater than 2 weeks.

  9. Zero Boil-Off Tank (ZBOT) Experiment

    NASA Technical Reports Server (NTRS)

    Mcquillen, John

    2016-01-01

    The Zero-Boil-Off Tank (ZBOT) experiment has been developed as a small scale ISS experiment aimed at delineating important fluid flow, heat and mass transport, and phase change phenomena that affect cryogenic storage tank pressurization and pressure control in microgravity. The experiments use a simulant transparent low boiling point fluid (PnP) in a sealed transparent Dewar to study and quantify: (a) fluid flow and thermal stratification during pressurization; (b) mixing, thermal destratification, depressurization, and jet-ullage penetration during pressure control by jet mixing. The experiment will provide valuable microgravity empirical two-phase data associated with the above-mentioned physical phenomena through highly accurate local wall and fluid temperature and pressure measurements, full-field phase-distribution and flow visualization. Moreover, the experiments are performed under tightly controlled and definable heat transfer boundary conditions to provide reliable high-fidelity data and precise input as required for validation verification of state-of-the-art two-phase CFD models developed as part of this research and by other groups in the international scientific and cryogenic fluid management communities.

  10. Cork boiling wastewater treatment and reuse through combination of advanced oxidation technologies.

    PubMed

    Ponce-Robles, L; Miralles-Cuevas, S; Oller, I; Agüera, A; Trinidad-Lozano, M J; Yuste, F J; Malato, S

    2017-03-01

    Industrial preparation of cork consists of its immersion for approximately 1 hour in boiling water. The use of herbicides and pesticides in oak tree forests leads to absorption of these compounds by cork; thus, after boiling process, they are present in wastewater. Cork boiling wastewater shows low biodegradability and high acute toxicity involving partial inhibition of their biodegradation when conventional biological treatment is applied. In this work, a treatment line strategy based on the combination of advanced physicochemical technologies is proposed. The final objective is the reuse of wastewater in the cork boiling process; thus, reducing consumption of fresh water in the industrial process itself. Coagulation pre-treatment with 0.5 g/L of FeCl 3 attained the highest turbidity elimination (86 %) and 29 % of DOC elimination. Similar DOC removal was attained when using 1 g/L of ECOTAN BIO (selected for ozonation tests), accompanied of 64 % of turbidity removal. Ozonation treatments showed less efficiency in the complete oxidation of cork boiling wastewater, compared to solar photo-Fenton process, under the studied conditions. Nanofiltration system was successfully employed as a final purification step with the aim of obtaining a high-quality reusable permeate stream. Monitoring of unknown compounds by LC-QTOF-MS allowed the qualitative evaluation of the whole process. Acute and chronic toxicity as well as biodegradability assays were performed throughout the whole proposed treatment line.

  11. Magnetic resonance imaging of boiling induced by high intensity focused ultrasound

    PubMed Central

    Khokhlova, Tatiana D.; Canney, Michael S.; Lee, Donghoon; Marro, Kenneth I.; Crum, Lawrence A.; Khokhlova, Vera A.; Bailey, Michael R.

    2009-01-01

    Both mechanically induced acoustic cavitation and thermally induced boiling can occur during high intensity focused ultrasound (HIFU) medical therapy. The goal was to monitor the temperature as boiling was approached using magnetic resonance imaging (MRI). Tissue phantoms were heated for 20 s in a 4.7-T magnet using a 2-MHz HIFU source with an aperture and radius of curvature of 44 mm. The peak focal pressure was 27.5 MPa with corresponding beam width of 0.5 mm. The temperature measured in a single MRI voxel by water proton resonance frequency shift attained a maximum value of only 73 °C after 7 s of continuous HIFU exposure when boiling started. Boiling was detected by visual observation, by appearance on the MR images, and by a marked change in the HIFU source power. Nonlinear modeling of the acoustic field combined with a heat transfer equation predicted 100 °C after 7 s of exposure. Averaging of the calculated temperature field over the volume of the MRI voxel (0.3×0.5×2 mm3) yielded a maximum of 73 °C that agreed with the MR thermometry measurement. These results have implications for the use of MRI-determined temperature values to guide treatments with clinical HIFU systems. PMID:19354416

  12. The Parable of the Boiled System Safety Professional: Drift to Failure

    NASA Technical Reports Server (NTRS)

    Shivers, C. Herbert

    2011-01-01

    Recall from the Parable of the Boiled Frog, that tossing a frog into boiling water causes the frog to jump out and hop away while placing a frog in suitable temperature water and slowly bringing the water to a boil results in the frog boiling due to not being aware of the slowly increasing danger, theoretically, of course. System safety professionals must guard against allowing dangers to creep unnoticed into their projects and be ever alert to notice signs of impending problems. People have used various phrases related to the idea, most notably, latent conditions, James Reason in Managing the Risks of Organizational Accidents (1, pp 10-11), Drift to Failure, Sydney Dekker (2, pp 82-86) in Resilience Engineering: Chronicling the Emergence of Confused Consensus in Resilience Engineering: Concepts and Precepts, Hollnagel, Woods and Leveson, and normalization of deviance, Diane Vaughan in The Challenger Launch Decision: Risky Technology, Culture, and Deviance at NASA (3). Reason also said, If eternal vigilance is the price of liberty, then chronic unease is the price of safety (1, p 37). Our challenge as system safety professionals is to be aware of the emergence of signals that warn us of slowly eroding safety margins. This paper will discuss how system safety professionals might better perform in that regard.

  13. Experimental study of flash boiling spray vaporization through quantitative vapor concentration and liquid temperature measurements

    NASA Astrophysics Data System (ADS)

    Zhang, Gaoming; Hung, David L. S.; Xu, Min

    2014-08-01

    Flash boiling sprays of liquid injection under superheated conditions provide the novel solutions of fast vaporization and better air-fuel mixture formation for internal combustion engines. However, the physical mechanisms of flash boiling spray vaporization are more complicated than the droplet surface vaporization due to the unique bubble generation and boiling process inside a superheated bulk liquid, which are not well understood. In this study, the vaporization of flash boiling sprays was investigated experimentally through the quantitative measurements of vapor concentration and liquid temperature. Specifically, the laser-induced exciplex fluorescence technique was applied to distinguish the liquid and vapor distributions. Quantitative vapor concentration was obtained by correlating the intensity of vapor-phase fluorescence with vapor concentration through systematic corrections and calibrations. The intensities of two wavelengths were captured simultaneously from the liquid-phase fluorescence spectra, and their intensity ratios were correlated with liquid temperature. The results show that both liquid and vapor phase of multi-hole sprays collapse toward the centerline of the spray with different mass distributions under the flash boiling conditions. Large amount of vapor aggregates along the centerline of the spray to form a "gas jet" structure, whereas the liquid distributes more uniformly with large vortexes formed in the vicinity of the spray tip. The vaporization process under the flash boiling condition is greatly enhanced due to the intense bubble generation and burst. The liquid temperature measurements show strong temperature variations inside the flash boiling sprays with hot zones present in the "gas jet" structure and vortex region. In addition, high vapor concentration and closed vortex motion seem to have inhibited the heat and mass transfer in these regions. In summary, the vapor concentration and liquid temperature provide detailed information

  14. Pool Boiling Heat Transfer on structured Surfaces

    NASA Astrophysics Data System (ADS)

    Addy, J.; Olbricht, M.; Müller, B.; Luke, A.

    2016-09-01

    The development in the process and energy sector shows the importance of efficient utilization of available resources to improve thermal devices. To achieve this goal, all thermal components have to be optimized continuously. Various applications of multi-phase heat and mass transfer have to be improved. Therefore, the heat transfer and the influence of surface roughness in nucleate boiling with the working fluid propane is experimentally investigated on structured mild steel tubes, because only few data are available in the literature. The mild steel tube is sandblasted to obtain different surface roughness. The measurements are carried out over wide ranges of heat flux and pressure. The experimental results are compared with correlations from literature and the effect of surface roughness on the heat transfer is discussed. It is shown that the heat transfer coefficient increases with increasing surface roughness, heat flux and reduced pressure at nucleate pool boiling.

  15. Single-bubble dynamics in pool boiling of one-component fluids

    NASA Astrophysics Data System (ADS)

    Xu, Xinpeng; Qian, Tiezheng

    2014-06-01

    We numerically investigate the pool boiling of one-component fluids with a focus on the effects of surface wettability on the single-bubble dynamics. We employed the dynamic van der Waals theory [Phys. Rev. E 75, 036304 (2007), 10.1103/PhysRevE.75.036304], a diffuse-interface model for liquid-vapor flows involving liquid-vapor transition in nonuniform temperature fields. We first perform simulations for bubbles on homogeneous surfaces. We find that an increase in either the contact angle or the surface superheating can enhance the bubble spreading over the heating surface and increase the bubble departure diameter as well and therefore facilitate the transition into film boiling. We then examine the dynamics of bubbles on patterned surfaces, which incorporate the advantages of both hydrophobic and hydrophilic surfaces. The central hydrophobic region increases the thermodynamic probability of bubble nucleation while the surrounding hydrophilic region hinders the continuous bubble spreading by pinning the contact line at the hydrophobic-hydrophilic intersection. This leads to a small bubble departure diameter and therefore prevents the transition from nucleate boiling into film boiling. With the bubble nucleation probability increased and the bubble departure facilitated, the efficiency of heat transfer on such patterned surfaces is highly enhanced, as observed experimentally [Int. J. Heat Mass Transfer 57, 733 (2013), 10.1016/j.ijheatmasstransfer.2012.10.080]. In addition, the stick-slip motion of contact line on patterned surfaces is demonstrated in one-component fluids, with the effect weakened by surface superheating.

  16. SU-E-J-199: A Software Tool for Quality Assurance of Online Replanning with MR-Linac

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chen, G; Ahunbay, E; Li, X

    2015-06-15

    Purpose: To develop a quality assurance software tool, ArtQA, capable of automatically checking radiation treatment plan parameters, verifying plan data transfer from treatment planning system (TPS) to record and verify (R&V) system, performing a secondary MU calculation considering the effect of magnetic field from MR-Linac, and verifying the delivery and plan consistency, for online replanning. Methods: ArtQA was developed by creating interfaces to TPS (e.g., Monaco, Elekta), R&V system (Mosaiq, Elekta), and secondary MU calculation system. The tool obtains plan parameters from the TPS via direct file reading, and retrieves plan data both transferred from TPS and recorded during themore » actual delivery in the R&V system database via open database connectivity and structured query language. By comparing beam/plan datasets in different systems, ArtQA detects and outputs discrepancies between TPS, R&V system and secondary MU calculation system, and delivery. To consider the effect of 1.5T transverse magnetic field from MR-Linac in the secondary MU calculation, a method based on modified Clarkson integration algorithm was developed and tested for a series of clinical situations. Results: ArtQA is capable of automatically checking plan integrity and logic consistency, detecting plan data transfer errors, performing secondary MU calculations with or without a transverse magnetic field, and verifying treatment delivery. The tool is efficient and effective for pre- and post-treatment QA checks of all available treatment parameters that may be impractical with the commonly-used visual inspection. Conclusion: The software tool ArtQA can be used for quick and automatic pre- and post-treatment QA check, eliminating human error associated with visual inspection. While this tool is developed for online replanning to be used on MR-Linac, where the QA needs to be performed rapidly as the patient is lying on the table waiting for the treatment, ArtQA can be used as a general

  17. SU-F-J-51: A Cone-Based Scintillator Detector for IGRT QA for Scattered and Scanning Proton Therapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Oesten, H; Clasie, B; Jee, K

    Purpose: IGRT commissioning and QA are critical components for precise delivery of proton treatment beams to patients. In order to ensure high quality IGRT, a new cone-based scintillator detector was evaluated for our QA activities for double-scattered and scanning proton modalities. This allows a routine evaluation of the gantry-angle dependent position offset between the radiation and imaging. Methods: The cone-based scintillator detector (XRV-124, Logos Systems, Int’l CA, USA) features a unique configuration of measuring stereotactic paths of proton and x-ray beams in a single setup with arbitrary gantry angles. For the beams-eye-view (BEV) analysis of x-ray crosshair images, a cylindricalmore » representation of the cone image was newly developed. The calibration accuracy was evaluated using different CT resolutions for a range of 55 – 95mm in patient’s cranial direction and ±9mm in the lateral direction. Energy-dependent spot sizes (σ) of pencil beams were characterized and compared to measurements by the MatriXX detector (IBA, Germany). Iso-centric deviations between radiation and x-ray imaging were characterized as a function of gantry angle. Results: The position calibration of the detector was successfully verified with a reproducible positioning by x-ray imaging. The measurements were reproducible within clinical tolerances (±1mm). The spot size vs. energy at zero gantry angle measured with the scintillating cone detector agreed with the MatriXX detector measurements within 17%. Conclusion: The new approach to investigate the accuracy of IGRT and pencil beam properties could successfully be implemented into the QA program. The system will improve efficiency in our QA activities for proton treatments.« less

  18. QUALITY SYSTEMS AND IMPLEMENTATION PLAN FOR A PILOT STUDY OF CHILDREN'S TOTAL EXPOSURE TO PERSISTENT PESTICIDES AND OTHER PERSISTENT ORGANIC PESTICIDES (CTEPP)

    EPA Science Inventory

    The Quality System Implementation Plan (QSIP) describes the quality assurance and quality control procedures developed for the CTEPP study. It provides the QA/QC procedures used in recruitment of subjects, sample field collection, sample extraction and analysis, data storage, and...

  19. Visual Analysis of the Daily QA Results of Photon and Electron Beams of a Trilogy Linac over a Five-year Period

    PubMed Central

    Chan, Maria F.; Li, Qiongge; Tang, Xiaoli; Li, Xiang; Li, Jingdong; Tang, Grace; Hunt, Margie A.; Deasy, Joseph O.

    2016-01-01

    Data visualization technique was applied to analyze the daily QA results of photon and electron beams. Special attention was paid to any trend the beams might display. A Varian Trilogy Linac equipped with dual photon energies and five electron energies was commissioned in early 2010. Daily Linac QA tests including the output constancy, beam flatness and symmetry (radial and transverse directions) were performed with an ionization chamber array device (QA BeamChecker Plus, Standard Imaging). The data of five years were collected and analyzed. For each energy, the measured data were exported and processed for visual trending using an in-house Matlab program. These daily data were cross-correlated with the monthly QA and annual QA results, as well as the preventive maintenance records. Majority of the output were within 1% of variation, with a consistent positive/upward drift for all seven energies (~+0.25% per month). The baseline of daily device is reset annually right after the TG-51 calibration. This results in a sudden drop of the output. On the other hand, the large amount of data using the same baseline exhibits a sinusoidal behavior (cycle = 12 months; amplitude = 0.8%, 0.5% for photons, electrons, respectively) on symmetry and flatness when normalization of baselines is accounted for. The well known phenomenon of new Linac output drift was clearly displayed. This output drift was a result of the air leakage of the over-pressurized sealed monitor chambers for the specific vendor. Data visualization is a new trend in the era of big data in radiation oncology research. It allows the data to be displayed visually and therefore more intuitive. Based on the visual display from the past, the physicist might predict the trend of the Linac and take actions proactively. It also makes comparisons, alerts failures, and potentially identifies causalities. PMID:27547595

  20. Land ECVs from QA4ECV using an optimal estimation framework

    NASA Astrophysics Data System (ADS)

    Muller, Jan-Peter; Kharbouche, Said; Lewis, Philip; Danne, Olaf; Blessing, Simon; Giering, Ralf; Gobron, Nadine; Lanconelli, Christian; Govaerts, Yves; Schulz, Joerg; Doutriaux-Boucher, Marie; Lattanzio, Alessio; Aoun, Youva

    2017-04-01

    In the ESA-DUE GlobAlbedo project (http://www.GlobAlbedo.org), a 15 year record of land surface albedo was generated from the European VEGETATION & MERIS sensors using optimal estimation. This was based on 3 broadbands (0.4-0.7, 0.7-3, 0.4-3µm) and fused data at level-2 after converting from spectral narrowband to these 3 broadbands with surface BRFs. A 10 year long record of land surface albedo climatology was generated from Collection 5 of the MODIS BRDF product for these same broadbands. This was employed as an a priori estimate for an optimal estimation based retrieval of land surface albedo when there were insufficient samples from the European sensors. This so-called MODIS prior was derived at 1km from the 500m MOD43A1,2 BRDF inputs every 8 days using the QA bits and the method described in the GlobAlbedo ATBD which is available from the website (http://www.globalbedo.org/docs/GlobAlbedo_Albedo_ATBD_V4.12.pdf). In the ESA-STSE WACMOS-ET project, FastOpt generated fapar & LAI based on this GlobAlbedo BRDF with associated per pixel uncertainty using the TIP framework. In the successor EU-FP7-QA4ECV* project, we have developed a 33 year record (1981-2014) of Earth surface spectral and broadband albedo (i.e. including the ocean and sea-ice) using optimal estimation for the land and where available, relevant sensors for "instantaneous" retrievals over the poles (Kharbouche & Muller, this conference). This requires the longest possible land surface spectral and broadband BRDF record that can only be supplied by a 16 year of MODIS Collection 6 BRDFs at 500m but produced on a daily basis. The CEMS Big Data computer at RAL was used to generate 7 spectral bands and 3 broadband BRDF with and without snow and snow_only. We will discuss the progress made since the start of the QA4ECV project on the production of a new fused land surface BRDF/albedo spectral and broadband CDR product based on four European sensors: MERIS, (A)ATSR(2), VEGETATION, PROBA-V and two US sensors

  1. Estimation of boiling points using density functional theory with polarized continuum model solvent corrections.

    PubMed

    Chan, Poh Yin; Tong, Chi Ming; Durrant, Marcus C

    2011-09-01

    An empirical method for estimation of the boiling points of organic molecules based on density functional theory (DFT) calculations with polarized continuum model (PCM) solvent corrections has been developed. The boiling points are calculated as the sum of three contributions. The first term is calculated directly from the structural formula of the molecule, and is related to its effective surface area. The second is a measure of the electronic interactions between molecules, based on the DFT-PCM solvation energy, and the third is employed only for planar aromatic molecules. The method is applicable to a very diverse range of organic molecules, with normal boiling points in the range of -50 to 500 °C, and includes ten different elements (C, H, Br, Cl, F, N, O, P, S and Si). Plots of observed versus calculated boiling points gave R²=0.980 for a training set of 317 molecules, and R²=0.979 for a test set of 74 molecules. The role of intramolecular hydrogen bonding in lowering the boiling points of certain molecules is quantitatively discussed. Crown Copyright © 2011. Published by Elsevier Inc. All rights reserved.

  2. Large-scale boiling experiments of the flooded cavity concept for in-vessel core retention

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chu, T.Y.; Slezak, S.E.; Bentz, J.H.

    1994-03-01

    This paper presents results of ex-vessel boiling experiments performed in the CYBL (CYlindrical BoiLing) facility. CYBL is a reactor-scale facility for confirmatory research of the flooded cavity concept for accident management. CYBL has a tank-within-a-tank design; the inner tank simulates the reactor vessel and the outer tank simulates the reactor cavity. Experiments with uniform and edge-peaked heat flux distributions up to 20 W/cm{sup 2} across the vessel bottom were performed. Boiling outside the reactor vessel was found to be subcooled nucleate boiling. The subcooling is mainly due to the gravity head which results from flooding the sides of the reactormore » vessel. The boiling process exhibits a cyclic pattern with four distinct phases: direct liquid/solid contact, bubble nucleation and growth, coalescence, and vapor mass dispersion (ejection). The results suggest that under prototypic heat load and heat flux distributions, the flooded cavity in a passive pressurized water reactor like the AP-600 should be capable of cooling the reactor pressure vessel in the central region of the lower head that is addressed by these tests.« less

  3. QA-driven Guidelines Generation for Bacteriotherapy

    PubMed Central

    Pasche, Emilie; Teodoro, Douglas; Gobeill, Julien; Ruch, Patrick; Lovis, Christian

    2009-01-01

    PURPOSE We propose a question-answering (QA) driven generation approach for automatic acquisition of structured rules that can be used in a knowledge authoring tool for antibiotic prescription guidelines management. METHODS: The rule generation is seen as a question-answering problem, where the parameters of the questions are known items of the rule (e.g. an infectious disease, caused by a given bacterium) and answers (e.g. some antibiotics) are obtained by a question-answering engine. RESULTS: When looking for a drug given a pathogen and a disease, top-precision of 0.55 is obtained by the combination of the Boolean engine (PubMed) and the relevance-driven engine (easyIR), which means that for more than half of our evaluation benchmark at least one of the recommended antibiotics was automatically acquired by the rule generation method. CONCLUSION: These results suggest that such an automatic text mining approach could provide a useful tool for guidelines management, by improving knowledge update and discovery. PMID:20351908

  4. Transition from Pool to Flow Boiling: The Effect of Reduced Gravity

    NASA Technical Reports Server (NTRS)

    Dhir, Vijay K.

    2004-01-01

    Applications of boiling heat transfer in space can be found in the areas of thermal management, fluid handling and control, power systems, on-orbit storage and supply systems for cryogenic propellants and life support fluids, and for cooling of electronic packages for power systems associated with various instrumentation and control systems. Recent interest in exploration of Mars and other planets, and the concepts of in-situ resource utiliLation on Mars highlights the need to understand the effect of gravity on boiling heat transfer at gravity levels varying from 1>= g/g(sub e) >=10(exp -6). The objective of the proposed work was to develop a mechanistic understanding of nucleate boiling and critical heat flux under low and micro-gravity conditions when the velocity of the imposed flow is small. For pool boiling, the effect of reduced gravity is to stretch both the length scale as well as the time scale for the boiling process. At high flow velocities, the inertia of the liquid determines the time and the length scales and as such the gravitational acceleration plays little role. However, at low velocities and at low gravity levels both liquid inertia and buoyancy are of equal importance. At present, we have little understanding of the interacting roles of gravity and liquid inertia on the nucleate boiling process. Little data that has been reported in the literature does not have much practical value in that it can not serve as a basis for design of heat exchange components to be used in space. Both experimental and complete numerical simulations of the low velocity, low-gravity nucleate boiling process were carried out. A building block type of approach was used in that first the growth and detachment process of a single bubble and flow and heat transfer associated with the sliding motion of the bubble over the heater surface after detachment was studied. Liquid subcooling and flow velocity were varied parametrically. The experiments were conducted at 1 g(sub e

  5. Issue a Boil-Water Advisory or Wait for Definitive Information? A Decision Analysis

    PubMed Central

    Wagner, Michael M.; Wallstrom, Garrick L.; Onisko, Agnieszka

    2005-01-01

    Objective Study the decision to issue a boil-water advisory in response to a spike in sales of diarrhea remedies or wait 72 hours for the results of definitive testing of water and people. Methods Decision analysis. Results In the base-case analysis, the optimal decision is test-and-wait. If the cost of issuing a boil-water advisory is less than 13.92 cents per person per day, the optimal decision is to issue the boil-water advisory immediately. Conclusions Decisions based on surveillance data that are suggestive but not conclusive about the existence of a disease outbreak can be modeled. PMID:16779145

  6. Enhanced susceptibility to acute pneumococcal otitis media in mice deficient in complement C1qa, factor B, and factor B/C2.

    PubMed

    Tong, Hua Hua; Li, Yong Xing; Stahl, Gregory L; Thurman, Joshua M

    2010-03-01

    To define the roles of specific complement activation pathways in host defense against Streptococcus pneumoniae in acute otitis media (AOM), we investigated the susceptibility to AOM in mice deficient in complement factor B and C2 (Bf/C2(-/)(-)), C1qa (C1qa(-/)(-)), and factor B (Bf(-)(/)(-)). Bacterial titers of both S. pneumoniae serotype 6A and 14 in the middle ear lavage fluid samples from Bf/C2(-/)(-), Bf(-)(/)(-), and C1qa(-/)(-) mice were significantly higher than in samples from wild-type mice 24 h after transtympanical infection (P < 0.05) and remained persistently higher in samples from Bf/C2(-/)(-) mice than in samples from wild-type mice. Bacteremia occurred in Bf/C2(-/)(-), Bf(-)(/)(-), and C1qa(-/)(-) mice infected with both strains, but not in wild-type mice. Recruitment of inflammatory cells was paralleled by enhanced production of inflammatory mediators in the middle ear lavage samples from Bf/C2(-/)(-) mice. C3b deposition on both strains was greatest for sera obtained from wild-type mice, followed by C1qa(-)(/)(-) and Bf(-)(/)(-) mice, and least for Bf/C2(-)(/)(-) mice. Opsonophagocytosis and whole-blood killing capacity of both strains were significantly decreased in the presence of sera or whole blood from complement-deficient mice compared to wild-type mice. These findings indicate that both the classical and alternative complement pathways are critical for middle ear immune defense against S. pneumoniae. The reduced capacity of complement-mediated opsonization and phagocytosis in the complement-deficient mice appears to be responsible for the impaired clearance of S. pneumoniae from the middle ear and dissemination to the bloodstream during AOM.

  7. Quality assurance of HDR prostate plans: program implementation at a community hospital.

    PubMed

    Rush, Jennifer B; Thomas, Michael D

    2005-01-01

    Adenocarcinoma of the prostate is currently the most commonly diagnosed cancer in men in the United States, and the second leading cause of cancer mortality. The utilization of radiation therapy is regarded as the definitive local therapy of choice for intermediate- and high-risk disease, in which there is increased risk for extracapsular extension, seminal vesicle invasion, or regional node involvement. High-dose-rate (HDR) brachytherapy is a logical treatment modality to deliver the boost dose to an external beam radiation therapy (EBRT) treatment to increase local control rates. From a treatment perspective, the utilization of a complicated treatment delivery system, the compressed time frame in which the procedure is performed, and the small number of large dose fractions make the implementation of a comprehensive quality assurance (QA) program imperative. One aspect of this program is the QA of the HDR treatment plan. Review of regulatory and medical physics professional publications shows that substantial general guidance is available. We provide some insight to the implementation of an HDR prostate plan program at a community hospital. One aspect addressed is the utilization of the low-dose-rate (LDR) planning system and the use of existing ultrasound image sets to familiarize the radiation therapy team with respect to acceptable HDR implant geometries. Additionally, the use of the LDR treatment planning system provided a means to prospectively determine the relationship between the treated isodose volume and the product of activity and time for the department's planning protocol prior to the first HDR implant. For the first 12 HDR prostate implants, the root-mean-square (RMS) deviation was 3.05% between the predicted product of activity and time vs. the actual plan values. Retrospective re-evaluation of the actual implant data reduced the RMS deviation to 2.36%.

  8. Kinetics-based phase change approach for VOF method applied to boiling flow

    NASA Astrophysics Data System (ADS)

    Cifani, Paolo; Geurts, Bernard; Kuerten, Hans

    2014-11-01

    Direct numerical simulations of boiling flows are performed to better understand the interaction of boiling phenomena with turbulence. The multiphase flow is simulated by solving a single set of equations for the whole flow field according to the one-fluid formulation, using a VOF interface capturing method. Interface terms, related to surface tension, interphase mass transfer and latent heat, are added at the phase boundary. The mass transfer rate across the interface is derived from kinetic theory and subsequently coupled with the continuum representation of the flow field. The numerical model was implemented in OpenFOAM and validated against 3 cases: evaporation of a spherical uniformly heated droplet, growth of a spherical bubble in a superheated liquid and two dimensional film boiling. The computational model will be used to investigate the change in turbulence intensity in a fully developed channel flow due to interaction with boiling heat and mass transfer. In particular, we will focus on the influence of the vapor bubble volume fraction on enhancing heat and mass transfer. Furthermore, we will investigate kinetic energy spectra in order to identify the dynamics associated with the wakes of vapor bubbles. Department of Applied Mathematics, 7500 AE Enschede, NL.

  9. Nucleate Pool Boiling Performance of Smooth and Finned Tube Bundles in R-113 and R-114/Oil Mixtures

    DTIC Science & Technology

    1989-06-01

    tfilm Film thermodynamic temperature (K) Tfilm Film Celcius temperature (C) Tldl Liquid temperature (C) Tld2 Liquid temperature (C) Tn Tube wall local...surface immersed in a pool of saturated liquid is the most thoroughly studied boiling heat-transfer mechanism, when compared to partial film boiling and... film boiling. Figure 2.1 shows the characteristic boiling curve of a heated surface immersed in a froon. As the surface is heated up, heat is

  10. Prediction of boiling points of organic compounds by QSPR tools.

    PubMed

    Dai, Yi-min; Zhu, Zhi-ping; Cao, Zhong; Zhang, Yue-fei; Zeng, Ju-lan; Li, Xun

    2013-07-01

    The novel electro-negativity topological descriptors of YC, WC were derived from molecular structure by equilibrium electro-negativity of atom and relative bond length of molecule. The quantitative structure-property relationships (QSPR) between descriptors of YC, WC as well as path number parameter P3 and the normal boiling points of 80 alkanes, 65 unsaturated hydrocarbons and 70 alcohols were obtained separately. The high-quality prediction models were evidenced by coefficient of determination (R(2)), the standard error (S), average absolute errors (AAE) and predictive parameters (Qext(2),RCV(2),Rm(2)). According to the regression equations, the influences of the length of carbon backbone, the size, the degree of branching of a molecule and the role of functional groups on the normal boiling point were analyzed. Comparison results with reference models demonstrated that novel topological descriptors based on the equilibrium electro-negativity of atom and the relative bond length were useful molecular descriptors for predicting the normal boiling points of organic compounds. Copyright © 2013 Elsevier Inc. All rights reserved.

  11. A Study of Nucleate Boiling with Forced Convection in Microgravity

    NASA Technical Reports Server (NTRS)

    Merte, Herman, Jr.

    1996-01-01

    Boiling is a rather imprecise term applied to the process of evaporation in which the rate of liquid-vapor phase change is large. In seeking to determine the role and significance of body forces on the process, of which buoyancy or gravity is just one agent, it becomes necessary to define the term more precisely. It is generally characterized by the formation and growth of individual vapor bubbles arising from heat transfer to the liquid, either at a solid/liquid or liquid/liquid interface, or volumetrically. The terms 'bubble' boiling and 'nucleate' boiling are frequently used, in recognition of the interactions of surface tension and other forces in producing discrete bubbles at distinctive locations (although not always). Primary considerations are that evaporation can occur only at existing liquid-vapor interfaces, so that attention must be given to the formation of an interface (the nucleation process), and that the latent heat for this evaporation can come only from the superheated liquid, so that attention must also be given to the temperature distributions in the liquid.

  12. Micro-bubble emission boiling with the cavitation bubble blow pit

    PubMed Central

    Inada, Shigeaki; Shinagawa, Kazuaki; Illias, Suhaimi Bin; Sumiya, Hiroyuki; Jalaludin, Helmisyah A.

    2016-01-01

    The miniaturization boiling (micro-bubble emission boiling [MEB]) phenomenon, with a high heat removal capacity that contributes considerably to the cooling of the divertor of the nuclear fusion reactor, was discovered in the early 1980s. Extensive research on MEB has been performed since its discovery. However, the progress of the application has been delayed because the generation mechanism of MEB remains unclear. Reasons for this lack of clarity include the complexity of the phenomenon itself and the high-speed phase change phenomenon in which boiling and condensation are rapidly generated. In addition, a more advanced thermal technique is required to realize the MEB phenomenon at the laboratory scale. To the authors’ knowledge, few studies have discussed the rush mechanism of subcooled liquid to the heating surface, which is critical to elucidating the mechanism behind MEB. This study used photographic images to verify that the cavitation phenomenon spreads to the inside of the superheated liquid on the heating surface and thus clarify the mechanism of MEB. PMID:27628271

  13. Unorthodox bubbles when boiling in cold water.

    PubMed

    Parker, Scott; Granick, Steve

    2014-01-01

    High-speed movies are taken when bubbles grow at gold surfaces heated spotwise with a near-infrared laser beam heating water below the boiling point (60-70 °C) with heating powers spanning the range from very low to so high that water fails to rewet the surface after bubbles detach. Roughly half the bubbles are conventional: They grow symmetrically through evaporation until buoyancy lifts them away. Others have unorthodox shapes and appear to contribute disproportionately to heat transfer efficiency: mushroom cloud shapes, violently explosive bubbles, and cavitation events, probably stimulated by a combination of superheating, convection, turbulence, and surface dewetting during the initial bubble growth. Moreover, bubbles often follow one another in complex sequences, often beginning with an unorthodox bubble that stirs the water, followed by several conventional bubbles. This large dataset is analyzed and discussed with emphasis on how explosive phenomena such as cavitation induce discrepancies from classical expectations about boiling.

  14. Acoustically enhanced boiling heat transfer on a heated surface containing open microchannels

    NASA Astrophysics Data System (ADS)

    Boziuk, Thomas R.; Smith, Marc K.; Glezer, Ari

    2011-11-01

    Acoustic actuation is used to enhance boiling heat transfer on a submerged heated surface containing an array of open microchannels by controlling the formation and evolution of vapor bubbles and inhibiting the instability that leads to film boiling at the critical heat flux. The effect of actuation at millimeter and micrometer scales is investigated with emphasis on the behavior of bubble nucleation, growth, contact-line motion, condensation, and detachment. The results show that microchannels control the location of boiling and reduce the mean surface superheat. In addition, acoustic actuation increases the heat flux at a given surface temperature and leads to a significant increase in the critical heat flux, a reduction of the vapor mass above the surface, and the breakup of low-frequency vapor slug formation. Supported by ONR.

  15. Design and test of a compact optics system for the pool boiling experiment

    NASA Technical Reports Server (NTRS)

    Ling, Jerri S.; Laubenthal, James R.

    1990-01-01

    The experiment described seeks to improve the understanding of the fundamental mechanisms that constitute nucleate pool boiling. The vehicle for accomplishing this is an investigation, including tests to be conducted in microgravity and coupled with appropriate analyses, of the heat transfer and vapor bubble dynamics associated with nucleation, bubble growth/collapse and subsequent motion, considering the interrelations between buoyancy, momentum and surface tension which will govern the motion of the vapor and surrounding liquid, as a function of the heating rate at the heat transfer surface and the temperature level and distribution in the bulk liquid. The experiment is designed to be contained within the confines of a Get-Away-Special Canister (GAS Can) installed in the bay of the space shuttle. When the shuttle reaches orbit, the experiment will be turned on and testing will proceed automatically. In the proposed Pool Boiling Experiment a pool of liquid, initially at a precisely defined pressure and temperature, will be subjected to a step imposed heat flux from a semitransparent thin-film heater forming part of one wall of the container such that boiling is initiated and maintained for a defined period of time at a constant pressure level. Transient measurements of the heater surface and fluid temperatures near the surface will be made, noting especially the conditions at the onset of boiling, along with motion photography of the boiling process in two simultaneous views, from beneath the heating surface and from the side. The conduct of the experiment and the data acquisition will be completely automated and self-contained. For the initial flight, a total of nine tests are proposed, with three levels of heat flux and three levels of subcooling. The design process used in the development and check-out of the compact photographic/optics system for the Pool Boiling Experiment is documented.

  16. 77 FR 73732 - In the Matter of the Amendment of the Designation of al-Qa'ida in Iraq, aka Jam'at al Tawhid wa...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-12-11

    ...-Qa'ida in Iraq, aka Jam'at al Tawhid wa'al-Jihad, aka The Monotheism and Jihad Group, aka The al-Zarqawi Network, aka al-Tawhid, aka Tanzim Qa`idat al-Jihad fi Bilad al-Rafidayn, aka The Organization of... Group of Jihad in Iraq, aka al-Qaida Group of Jihad in the Land of the Two Rivers, aka The Organization...

  17. High-resolution fluence verification for treatment plan specific QA in ion beam radiotherapy

    NASA Astrophysics Data System (ADS)

    Martišíková, Mária; Brons, Stephan; Hesse, Bernd M.; Jäkel, Oliver

    2013-03-01

    Ion beam radiotherapy exploits the finite range of ion beams and the increased dose deposition of ions toward the end of their range in material. This results in high dose conformation to the target region, which can be further increased using scanning ion beams. The standard method for patient-plan verification in ion beam therapy is ionization chamber dosimetry. The spatial resolution of this method is given by the distance between the chambers (typically 1 cm). However, steep dose gradients created by scanning ion beams call for more information and improved spatial resolution. Here we propose a clinically applicable method, supplementary to standard patient-plan verification. It is based on ion fluence measurements in the entrance region with high spatial resolution in the plane perpendicular to the beam, separately for each energy slice. In this paper the usability of the RID256 L amorphous silicon flat-panel detector for the measurements proposed is demonstrated for carbon ion beams. The detector provides sufficient spatial resolution for this kind of measurement (pixel pitch 0.8 mm). The experiments were performed at the Heidelberg Ion-Beam Therapy Center in Germany. This facility is equipped with a synchrotron capable of accelerating ions from protons up to oxygen to energies between 48 and 430 MeV u-1. Beam application is based on beam scanning technology. The measured signal corresponding to single energy slices was translated to ion fluence on a pixel-by-pixel basis, using calibration, which is dependent on energy and ion type. To quantify the agreement of the fluence distributions measured with those planned, a gamma-index criterion was used. In the patient field investigated excellent agreement was found between the two distributions. At least 95% of the slices contained more than 96% of points agreeing with our criteria. Due to the high spatial resolution, this method is especially valuable for measurements of strongly inhomogeneous fluence

  18. Boiling enriches the linear polysulfides and the hydrogen sulfide-releasing activity of garlic.

    PubMed

    Tocmo, Restituto; Wu, Yuchen; Liang, Dong; Fogliano, Vincenzo; Huang, Dejian

    2017-04-15

    Garlic is rich in polysulfides, and some of them can be H 2 S donors. This study was conducted to explore the effect of cooking on garlic's organopolysulfides and H 2 S-releasing activity. Garlic bulbs were crushed and boiled for a period ranging from 3 to 30min and the solvent extracts were analyzed by GC-MS/FID and HPLC. A cell-based assay was used to measure the H 2 S-releasing activity of the extracts. Results showed that the amounts of allyl polysulfides increased in crushed garlic boiled for 6-10min; however, prolonging the thermal treatment to 20 or 30min decreased their concentrations. Data of the H 2 S-releasing activity, expressed as diallyl trisulfide equivalents (DATS-E), parallel this trend, being significantly higher at 6 and 10min boiling. Our results showed enhancement of H 2 S-releasing activity upon moderate boiling, suggesting that shorter cooking time may maximize its health benefits as a dietary source of natural H 2 S donors. Copyright © 2016 Elsevier Ltd. All rights reserved.

  19. Investigation into flow boiling heat transfer in a minichannel with enhanced heating surface

    NASA Astrophysics Data System (ADS)

    Piasecka, Magdalena

    2012-04-01

    The paper presents results of flow boiling in a minichannel of 1.0 mm depth. The heating element for the working fluid (FC-72) that flows along the minichannel is a single-sided enhanced alloy foil made from Haynes-230. Microrecesses were formed on the selected area of the heating foil by laser technology. The observations of the flow structure were carried out through a piece of glass. Simultaneously, owing to the liquid crystal layer placed on the opposite side of the enhanced foil surface, it was possible to measure temperature distribution on the heating wall through another piece of glass. The experimental research has been focused on the transition from single phase forced convection to nucleate boiling, i.e. the zone of boiling incipience and further development of boiling. The objective of the paper is determining of the void fraction for some cross-sections of selected images for increasing heat fluxes supplied to the heating surface. The flow structure photos were processed in Corel graphics software and binarized. The analysis of phase volumes was developed in Techystem Globe software.

  20. A New Theory of Nucleate Pool Boiling in Arbitrary Gravity

    NASA Technical Reports Server (NTRS)

    Buyevich, Y. A.; Webbon, Bruce W.

    1995-01-01

    Heat transfer rates specific to nucleate pool boiling under various conditions are determined by the dynamics of vapour bubbles that are originated and grow at nucleation sites of a superheated surface. A new dynamic theory of these bubbles has been recently developed on the basis of the thermodynamics of irreversible processes. In contrast to other existing models based on empirically postulated equations for bubble growth and motion, this theory does not contain unwarrantable assumptions, and both the equations are rigorously derived within the framework of a unified approach. The conclusions of the theory are drastically different from those of the conventional models. The bubbles are shown to detach themselves under combined action of buoyancy and a surface tension force that is proven to add to buoyancy in bubble detachment, but not the other way round as is commonly presumed. The theory ensures a sound understanding of a number of so far unexplained phenomena, such as effect caused by gravity level and surface tension on the bubble growth rate and dependence of the bubble characteristics at detachment on the liquid thermophysical parameters and relevant temperature differences. The theoretical predictions are shown to be in a satisfactory qualitative and quantitative agreement with observations. When being applied to heat transfer at nucleate pool boiling, this bubble dynamic theory offers an opportunity to considerably improve the main formulae that are generally used to correlate experimental findings and to design boiling heat removal in various industrial applications. Moreover, the theory makes possible to pose and study a great deal of new problems of essential impact in practice. Two such problems are considered in detail. One problem concerns the development of a principally novel physical model for the first crisis of boiling. This model allows for evaluating critical boiling heat fluxes under various conditions, and in particular at different

  1. Technical support to the Nuclear Regulatory Commission for the boiling water reactor blowdown heat transfer program

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rice, R.E.

    Results are presented of studies conducted by Aerojet Nuclear Company (ANC) in FY 1975 to support the Nuclear Regulatory Commission (NRC) on the boiling water reactor blowdown heat transfer (BWR-BDHT) program. The support provided by ANC is that of an independent assessor of the program to ensure that the data obtained are adequate for verification of analytical models used for predicting reactor response to a postulated loss-of-coolant accident. The support included reviews of program plans, objectives, measurements, and actual data. Additional activity included analysis of experimental system performance and evaluation of the RELAP4 computer code as applied to the experiments.

  2. THE FREEZING POINT DEPRESSION OF MAMMALIAN TISSUES AFTER SUDDEN HEATING IN BOILING DISTILLED WATER

    PubMed Central

    Appelboom, Johannes W. Th.; Brodsky, William A.; Tuttle, William S.; Diamond, Israel

    1958-01-01

    The calculated freezing point depression of freshly excised boiled mammalian tissue is approximately the same as that of plasma. The boiling procedure was chosen to eliminate the influence of metabolism on the level of the freezing point depression. Problems created by the boiling, such as equilibrium between tissue and diluent, change in activity coefficient by dilution, and loss of CO2 content, are discussed. A frozen crushed tissue homogenate is hypertonic to plasma. Boiling and dilution of such hypertonic homogenate exposed to room temperature for 5 to 15 minutes did not produce significant or unexplicable decreases in its osmotic activity. Moreover, freezing and crushing of a boiled diluted tissue did not produce any increase of the isoosmotic level of freezing point depression. It is possible to explain these data either with the hypothesis of hypertonic cell fluid or with that of isotonic cell fluid. In the case of an assumed isotonic cell fluid, data can be explained with one assumption, experimentally backed. In the case of an assumed hypertonic theory data can be explained only with the help of at least three ad hoc postulates. The data support the validity of the classical concept which holds that cell fluid is isotonic to extracellular fluid. PMID:13563805

  3. Nano-inspired fluidic interactivity for boiling heat transfer: impact and criteria

    PubMed Central

    Kim, Beom Seok; Choi, Geehong; Shin, Sangwoo; Gemming, Thomas; Cho, Hyung Hee

    2016-01-01

    The enhancement of boiling heat transfer, the most powerful energy-transferring technology, will lead to milestones in the development of high-efficiency, next-generation energy systems. Perceiving nano-inspired interface functionalities from their rough morphologies, we demonstrate interface-induced liquid refreshing is essential to improve heat transfer by intrinsically avoiding Leidenfrost phenomenon. High liquid accessibility of hemi-wicking and catalytic nucleation, triggered by the morphological and hydrodynamic peculiarities of nano-inspired interfaces, contribute to the critical heat flux (CHF) and the heat transfer coefficient (HTC). Our experiments show CHF is a function of universal hydrodynamic characteristics involving interfacial liquid accessibility and HTC is improved with a higher probability of smaller nuclei with less superheat. Considering the interface-induced and bulk liquid accessibility at boiling, we discuss functionalizing the interactivity between an interface and a counteracting fluid seeking to create a novel interface, a so-called smart interface, for a breakthrough in boiling and its pragmatic application in energy systems. PMID:27708341

  4. Project Quality Assurance Plan for research and development services provided by Oak Ridge National Laboratory in support of the Westinghouse Materials Company of Ohio Operable Unit 1 Stabilization Development and Treatability Studies Program

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gilliam, T.M.

    1991-05-01

    This Project Quality Assurance Plan (PQAP) sets forth the quality assurance (QA) requirements that are applied to those elements of the Westinghouse Materials Company of Ohio (WMCO) Operable Unit 1 support at Oak Ridge National Laboratory (ORNL) project that involve research and development (R D) performed at ORNL. This is in compliance with the applicable criteria of 10 CFR Part 50, Appendix B, ANSI/ASME NQA-1, as specified by Department of Energy (DOE) Oak Ridge Operations (ORO) Order 5700.6B. For this application, NQA-1 is the core QA Program requirements document. QA policy, normally found in the requirements document, is contained herein.more » The requirements of this PQAP apply to project activities that affect the quality and reliability/credibility of research, development, and investigative data and documentation. These activities include the functions of attaining quality objectives and assuring that an appropriate QA program scope is established. The scope of activities affecting quality includes organization; personnel training and qualifications; design control; procurement; material handling and storage; operating procedures; testing, surveillance, and auditing; R D investigative activities and documentation; deficiencies; corrective actions; and QA record keeping. 12 figs.« less

  5. Confocal Microscopy and Flow Cytometry System Performance: Assessment of QA Parameters that affect data Quanitification

    EPA Science Inventory

    Flow and image cytometers can provide useful quantitative fluorescence data. We have devised QA tests to be used on both a flow cytometer and a confocal microscope to assure that the data is accurate, reproducible and precise. Flow Cytometry: We have provided two simple perform...

  6. The Gibbs Energy Basis and Construction of Boiling Point Diagrams in Binary Systems

    ERIC Educational Resources Information Center

    Smith, Norman O.

    2004-01-01

    An illustration of how excess Gibbs energies of the components in binary systems can be used to construct boiling point diagrams is given. The underlying causes of the various types of behavior of the systems in terms of intermolecular forces and the method of calculating the coexisting liquid and vapor compositions in boiling point diagrams with…

  7. The use of preservatives consist of green tea, piper betel and potassium sorbate on boiled salted fish processing

    NASA Astrophysics Data System (ADS)

    Ariyani, F.; Hermana, I.; Hidayah, I.

    2018-03-01

    The main problem in boiled salted fish ikan pindang is mucus and mold on the surface of the fish which is produced relatively fast as well as the high level of histamine content especially when scombroid fish species are used as raw material. This study was performed to evaluate the effectiveness of various preservatives to overcome such problems. Three combinations of preservatives P1 (green tea and sorbate), P3 (green tea, piper betel, sorbate), P4 (green tea and piper betel) and P0 (no preservative/control) resulted from the previous study were used in this study. Before being used, the preservatives were tested against deteriorating microorganisms commonly found in boiled salted products, of which the result showed that all microorganisms were inhibited. The preservatives were then applied at three different stages of the process of boiled salted fish, i.e. before boiling, during boiling and after boiling. Sensory attributes and microbial characteristics of the products were then evaluated. The results showed that the performance of all tested preservatives against deteriorating microorganisms was relatively similar. It was also shown that the application before and during boiling performed better.

  8. Generation of shockwave and vortex structures at the outflow of a boiling water jet

    NASA Astrophysics Data System (ADS)

    Alekseev, M. V.; Lezhnin, S. I.; Pribaturin, N. A.; Sorokin, A. L.

    2014-12-01

    Results of numerical simulation for shock waves and generation of vortex structures during unsteady outflow of boiling liquid jet are presented. The features of evolution of shock waves and vortex structures formation during unsteady outflow of boiling water are compared with corresponding structures during unsteady gas outflow.

  9. Pool boiling of nanofluids on rough and porous coated tubes: experimental and correlation

    NASA Astrophysics Data System (ADS)

    Cieśliński, Janusz T.; Kaczmarczyk, Tomasz Z.

    2014-06-01

    The paper deals with pool boiling of water-Al2O3 and water- Cu nanofluids on rough and porous coated horizontal tubes. Commercially available stainless steel tubes having 10 mm outside diameter and 0.6 mm wall thickness were used to fabricate the test heater. The tube surface was roughed with emery paper 360 or polished with abrasive compound. Aluminium porous coatings of 0.15 mm thick with porosity of about 40% were produced by plasma spraying. The experiments were conducted under different absolute operating pressures, i.e., 200, 100, and 10 kPa. Nanoparticles were tested at the concentration of 0.01, 0.1, and 1% by weight. Ultrasonic vibration was used in order to stabilize the dispersion of the nanoparticles. It was observed that independent of operating pressure and roughness of the stainless steel tubes addition of even small amount of nanoparticles augments heat transfer in comparison to boiling of distilled water. Contrary to rough tubes boiling heat transfer coefficient of tested nanofluids on porous coated tubes was lower compared to that for distilled water while boiling on porous coated tubes. A correlation equation for prediction of the average heat transfer coefficient during boiling of nanofluids on smooth, rough and porous coated tubes is proposed. The correlation includes all tested variables in dimensionless form and is valid for low heat flux, i.e., below 100 kW/m2.

  10. Nutrition content of brisket point end of part Simental Ongole Crossbred meat in boiled various temperature

    NASA Astrophysics Data System (ADS)

    Riyanto, J.; Sudibya; Cahyadi, M.; Aji, A. P.

    2018-01-01

    This aim of this study was to determine the quality of nutritional contents of beef brisket point end of Simental Ongole Crossbred meat in various boiling temperatures. Simental Ongole Crossbred had been fattened for 9 months. Furthermore, they were slaughtered at slaughterhouse and brisket point end part of meat had been prepared to analyse its nutritional contents using Food Scan. These samples were then boiled at 100°C for 0 (TR), 15 (R15), and 30 (R30) minutes, respectively. The data was analysed using Randomized Complete Design (CRD) and Duncan’s multiple range test (DMRT) had been conducted to differentiate among three treatments. The results showed that boiling temperatures significantly affected moisture, and cholesterol contents of beef (P<0.05) while fat content was not significantly affected by boiling temperatures. The boiling temperature decreased beef water contents from 72.77 to 70.84%, on the other hand, the treatment increased beef protein and cholesterol contents from 20.77 to 25.14% and 47.55 to 50.45 mg/100g samples, respectively. The conclusion of this study was boiling of beef at 100°C for 15 minutes and 30 minutes decreasing water content and increasing protein and cholesterol contents of brisket point end of Simental Ongole Crossbred beef.

  11. Predictors of Drinking Water Boiling and Bottled Water Consumption in Rural China: A Hierarchical Modeling Approach.

    PubMed

    Cohen, Alasdair; Zhang, Qi; Luo, Qing; Tao, Yong; Colford, John M; Ray, Isha

    2017-06-20

    Approximately two billion people drink unsafe water. Boiling is the most commonly used household water treatment (HWT) method globally and in China. HWT can make water safer, but sustained adoption is rare and bottled water consumption is growing. To successfully promote HWT, an understanding of associated socioeconomic factors is critical. We collected survey data and water samples from 450 rural households in Guangxi Province, China. Covariates were grouped into blocks to hierarchically construct modified Poisson models and estimate risk ratios (RR) associated with boiling methods, bottled water, and untreated water. Female-headed households were most likely to boil (RR = 1.36, p < 0.01), and among boilers those using electric kettles rather than pots had higher income proxies (e.g., per capita TV ownership RR = 1.42, p < 0.01). Higher-income households with younger, literate, and male heads were more likely to purchase (frequently contaminated) bottled water, or use electric kettles if they boiled. Our findings show that boiling is not an undifferentiated practice, but one with different methods of varying effectiveness, environmental impact, and adoption across socioeconomic strata. Our results can inform programs to promote safer and more efficient boiling using electric kettles, and suggest that if rural China's economy continues to grow then bottled water use will increase.

  12. Technical Note: Development and performance of a software tool for quality assurance of online replanning with a conventional Linac or MR-Linac.

    PubMed

    Chen, Guang-Pei; Ahunbay, Ergun; Li, X Allen

    2016-04-01

    To develop an integrated quality assurance (QA) software tool for online replanning capable of efficiently and automatically checking radiation treatment (RT) planning parameters and gross plan quality, verifying treatment plan data transfer from treatment planning system (TPS) to record and verify (R&V) system, performing a secondary monitor unit (MU) calculation with or without a presence of a magnetic field from MR-Linac, and validating the delivery record consistency with the plan. The software tool, named ArtQA, was developed to obtain and compare plan and treatment parameters from both the TPS and the R&V system database. The TPS data are accessed via direct file reading and the R&V data are retrieved via open database connectivity and structured query language. Plan quality is evaluated with both the logical consistency of planning parameters and the achieved dose-volume histograms. Beams in between the TPS and R&V system are matched based on geometry configurations. To consider the effect of a 1.5 T transverse magnetic field from MR-Linac in the secondary MU calculation, a method based on modified Clarkson integration algorithm was developed and tested for a series of clinical situations. ArtQA has been used in their clinic and can quickly detect inconsistencies and deviations in the entire RT planning process. With the use of the ArtQA tool, the efficiency for plan check including plan quality, data transfer, and delivery check can be improved by at least 60%. The newly developed independent MU calculation tool for MR-Linac reduces the difference between the plan and calculated MUs by 10%. The software tool ArtQA can be used to perform a comprehensive QA check from planning to delivery with conventional Linac or MR-Linac and is an essential tool for online replanning where the QA check needs to be performed rapidly.

  13. Study on Fins' Effect of Boiling Flow in Millimeter Channel Heat Exchanger

    NASA Astrophysics Data System (ADS)

    Watanabe, Satoshi

    2005-11-01

    Recently, a lot of researches about compact heat exchangers with mini-channels have been carried out with the hope of obtaining a high-efficiency heat transfer, due to the higher ratio of surface area than existing heat exchangers. However, there are many uncertain phenomena in fields such as boiling flow in mini-channels. Thus, in order to understand the boiling flow in mini-channels to design high-efficiency heat exchangers, this work focused on the visualization measurement of boiling flow in a millimeter channel. A transparent acrylic channel (heat exchanger form), high-speed camera (2000 fps at 1024 x 1024 pixels), and halogen lamp (backup light) were used as the visualization system. The channel's depth is 2 mm, width is 30 mm, and length is 400 mm. In preparation for commercial use, two types of channels were experimented on: a fins type and a normal slit type (without fins). The fins are circular cylindrical obstacles (diameter is 5 mm) to promote heat transfer, set in a triangular array (distance between each center point is 10 mm). Especially in this work, boiling flow and heat transfer promotion in the millimeter channel heat exchanger with fins was evaluated using a high-speed camera.

  14. SU-E-T-268: Proton Radiosurgery End-To-End Testing Using Lucy 3D QA Phantom

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Choi, D; Gordon, I; Ghebremedhin, A

    2014-06-01

    Purpose: To check the overall accuracy of proton radiosurgery treatment delivery using ready-made circular collimator inserts and fixed thickness compensating boluses. Methods: Lucy 3D QA phantom (Standard Imaging Inc. WI, USA) inserted with GaFchromicTM film was irradiated with laterally scattered and longitudinally spread-out 126.8 MeV proton beams. The tests followed every step in the proton radiosurgery treatment delivery process: CT scan (GE Lightspeed VCT), target contouring, treatment planning (Odyssey 5.0, Optivus, CA), portal calibration, target localization using robotic couch with image guidance and dose delivery at planned gantry angles. A 2 cm diameter collimator insert in a 4 cm diametermore » radiosurgery cone and a 1.2 cm thick compensating flat bolus were used for all beams. Film dosimetry (RIT114 v5.0, Radiological Imaging Technology, CO, USA) was used to evaluate the accuracy of target localization and relative dose distributions compared to those calculated by the treatment planning system. Results: The localization accuracy was estimated by analyzing the GaFchromic films irradiated at gantry 0, 90 and 270 degrees. We observed 0.5 mm shift in lateral direction (patient left), ±0.9 mm shift in AP direction and ±1.0 mm shift in vertical direction (gantry dependent). The isodose overlays showed good agreement (<2mm, 50% isodose lines) between measured and calculated doses. Conclusion: Localization accuracy depends on gantry sag, CT resolution and distortion, DRRs from treatment planning computer, localization accuracy of image guidance system, fabrication of ready-made aperture and cone housing. The total deviation from the isocenter was 1.4 mm. Dose distribution uncertainty comes from distal end error due to bolus and CT density, in addition to localization error. The planned dose distribution was well matched (>90%) to the measured values 2%/2mm criteria. Our test showed the robustness of our proton radiosurgery treatment delivery system

  15. Anti-obesity effects of boiled tuna extract in mice with obesity induced by a high-fat diet.

    PubMed

    Kim, Youngmin; Kwon, Mi-Jin; Choi, Jeong-Wook; Lee, Min-Kyeong; Kim, Chorong; Jung, Jaehun; Aprianita, Heny; Nam, Heesop; Nam, Taek-Jeong

    2016-10-01

    The aim of this study was to examine the anti-obesity effects of boiled tuna extract in C57BL/6N mice with obesity induced by a high-fat diet (HFD). We determined the anti-obesity effects of boiled tuna extract (100, 200, or 400 mg/kg) on the progression of HFD-induced obesity for 10 weeks. The mice were divided into 5 groups as follows: the normal diet (ND) group (n=10); the HFD group (n=10); the mice fed HFD and 100 mg/kg boiled tuna extract group (n=10); those fed a HFD and 200 mg/kg boiled tuna extract group (n=10); and those fed a HFD and 400 mg/kg boiled tuna extract group (n=10). Changes in body weight, fat content, serum lipid levels and lipogenic enzyme levels were measured. The consumption of boiled tuna extract lowered epididymal tissue weight and exerted anti-obesity effects, as reflected by the serum glucose, triglyceride (TG), total cholesterol (TC), high-density lipoprotein cholesterol (HDL‑C), low-density lipoprotein cholesterol (LDL-C), insulin and leptin levels. In addition, we demonstrated changes in liver adipogenic- and lipogenic-related protein expression by western blot analysis. Boiled tuna extract downregulated the levels of the CCAAT/enhancer-binding protein α, β and δ (C/EBPα, β, δ), and peroxisome proliferator-activated receptor-γ (PPAR-γ) adipocyte marker genes. Boiled tuna extract also attenuated adipogenic and lipogenic gene expression, namely the levels of fatty acid synthase (FAS), lipoprotein lipase (LPL), acetyl-CoA carboxylase (ACC), glucose transporter type 4 (Glut4) and phosphorylated adenosine monophosphate-activated protein kinase α and β (AMPKα, β) in a dose-dependent manner. Moreover, the consumption of boiled tuna extract restored the levels of superoxide dismutase (SOD), catalase (CAT), glutamic oxaloacetic transaminase (GOT), glutamic-pyruvate transaminase (GPT), aspartate transaminase (AST) and alanine transaminase (ALT) to those of the control group. These results

  16. Raw milk from vending machines: Effects of boiling, microwave treatment, and refrigeration on microbiological quality.

    PubMed

    Tremonte, Patrizio; Tipaldi, Luca; Succi, Mariantonietta; Pannella, Gianfranco; Falasca, Luisa; Capilongo, Valeria; Coppola, Raffaele; Sorrentino, Elena

    2014-01-01

    In Italy, the sale of raw milk from vending machines has been allowed since 2004. Boiling treatment before its use is mandatory for the consumer, because the raw milk could be an important source of foodborne pathogens. This study fits into this context with the aim to evaluate the microbiological quality of 30 raw milk samples periodically collected (March 2013 to July 2013) from 3 vending machines located in Molise, a region of southern Italy. Milk samples were stored for 72 h at 4 °C and then subjected to different treatments, such as boiling and microwaving, to simulate domestic handling. The results show that all the raw milk samples examined immediately after their collection were affected by high microbial loads, with values very close to or even greater than those acceptable by Italian law. The microbial populations increased during refrigeration, reaching after 72 h values of about 8.0 log cfu/mL for Pseudomonas spp., 6.5 log cfu/mL for yeasts, and up to 4.0 log cfu/mL for Enterobacteriaceae. Boiling treatment, applied after 72 h to refrigerated milk samples, caused complete decontamination, but negatively affected the nutritional quality of the milk, as demonstrated by a drastic reduction of whey proteins. The microwave treatment at 900 W for 75 s produced microbiological decontamination similar to that of boiling, preserving the content in whey proteins of milk. The microbiological characteristics of raw milk observed in this study fully justify the obligation to boil the raw milk from vending machines before consumption. However, this study also showed that domestic boiling causes a drastic reduction in the nutritional value of milk. Microwave treatment could represent a good alternative to boiling, on the condition that the process variables are standardized for safe domestic application. Copyright © 2014 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.

  17. MO-F-CAMPUS-T-02: An Electronic Whiteboard Platform to Manage Treatment Planning Process

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    DiCostanzo, D; Woollard, J; Gupta, N

    2015-06-15

    Purpose: In an effort to improve patient safety and streamline the radiotherapy treatment planning (TP) process, a software based whiteboard had been developed and put in use in our facility Methods: The electronic whiteboard developed using SQL database (DB) and PHP/JavaScript based web interface, is published via department intranet and login credentials. The DB stores data for each TP process such as patient information, plan type, simulation/start dates, physician, dosimetrist, QA and the current status in planning process. Users interact with the DB per plan and perform status updates in real time as the planning process progresses. All user interactionsmore » with the DB are recorded with timestamps so as to calculate statistical information for TP process management such as contouring times, planning and review times, dosimetry, physics and therapist QA times. External beam and brachytherapy plans are categorized according to complexity (ex: IMRT, 3D, HDR, LDR etc) and treatment types and applicators. Each plan category is assigned specific timelines for each planning process. When a plan approaches or passes the predetermined timeline, users are alerted via color coded graphical cues. When certain process items are not completed in time, pre-determined actions are triggered such as a delay in treatment start date. Results: Our institution has been using the electronic whiteboard for two years. Implementation of pre-determined actions based on the statistical information collected by the whiteboard improved our TP process. For example, the average time for normal tissue contouring decreased from 0.73±1.37 to 0.24±0.33 days. The average time for target volume contouring decreased from 3.2±2.84 to 2.37±2.54 days. This increase in efficiency allows more time for quality assurance processes, improving patient safety. Conclusion: The electronic whiteboard has been an invaluable tool for streamlining our TP processes. It facilitates timely and accurate

  18. Boiling point measurement of a small amount of brake fluid by thermocouple and its application.

    PubMed

    Mogami, Kazunari

    2002-09-01

    This study describes a new method for measuring the boiling point of a small amount of brake fluid using a thermocouple and a pear shaped flask. The boiling point of brake fluid was directly measured with an accuracy that was within approximately 3 C of that determined by the Japanese Industrial Standards method, even though the sample volume was only a few milliliters. The method was applied to measure the boiling points of brake fluid samples from automobiles. It was clear that the boiling points of brake fluid from some automobiles dropped to approximately 140 C from about 230 C, and that one of the samples from the wheel cylinder was approximately 45 C lower than brake fluid from the reserve tank. It is essential to take samples from the wheel cylinder, as this is most easily subjected to heating.

  19. Preliminary design and hazards report. Boiling Reactor Experiment V (BORAX V)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rice, R. E.

    1960-02-01

    The preliminary objectives of the proposed BORAX V program are to test nuclear superheating concepts and to advance the technology of boiling-water-reactor design by performing experiments which will improve the understanding of factors limiting the stability of boiling reactors at high power densities. The reactor vessel is a cylinder with ellipsoidal heads, made of carbon steel clad internally with stainless steel. Each of the three cores is 24 in. high and has an effective diameter of 39 in. This is a preliminary report. (W.D.M.)

  20. Heterogeneous boiling-up of superheated liquid at achievable superheat threshold.

    PubMed

    Ermakov, G V; Lipnyagov, E V; Perminov, S A; Gurashkin, A L

    2009-07-21

    The classical theory of homogeneous nucleation describes well the superheat threshold observed in experiments. It may be assumed therefore that homogeneous boiling-up of a liquid takes place in experiments, and the theory has been verified experimentally well. The streak photography used in this study showed that boiling-up of a superheated liquid at the threshold of the achievable superheat occurs at a limited number of surface fluctuation centers in a vessel, rather than in the bulk as one would expect with homogeneous nucleation. Thus, the homogeneous theory, which rather accurately describes the heterogeneous threshold of the achievable superheat, obviously is not confirmed in experiments.