1 JOURNAL OF APPLIED CLINICAL MEDICAL PHYSICS, VOLUME 16, NUMBER 4, 2015 Quality control of CT systems by automated monitoring of key performance indicators: a two-year study Patrik Nowik, a Robert Bujila, Gavin Poludniowski, Annette Fransson Department of Medical Physics, Karolinska University Hospital, Stockholm, Sweden Received 01 December, 2014; accepted 20 March, 2015 The purpose of this study was to develop a method of performing routine periodical quality controls (QC) of CT systems by automatically analyzing key performance indicators (KPIs), obtainable from images of manufacturers quality assurance (QA) phantoms. A KPI pertains to a measurable or determinable QC parameter that is influenced by other underlying fundamental QC parameters. The established KPIs are based on relationships between existing QC parameters used in the annual testing program of CT scanners at the Karolinska University Hospital in Stockholm, Sweden. The KPIs include positioning, image noise, uniformity, homogeneity, the CT number of water, and the CT number of air. An application (MonitorCT) was developed to automatically evaluate phantom images in terms of the established KPIs. The developed methodology has been used for two years in clinical routine, where CT technologists perform daily scans of the manufacturer s QA phantom and automatically send the images to MonitorCT for KPI evaluation. In the cases where results were out of tolerance, actions could be initiated in less than 10 min. 900 QC scans from two CT scanners have been collected and analyzed over the two-year period that MonitorCT has been active. Two types of errors have been registered in this period: a ring artifact was discovered with the image noise test, and a calibration error was detected multiple times with the CT number test. In both cases, results were outside the tolerances defined for MonitorCT, as well as by the vendor. Automated monitoring of KPIs is a powerful tool that can be used to supplement established QC methodologies. Medical physicists and other professionals concerned with the performance of a CT system will, using such methods, have access to comprehensive data on the current and historical (trend) status of the system such that swift actions can be taken in order to ensure the quality of the CT examinations, patient safety, and minimal disruption of service PACS numbers: C-, N-, Q- Key words: computed tomography, quality assurance, quality control, key performance indicators, image analysis I. INTRODUCTION The number of CT examinations performed worldwide and the high doses associated with those CT examinations makes up-to-date information on the performance of CT systems a prerequisite to ensure patient safety and the quality of examinations. The total time required and the limited access to the CT scanners often restrict the possibilities for the medical physicist to perform quality control (QC) at more than the recommended annual frequency. (1) A CT scanner is likely to undergo preventive maintenance, hardware and/or software upgrades, and repairs between annual QC. In addition, degrading and/or malfunctioning of the hardware can occur a Corresponding author: Patrik Nowik, Department of Medical Physics, Section of Imaging Physics, Karolinska University Hospital, SE Stockholm, Sweden; phone: +46 (0) ; fax: +46 (0) ;
2 255 Nowik et al.: QC of CT by monitoring of KPI 255 and might go undetected until the next maintenance or annual routine QC. Major performance abnormalities are likely to be detected by technologists during routine work, but more subtle issues may go unnoticed. One of the main responsibilities of a medical physicist is to establish, implement, and supervise a quality assurance (QA) program. (2) Today, most QC procedures, apart from being performed only annually, contain subjective elements (e.g., the manual positioning of ROIs). This reduces the reproducibility of consecutive QC tests. Larger uncertainties and low frequency of testing reduce statistical power in trend analysis, rendering it less valuable. A modern routine QC should ideally have fully automated evaluation routines and store the results in a database. (3) The results will then be at their most reliable, reproducible, and useful for longterm trend analysis. More frequent QC of CT systems in order to ensure scanner performance and acceptable image quality has been suggested by the American College of Radiology (ACR). ACR also state that automated QC procedures may be used if they are reviewed correctly. In its 2012 CT QC report, ACR recommends daily QC, including monitoring the CT number of water, the image noise, and image artifacts. (1) IEC, instead, recommends that the noise, CT number of water, and image uniformity should be evaluated monthly. (4) Many sites around the world already have daily QC program where CT technologists perform a limited number of tests. When automatic evaluation procedures are lacking, however, recommendations such as those of the ACR can be difficult to implement in clinical practice. The aim of this work was to develop a methodology and software tools to enable a comprehensive raft of QC tests to be carried out rapidly on a daily basis with minimal operator intervention. More than a simple logging of QC results, the system provides automated alerts to physics staff and the facility for trend analysis. The advantages of such an automated QA system is that performance issues may be identified earlier without unduly increasing demands on CT technologists, and it also allows physicists to implement more sophisticated analysis of images. It also provides insight into the behavior of CT scanners on a day-to-day basis and assists in identifying trends in slow performance degradation. While the burden for a CT technologist between a nonautomated and an automated QA system is not necessarily large, the results from an automated QA system are more valuable, since operator variability is removed. Further, in addition to a basic set of tests, physicists have the option of including more sophisticated tests without an increase in burden to the technologist. We note that some manufacturers have their own automated QC routines incorporated in their CT systems. The results from the QC performed with such routines are not typically easily accessed and it is, therefore, hard to perform trend analysis on the data. In this study we sought to develop analyses that are vendor-independent. An important part of this work was to identify key performance indicators (KPIs) in CT. A KPI pertains to a measurable or determinable parameter that is influenced by other underlying more fundamental parameters affecting scanner performance. While the established QA program with annual QC consists of an extensive set of independent parameter tests, a QC of KPIs alleviates the intricacies involved with performing such full-blown QC. The KPI method offers a simplified routine and makes periodic QC accessible while maintaining insight into most parameters that are tested annually. Furthermore, when a KPI deviates, the deviation can be systematically located by following the KPIs dependencies. Relevant (independent) tests can then be performed to isolate the real origin of the problem. By introducing such a simplified yet comprehensive test, the performance of CTs can be sampled at tighter intervals and complement the annual QC test. Once the KPIs were identified in this work, the methodology was implemented as a daily QC program on two of the CT scanners at the Karolinska University Hospital in Stockholm, Sweden, using the manufacturer s QA phantom. Results from this daily QC program during a two-year period are presented.
3 256 Nowik et al.: QC of CT by monitoring of KPI 256 II. MATERIALS AND METHODS A. Key performance indicators (KPIs) A KPI pertains to a measurable or determinable parameter that is influenced by other underlying parameters. A stable KPI implies that the underlying parameters are stable as well. To perform daily QC procedures that cover as many parameters as possible from routine annual QC, KPIs were identified and verified by studying and mapping out the relationship between established QC parameters used in annual testing. The KPIs include positioning, image noise, uniformity, homogeneity, CT numbers of water, and CT number of air. The KPI positioning needs a dedicated positioning part in a manufacturer s phantom, while the others are determined from images of the water part. The available QC procedures of various parameters in a QA program, (5-9) and the KPIs main dependencies on QC parameters are illustrated in Table 1. Table 1. CT parameters that generally are available with a CT QA program, how often they should be tested, and on which KPIs they mainly depend. Parameters Tested During QA QA Parameter Dependencies (A = acceptance, C = constancy) on KPIs LASER accuracy (A, C) Table top indexing (A) Table top orientation (A) Scan plane localization (A, C) CTDI vol (A) CTDI air (A, C) Dose-noise response (A) Dose-noise linearity (A) X-ray tube voltage (A, C) HVL (A, C) Z-dose profile (A) Geometric efficiency (A) CT number of air (A, C) CT number of water (A, C) CT numbers of various materials (A, C) CT number linearity (A, C) Slice thickness (A, C) Image noise (A, C) Direct Tests Positioning CT Numbers Uniformity (A, C) Inter-slice noise (A, C) Geometric accuracy (A) Spatial resolution (A, C) Scattered radiation (A) System documentation (A, C) Radiation protection documentation (A, C) Radiation shielding (A, C) Radiation notification (A, C) Staff interview (C) Artifact evaluation (A,C) Indirect tests Shaped filter Calibration Dose display (A,C) Table top indexing Table top orientation Scan plane localization Image noise X-ray tube voltage CTDI vol CTDI air Spatial resolution HVL Low contrast resolution Collimation Slice thickness Geometric efficiency Z-dose profile Dose-noise response Detector response Inter-slice noise CT number of water CT number of air Calibration Uniformity Shaped filter Reconstruction Detector Homogeneity Artifact evaluation A.1 Positioning The primary goal with the positioning test is to assess the CT scanner s ability to correctly position the object being imaged. From a clinical view point this is important as it has been shown that patient miscentering may lead to both higher surface doses and noisier images. (10) Positioning depends on the precision of the gantry lasers and the functionality of the tabletop of the CT scanner. Positioning that is out of tolerance may have the consequence that other KPI measurements cannot be fully trusted since they are calculated with the phantom miscentered.
4 257 Nowik et al.: QC of CT by monitoring of KPI 257 The KPIs for positioning are the values of the X, Y, and Z offsets of the phantom. The manufacturer s phantom usually includes a positioning part where these parameters can be calculated. The tolerances for the positioning lasers for the X, Y, and Z offset were chosen as ± 2 mm. This was in accordance with the tolerances defined by the manufacturer and is also what the Nordic Association of Clinical Physics (NACP) and ACR recommend. (1,11) A.2 Image noise The image noise KPI is influenced by all parameters that can affect the pixel values in an image, parameters related to the X-ray tube output and the reconstruction. The Brooks formula (12) (reformulated by Nagel (13) ) shows that image noise (σ) varies with the attenuation of the object, radiation dose, pixel size, beam collimation, and slice thickness. It can be expressed as (1) where μ is the linear attenuation coefficient, d is the thickness of the object with attenuation μ, is the attenuation of the object, D is the CT dose index CTDI VOL, b 2 is the sampling distance at the rotation center (pixel size), w is the beam collimation, and T is the slice thickness. The image noise dependence on dose can be expanded by taking into account that the dose varies linearly with tube current and with exposure time. In addition, image noise dependence on tube peak voltage can be introduced since the radiation dose varies as where kvp is the peak voltage and n is a kilovoltage dependency factor that has been estimated to have a value of approximately 2.6. (14) The value n depends on the radiation quality of the beam, tube output, and detector response. The Brooks formula can be rewritten by inserting the tube voltage (kvp), tube current (ma), and exposure time (s) dependences on dose, yielding (2) (3) A base value of the KPI image noise is defined at the acceptance testing of the CT scanner and is used as a comparison in future tests. The tolerance for the variation of the image noise that is applied at the annual quality controls of the CT scanners at the Karolinska University Hospital is set to 10% of this base value. This tolerance is the same as recommended by the NACP and IEC. (4,11) The automated image analysis used in this work allows, however, for a constant and reproducible ROI diameter of 40% of the phantom (6) diameter at every test. Initial tests showed that the automatization of ROI definition allowed the use of a lower tolerance. A reduction to 5% of the base value for the image noise was tested and implemented in this work. As a lowered tolerance for the variation in image noise translates into less variation possibilities for the underlying parameters, the test will pick up bad machine performance earlier than what otherwise would be possible. The effect on this KPI when an underlying parameter is out of tolerance can be evaluated directly from the Brooks formula (4)
5 258 Nowik et al.: QC of CT by monitoring of KPI 258 The relation shows how the image noise will change depending on variations in dose. In order to use this expression to monitor changes of underlying parameters by measuring the noise, it is of importance to use the same scan parameters at every QC procedure. According to the IEC (4) if both CT numbers and image noise are within specified tolerances, the specified low-contrast resolution is also expected to be within tolerance and, for that reason, specific tests for low-contrast resolution are not performed with this method. The magnitude of image noise is sensitive to small variations in the spatial frequency distribution of the MTF in the reconstruction of images. (15,16) Therefore, if a specific reconstruction algorithm produced images with slight changes in its MTF, it could possibly be discovered with the KPI image noise. A.3 CT number of water and air According to the definition of the Hounsfield unit (HU), air (approximately vacuum) has a CT number of HU and water has a CT number of 0 HU. These two values provide two calibration points that affect the CT number of all other materials. These KPIs are mainly influenced by the scanner s reconstruction algorithm and its calibration. Although they are also influenced by the X-ray spectra and the object attenuation, the impact of those aspects are covered by other KPIs. Given that the manufacturer s phantom is used both for the CT scanner calibration and the QC tests, changes in the air and water CT number KPIs are expected to directly reflect any change in CT scanner calibration The automated measurements of the CT number is performed with a ROI diameter that is 10% of the phantom diameter. (6) The CT number tolerance used during annual quality controls at the hospital is ± 4 HU for water and ± 10 HU for air. These are the same values as those recommended by the NACP and IEC. (4,11) As in the case with image noise discussed in the previous section, the automated definition of the ROI in this work resulted in a very stable estimate of the CT number of water and a reduction of the tolerance was possible. The tolerance for CT numbers of water was empirically chosen to be ± 2 HU. The same reduction could not be applied for air since the ROI for air were put outside the phantom. This is contrary to the recommendation of the IEC, and resulted in increased variability with respect to water. Hence, the tolerance for CT numbers of air was kept at ± 10 HU. The chosen tolerances can be compared to the tolerances suggested by the vendor, which are ± 3 HU for water and ± 10 HU for air. A.4 Uniformity The uniformity KPI is a test that evaluates the CT scanner performance in reconstructing a uniform image. The uniformity of the image depends on the shaped beam filter, X-ray tube output, and on the centering of the object in the beam. The test focuses on the capability of the CT scanner to produce an image free from beam hardening and artifacts. The placement of ROIs in this test includes one ROI that is positioned in the center of the phantom and four ROIs that are placed in the periphery of the phantom and located at 3, 6, 9, and 12 o clock. The peripheral ROIs are placed 1 cm from the edge. (6) In order to exclude analyzing pixels in which the CT number might have an appreciable influence from the PMMA housing of the phantom, the center of the peripheral ROIs is located at a distance corresponding to 15% of the phantom diameter away from the edge of the phantom. The ROIs have a diameter that corresponds to 10% of the phantom diameter. (11) Following the recommendations of the IEC, image uniformity (an indicator of the existence of beam hardening effects) was evaluated as the maximum difference between the mean value of the center ROI and any of the four peripheral ROIs. (6) The tolerance for the variation of image uniformity used in the annual quality controls at the Hospital is 4 HU according to NACP recommendations. (11) In this work, a tolerance of 3 HU was used (as suggested by the manufacturer).
6 259 Nowik et al.: QC of CT by monitoring of KPI 259 A.5 Homogeneity The KPI homogeneity was introduced in order to increase the likelihood of identifying artifacts associated with degradations of CT scanner performance (e.g., ring artifacts and local nonuniformities caused by air bubbles in the cooling oil). Contiguous rectangular pixel ROIs are used to cover a circular area with a diameter of 85% of the diameter of the homogenous water phantom section of the manufacturer s QA phantom. An illustration of the ROI placement is shown in Fig. 1. The size of the area covering the phantom was chosen in order to exclude analyzing pixels in which the CT number might have an appreciable influence from the PMMA housing. This does mean, however, that artifacts may go undetected if they arise in the area outside the covered area. The size of the individual ROIs was empirically determined with the aim to provide a good tradeoff between being too large (artifacts can go undetected) and being too small (false positives can be found due to the statistical nature of CT numbers). The KPI homogeneity is defined as the maximum difference in mean CT number between any of the ROIs. This performance test for a CT scanner is not included in any standard QA procedure and has, to our knowledge, not been suggested before. As the homogeneity test is an extension of the concept of the uniformity test, it has the potential to replace the latter. Currently we have chosen to keep the uniformity KPI as it is recommended by the IEC to perform a uniformity test at least monthly. (4) The homogeneity tolerance level is still undergoing refinement. The approach for choosing the tolerance was to select a first tolerance level that from initial tests seem reasonable and then, as more data are collected, systematically lower the tolerance. The opposite approach, to start with a low tolerance and systematically increase the tolerance could have been chosen, but would have generated more false positives. The first tolerance level that was used was 6 HU (larger than the tolerance of uniformity) and following the first year of tests, the tolerance was lowered to 5 HU. Fig. 1. Placement of ROIs of the KPI homogeneity test. B. Methodology of routine QC by monitoring of KPI A method of performing routine QC procedures of CT scanners in terms of the identified KPIs was developed. The method relies on imaging the manufacturer s QA phantom. Two imaging protocols were defined. The first protocol (Level 0) was used as a standard test of the KPIs. The second protocol (Level 1) was set up to systematically search for the underlying fundamental QC parameter that caused the KPI to deviate out of tolerance. The Level 0 test includes a single slice over the reference markings of the manufacturer s QA phantom, together with one scan
7 260 Nowik et al.: QC of CT by monitoring of KPI 260 (having multiple slices) over the homogenous water section of the phantom. If the Level 0 passes, the CT scanner fulfills its performance requirements for clinical use. If a Level 0 test fails, the more comprehensive Level 1 test is required. In the same way as for the Level 0 test, the Level 1 test is performed by scanning the manufacturer s QA phantom. This test includes imaging of the same single slice as in Level 0 (used to evaluate the KPI for positioning), together with a series of multiple scans over the homogenous water section of the manufacturer s QA phantom. For each scan in the Level 1 test, imaging parameters that the different KPIs depend on are varied systematically in order to locate which of the fundamental QC parameters caused the KPI(s) to deviate from tolerance. It is the medical physicist s task to analyze why a QC has failed, with the combined data from the Level 0 and Level 1 tests. Following this analysis, the medical physicist has a good starting point for performing any additional and more focused tests (with, for example, dedicated phantoms or other instruments). Once the fundamental QC parameter causing the problems has been verified, efforts to service the CT scanner promptly can be initiated. The proposed methodology of routine QC by monitoring KPIs is illustrated in Fig. 2. In order to make an efficient implementation of the procedure, a software application, MonitorCT, was developed in-house in order to receive the set of KPI scans and to automatically analyze the images. The results from the analyses are stored in a database and tested against the defined tolerance levels. The imaging protocols have been configured to transfer the acquired images automatically to the software. As the procedure now is almost entirely automated, the time needed for the tests to be performed and evaluated has greatly decreased. For the CT technologist, the testing procedure consists of: (i) mounting the manufacturer s QA phantom, (ii) positioning the phantom according to the CT positioning laser and the phantom markings, and (iii) selecting the predefined protocol for the desired test (Level 0 or Level 1). The time required by the CT technologist to perform a test, from mounting of the phantom to a report of the results, is approximately 2 min for the Level 0 test and approximately 3 min for the Level 1 test. Fig. 2. Schematic illustration of the proposed routine QC methodology.
8 261 Nowik et al.: QC of CT by monitoring of KPI 261 C. Pilot study The method for performing routine QC proposed in the Material & Methods section B using the set of KPIs derived in section A was tested on two CT scanners at the Hospital over a period of two years. The CT scanners included a GE LightSpeed VCT (GE Healthcare, Waukesha, WI) and a GE Discovery CT750HD (GE Healthcare, Waukesha, WI). Both scanners are installed at the Department of Neuroradiology at the Karolinska University Hospital. The GE LightSpeed VCT and the GE Discovery CT750HD are subsequently referred to as CT1 and CT2, respectively, in this work. Due to the unprecedented nature of the methodology, we chose to perform the tests on a daily basis where a Level 0 test was performed Monday to Thursday and a Level 1 test was performed every Friday (and in case a Level 0 test failed). Following the standard daily air calibration, the CT technologists were instructed to perform daily QC scans according to the workflow proposed above. They had the responsibility to perform the Level 0 test and review the results, perform the Level 1 test if the Level 0 test failed, and to inform the medical physicist if a Level 1 test was performed. The imaging protocols that were used for the Level 0 and Level 1 tests are presented in Table 2. Initially, the medical physicist reviewed the QC reports from MonitorCT, sent via mail every morning, and contacted the head CT technologist if a Level 0 test failed with the instruction to perform a Level 1 test. After the first year, a Web interface was developed to display the results from the evaluations with MonitorCT whereby the technologists could review the result of the test and perform the Level 1 test if necessary. The Web interface provides current scanner status and support for trend analysis of the KPIs. Table 2. Scan parameters for the Level 0 and Level 1 tests on the connected CT scanners during the pilot study. Scanned No. of Rot time FOV Im Thk Coll Conv. Level Part Slices (kvp) SFOV (ma) (s) (cm) (mm) (mm) Kernel 1 Positioning Head std Water Head std 1 Positioning Head std Water 1 80 Head std Head std Head std Head std Head std Head std Head std Head std Head std Head std Medium std Body std Head std Head edge Head std III. RESULTS & DISCUSSION In the Material & Methods section A of this work, relevant KPIs for CT system performance were identified and their relation to existing fundamental QC parameters used during annual QC was established. A method for incorporating KPIs into periodical routine QC based on the identified KPIs was presented in section B. In section C, the implementation of the methodology for testing on a daily basis in a clinical setting was elaborated. The pilot study was carried out for a period of over two years at the Department of Neuroradiology at the Karolinska University Hospital and included more than 900 daily QC scans from the two CT scanners.
9 262 Nowik et al.: QC of CT by monitoring of KPI 262 The KPIs from the Level 0 tests for one of the CT scanners are shown in Fig. 3. Figure 3(a), the positioning of the phantom center is presented, revealing a systematic error of 3 4 mm in its position. Following these results, the phantom positioning was investigated. The positioning lasers were tested independently with a Catphan 600 (The Phantom Laboratory, Salem, NY) (a) (b) (c) (d) (e) (f) Fig. 3. KPIs from one CT scanner s Level 0 tests: (a) distribution of the phantom center (tolerances are for positioning laser accuracy); (b) image noise in one slice; (c) boxplot of CT number of water for all slices (all data points are inside the whiskers); (d) CT number of water in one slice; (e) uniformity in one slice; (f) homogeneity in one slice.
10 263 Nowik et al.: QC of CT by monitoring of KPI 263 and found to be within tolerance. It was concluded that the technologists were handling the QA phantom appropriately when positioning with the lasers. However, the deviations in positioning were found to be caused by problems mounting the QA phantom in a precise manner, due to the instability of the phantom bracket. This result shows that the manufacturer s QA phantom offers limited precision in the test of the positioning KPI. It may be of benefit to use another phantom, or design an in-house phantom, that has high precision in its positioning. Figure 3(b) presents the KPI image noise relative to its base value for one slice position. A boxplot of the KPI CT number of water for all eight slice positions is presented in Fig. 3(c). The boxplot is configured so that all data are inside the whiskers. A trend curve for the KPI CT number of water for one slice location is presented in Fig. 3(d). Figures 3(e) and (f) show trend curves for the KPI uniformity and the KPI homogeneity evaluated in one slice. From the figure it can be seen that the trends in uniformity and homogeneity closely follow each other. A homogeneity test could, therefore, be used instead of a uniformity test, providing additional sensitivity to some forms of artifacts. During the two years of the pilot study, four deviations in scanner performance were found. The deviations are presented in Fig. 4. Figure 4(a) displays the KPI image noise from one slice location in the Level 0 test for one of the CT scanners over a two-year period. On one occasion, the KPI image noise deviated outside of the allowed tolerance. The deviation was discovered promptly and was caused by a barely visible ring artifact. After discussing the cause of the ring artifact with the manufacturer it was concluded that a small amount of dirt was present in the collimator during the water calibration of the CT system. Ideally, such artifacts should also be spotted with the homogeneity tests, but the initial tolerance level defined for the KPI, 6 HU, was too high and the homogeneity test did not fail. Consequently, the tolerance for the KPI homogeneity was lowered to 5 HU. Test data are being accumulated at our center to establish whether this tolerance is acceptable. The homogeneity test can perhaps never entirely replace human evaluation in the identification of artifacts; however, its potential for rapid and automatic identification of many such aberrations makes it a valuable addition to the available arsenal of QC tests. Figure 4(b) presents a trend curve for the KPI CT number of water for one slice location of the Level 0 test. The KPI CT number of water went outside of the allowed tolerance on three occasions during the pilot study. The manufacturer has verified that the problems were caused by an error related to the air calibration of the system. On all three occasions, the results could not be seen in patient images and the clinic decided to continue scanning patients, with the problem monitored, until the manufacturer came to perform a preventive maintenance. This maintenance included a new water calibration, after which the CT numbers of water returned to their expected values. (a) (b) Fig. 4. Examples of when the proposed methodology has helped medical physicists in finding deviations in scanner performance: (a) image noise values out of tolerance; (b) CT number of water out of tolerance multiple times.
11 264 Nowik et al.: QC of CT by monitoring of KPI 264 The deviations in CT scanner performance that have been presented here were small and had a minor effect on patient images. Without a daily QC routine, most likely, they would not have been discovered until the next annual QC or planned service, or until the radiologists would have noticed degraded image quality in the clinical images. Furthermore, the deviations were discovered promptly due to the high frequency of the KPI tests (sampled daily). This, in turn, entails that the CT system could be serviced appropriately in a timely fashion to prevent identified issues deteriorating into clinical problems. The results from the pilot study demonstrate the utility of testing KPIs on a daily basis. It is therefore recommended to perform daily KPI tests in order to promptly discover and service abnormalities in CT system performance that could potentially affect the clinical outcome of CT patient examinations. The possibility to perform trend analysis of various parameters gave valuable insight into the deviations and helped the medical technicians and the manufacturer to localize and handle the deviations. The individual KPIs used to discover the four deviations discussed above were the only KPIs to go outside of their allowed tolerances. In the case of the ring artifact which was discovered only by the KPI image noise, it was expected that the KPI homogeneity should go outside of its tolerance, as well. As the sensitivity of the tests using KPIs is restricted by the tolerance levels defined, the user can define how large the deviations that can be accepted from the clinical viewpoint. In other words, tolerances can be tweaked to increase the sensitivity of the individual KPI tests. Continuous data collection of Level 0 and Level 1 tests over larger periods of time will provide data on how sensitive the different KPIs are and what tolerances that are possible, and meaningful, to implement. Finally, two examples of analyses a medical physicist could perform using the daily test data to help in detecting and understanding deviations in CT scanner performance are shown in Fig. 5. In Fig. 5(a) the KPI of image noise is plotted at different tube voltages. A fit to the equation σ = a (kvp) n + c was made, resulting in fit-value of n = 2.82 for the kilovoltage dependency factor. In Fig. 5(b), the KPI for image noise is displayed at different tube current time products (mas). A fit to Brooks formula was made yielding an r 2 = These data for these analyses was available from a Level 1 test. As the KPI-based QC methodology is both fast and simple to use, it has now become an integral and welcomed part of the CT system s daily start-up routine at the clinic. (a) Fig. 5. Two examples of analysis that can be performed with a Level 1 test. (a) The KPI image noise versus tube voltage fitted to σ = a (kvp) n + c, where the kilovoltage dependency factor is n = b 2. The slices in this analysis were 2, 3, 7, and 33. (b) The KPI image noise vs. tube current time product fitted to Brooks formula. The slices in this analysis were 4, 5, 7, and 31. (b)
12 265 Nowik et al.: QC of CT by monitoring of KPI 265 IV. CONCLUSIONS A set of KPIs for CT system performance that are measurable using a standard manufacturer CT QA phantom were introduced. A methodology to perform routine QC using such KPIs was also presented. The simplicity of the method makes it feasible to perform routine QC on a daily basis. The results are stored in a database for ease of further analyses. The implementation in a clinical setting has been evaluated over a two-year period, with promising results. The method has on several occasions identified abnormalities in CT performance that would otherwise not have been detected until a later time. Medical physicists and other professionals concerned with the performance of CT systems will, using the presented daily QC routine, have access to comprehensive data on the current status and long-term trends of CT systems. Swift and pinpointed actions can then be taken in order to ensure the quality of the CT examinations and patient safety. The ease and short time required to perform this daily routine QC mean that it can readily be implemented in clinical practice once suitable software tools have been developed. ACKNOWLEDGEMENTS This research is part of the XQuality project at Karolinska University Hospital in Stockholm, Sweden. REFERENCES 1. ACR CT quality control manual. Reston, VA: ACR; IOMP. IOMP Policy Statement No. 1. The Medical Physicist: role and responsibilities. International Organization for Medical Physics; Reiner BI. Automating quality assurance for digital radiography. J Am Coll Radiol. 2009;6(7): IEC. Evaluation and routine testing in medical imaging departments Part 26: constancy tests Imaging performance of computed tomography X-ray equipment. IEC Geneva: International Electrotechnical Commission; IPEM. Recommended standards for the routine performance testing of diagnostic x-ray imaging systems. York, UK: Institute of Physics and Engineering in Medicine; IEC. Evaluation and routine testing in medical imaging departments Part 3-5: acceptance tests. Imaging performance of computed tomography X-ray equipment. IEC Geneva: International Electrotechnical Commission; Mutic S, Palta JR, Butker EK, et al. Quality assurance for computed-tomography simulators and the computedtomography-simulation process: report of the AAPM Radiation Therapy Committee Task Group No. 66. Med Phys. 2003;30(10): Almén AC and Cederlund T. Quality control of x-ray equipment used in diagnostic radiology an investigation of the activity in Sweden. SSI report 2003:09. Stockholm: Swedish Radiation Protection Institute (SSI); Strålsäkerhetsmyndigheten [Swedish Radiation Safety Authority]. Strålsäkerhetsmyndighetens föreskrifter om röntgendiagnostik [in Swedish]. SSMFS 2008:31. Stockholm: Strålsäkerhetsmyndigheten; Habibzadeh MA, Ay MR, Asl ARK, Ghadiri H, Zaidi H. Impact of miscentering on patient dose and image noise in x-ray CT imaging: phantom and clinical studies. Phys Med. 2012;28(3): Kuttner S, Bujila R, Kortesniemi M, et al. A proposed protocol for acceptance and constancy control of computed tomography systems: a Nordic Association for Clinical Physics (NACP) work group report. Acta Radiol. 2013;54(2): Brooks RA and Di Chiro G. Statistical limitations in x-ray reconstructive tomography. Med Phys. 1976;3(4): Nagel HD. CT parameters that influence the radiation dose. In: Tack D and Gevenois PA, editors. Radiation dose from adult and pediatric multidetector computed tomography. NY: Springer; p Elojeimy S, Tipnis S, Huda W. Relationship between radiographic techniques (kilovolt and milliampere-second) and CTDI(vol). Radiat Prot Dosimetry. 2010;141(1): Eldevik K, Nordhøy W, Skretting A. Relationship between sharpness and noise in CT images reconstructed with different kernels. Radiat Prot Dosimetry. 2010;139(1-3): Sprawls P. AAPM tutorial. CT image detail and noise. Radiographics. 1992;12(5):
Image Quality Tradeoffs in CT Image Quality and Radiation Dose Michael F. McNitt-Gray, PhD, DABR Associate Professor Dept. Radiology David Geffen School of Medicine at UCLA Image quality has many components
186 Chapter 5 5.4 CT Number Accuracy and Noise In many clinical practices, radiologists rely on the value of measured CT numbers to differentiate healthy tissue from disease pathology. Although this practice
X. QUALITY CONTROL & PERFORMANCE TESTING MANUAL OF OPERATIONS FOR CT The following quality control measures have been developed for the NLST. Only multi-detector CT scanners will be used for the low-dose
Physics testing of image detectors Parameters to test Spatial resolution Contrast resolution Uniformity/geometric distortion Features and Weaknesses of Phantoms for CR/DR System Testing Dose response/signal
CT simulator M. Saiful Huq,, Ph.D., FAAPM, FInstP Professor and Director Department of Radiation Oncology University of Pittsburgh Cancer Institute Pittsburgh, Pennsylvania, USA Outline What is a CT simulator?
Centers for Quantitative Imaging Excellence (CQIE) LEARNING MODULE QUANTITATIVE IMAGING IN MULTICENTER CLINICAL TRIALS: PET American College of Radiology Clinical Research Center v.1 Centers for Quantitative
CT RADIATION DOSE REPORT FROM DICOM Frank Dong, PhD, DABR Diagnostic Physicist Imaging Institute Cleveland Clinic Foundation Cleveland, OH CT Patient comes out... Patient goes in... Big Black Box Radiology
WHERE IN THE WORLD IS JILL LIPOTI? HELLO FROM NEW JERSEY CRCPD - National Symposium on Fusion Imaging and Multimodalities February 18-20, 2004 Kansas City, Missouri New Jersey s Requirements As They Pertain
Moving Forward What does this mean for the Medical Physicist and the Imaging Community? John M. Boone, Ph.D., FAAPM, FACR Professor and Vice Chairman of Radiology University of California Davis Medical
Note: This is a courtesy copy and is not the official version of this rule. The official, legally effective version of this rule is available through www.lexisnexic.com/bookstore (Phone: (800) 223-1940).
Quality Control in PET PET/CT QC/QA Magnus Dahlbom, Ph.D. Division of Nuclear Medicine Ahmanson Biochemical Imaging Clinic David Geffen School of Medicine at UCLA Los Angeles Verify the operational integrity
Performance testing for Precision 500D Classical R/F System John M. Boudry, Ph.D. Image Quality Systems Engineer GE Healthcare Technologies Outline System background Image Quality Signature Test (IQST)
Managing Pediatric Patient s CT Dose with the Help of SSDE Keith J. Strauss, MSc, FAAPM, FACR Clinical Imaging Physicist Cincinnati Children s Hospital University of Cincinnati College of Medicine INTRODUCTION
Digital Quality Assurance Area Quality Assurance Procedure Frequency Electronic Display Overall visual assessment - general image quality using Daily Device SMPTE & TG18-QC test patterns Performance using
EXTENDED 3D CT METHOD FOR THE INSPECTION OF LARGE COMPONENTS M. Simon, C. Sauerwein, and I. Tiseanu Hans Wälischmiller GmbH, Meersburg, Germany Abstract: Conventional X-ray 3D computed tomography requires
Translating Protocols Between Scanner Manufacturer and Model Cynthia H. McCollough, PhD, FACR, FAAPM Professor of Radiologic Physics Director, CT Clinical Innovation Center Department of Radiology Mayo
The Challenge of CT Dose Records Kimberly Applegate, MD, MS, FACR Professor of Radiology and Pediatrics Emory University *financial disclosures: -Springer Textbook contracts -AIM advisory board for patient
The phantom portion of the American College of Radiology ACR Computed Tomography CT accreditation program: Practical tips, artifact examples, and pitfalls to avoid a Cynthia H. McCollough b) and Michael
QUALITY ASSURANCE PROGRAMS FOR DIAGNOSTIC RADIOLOGY FACILITIES (a) Applicability Quality assurance programs as described in paragraph (c) of this section are recommended for all diagnostic radiology facilities.
Quality Assurance Testing of Gamma Camera and SPECT Systems Sharon L. White University of Alabama at Birmingham July 28, 2009 Disclaimers Gamma camera images and photographs of equipment are not intended
Design and Implementation of an Institution-Wide Patient-Specific Radiation Dose Monitoring Program for Computed Tomography, Digital Radiography, and Nuclear Medicine by Olav Christianson Medical Physics
CT: Size Specific Dose Estimate (SSDE): Why We Need Another CT Dose Index Keith J. Strauss, MSc, FAAPM, FACR Clinical Imaging Physicist Cincinnati Children s Hospital University of Cincinnati College of
CT Protocol Optimization over the Range of CT Scanner Types: Recommendations & Misconceptions Frank N. Ranallo, Ph.D. Associate Professor of Medical Physics & Radiology University of Wisconsin School of
JOURNAL OF APPLIED CLINICAL MEDICAL PHYSICS, VOLUME 7, NUMBER 1, WINTER 2006 Performance evaluation and quality assurance of Varian enhanced dynamic wedges Parham Alaei and Patrick D. Higgins Department
Diagnostic Exposure Tracking in the Medical Record J.A. Seibert, Ph.D. Department of Radiology Sacramento, California USA Vancouver. British Columbia Relevant disclosures None Learning objectives Understand
COMMISSIONING AND USE OF CONVENTIONAL CT Sasa Mutic Department of Radiation Oncology Siteman Cancer Center Mallinckrodt Institute of Radiology Washington University School of Medicine St. Louis, Missouri
A software tool for Quality Assurance of Computed / Digital Radiography (CR/DR) systems Nikunj Desai a and Daniel J Valentino a,b a icr Company Inc, 2580 West 237th Street, Torrance, CA 90505, USA b Department
Study the Quality Assurance of Conventional X-ray Machines Using Non Invasive KV meter T.M.Taha Radiation Protection Department, Nuclear Research Center, Atomic Energy Authority, Cairo.P.O.13759 Egypt.
2012 Computed Tomography QUALITY CONTROL MANUAL Radiologist s Section Radiologic Technologist s Section Medical Physicist s Section 2012 Computed Tomography QUALITY CONTROL MANUAL Radiologist s Section
MODELING AND IMPLEMENTATION OF THE MECHANICAL SYSTEM AND CONTROL OF A CT WITH LOW ENERGY PROTON BEAM João Antônio Palma Setti, firstname.lastname@example.org Pontifícia Universidade Católica do Paraná / Rua Imaculada
GE Healthcare DoseWatch Gathering Radiation Dose Data Gathering Radiation Dose Data After your organization crosses the bridge of Why? for patient radiation dose management either due to regulation, risk
CT Manufacturer s Voluntary Commitment Regarding CT Dose Updated list of dose management features The CT manufacturers have worked through MITA to provide an updated list of available technologies implemented
Role of the Medical Physicist in Clinical Implementation of Breast Tomosynthesis Bob Liu, Ph.D. Department of Radiology Massachusetts General Hospital And Harvard Medical School Digital Breast Tomosynthesis
Medical Exposure Directive 97/43 Euratom Quality Assurance Ministry of Health, Radiation Protection Department, Luxembourg Alexandra Schreiner Medical Physicist Quality Assurance (QA): All those planned
(201) 825-9500 / (800) 631-3826 THRYOID UPTAKE SYSTEMS 51 Captus 3000 Thyroid Uptake And Well System The Ultimate... User Friendly... Intuitive Performing Thyroid Uptake measurements has never been easier.
DISCLAIMER: TO THE EXTENT ALLOWED BY LOCAL LAW, THIS INFORMATION IS PROVIDED TO YOU BY THE AMERICAN ASSOCIATION OF PHYSICISTS IN MEDICINE, A NON-PROFIT ORGANIZATION ORGANIZED TO PROMOTE THE APPLICATION
Quality Control and Maintenance Programs Cari Borrás, D.Sc., FACR, FAAPM Visiting Professor DOIN-DEN / UFPE Recife, Pernambuco, Brazil Co-Chair, IUPESM Health Technology Task Group 1 Medical Imaging Equipment
Patient Dose Tracking for Imaging Studies David E. Hintenlang, Ph.D., DABR University of Florida Conflict of Interest Statement No affiliation or financial interests in any of the commercial products or
2166-Handout College on Medical Physics. Digital Imaging Science and Technology to Enhance Healthcare in the Developing Countries 13 September - 1 October, 2010 Digital Radiography Image Parameters SNR,
CT for PET/CT and SPECT/CT Principles of Dose Reduction Frederic H. Fahey DSc Children s Hospital Boston Harvard Medical School email@example.com Lifetime Attributable Risk 10 mgy in
Dosimetry in Diagnostic Medical Imaging: State of the Art and New Developments Annalisa Trianni Medical Physics Department Udine University Hospital firstname.lastname@example.org Content Understand
Routine QA of Dedicated PET Scanners M.E. Daube-Witherspoon, Ph.D. University of Pennsylvania Introduction Purposes of QA testing: Verify continued scanner performance Identify problems prior to scanning
CT Dose Notifications and Alerts What are they, how do they work, and important information for successful implementation April 16, 2014 Disclaimer The information contained herein is current as of the
(201) 825-9500 / (800) 631-3826 THYROID UPTAKE SYSTEMS 49 Captus 3000 Thyroid Uptake And Well System The Ultimate... User Friendly... Intuitive Performing Thyroid Uptake measurements has never been easier.
Q.A. Collectible Sponsored by CRCPD s Committee on Quality Assurance in Diagnostic X-Ray (H-7) Mammography Phantom Image Quality Evaluation (from the American College of Radiology 1999 Mammography Quality
Designation: E 1815 96 (Reapproved 2001) Standard Test Method for Classification of Film Systems for Industrial Radiography 1 This standard is issued under the fixed designation E 1815; the number immediately
Donna M. Reeve, MS, DABR, DABMP Department of Imaging Physics Educational Objectives Discuss the importance of breast MRI quality control (QC). Provide an overview of the new ACR Breast MRI Accreditation
Software Tools for DICOM Media Exchange in Clinical Research: the SWABIK Project Software Tools for DICOM Media Exchange in Clinical Research: the SWABIK Project O. ElGazzar 1, M. Onken 1, M. Eichelberg
RADIATION THERAPY Dosimetry Pioneers since 1922 NEW New Product Releases 2015 One phantom multiple options 4D Patient and Machine QA Modular Phantom for Stay flexible. Go modular. } Revolutionary new modular
ACR Diagnostic Imaging Accreditation The Physicist Role By Dina Hernandez BSRT, RT (R)(CT)(QM) Team Lead CT/MR Accreditation 1 Overview of ACR MIPPA Physicist Role Application Testing QC requirements Outline
Technical Note: 9 Use of the VALIDATOR Dosimetry System for Quality Assurance and Quality Control of Blood Irradiators 1- Introduction The VALIDATOR, model TN-ID-60, is a compact, and stand-alone dosimetry
American College of Radiology CT Accreditation Program Testing Instructions (Revised July 24, 2015) This guide provides all of the instructions necessary for clinical tests, phantom tests and general submission
Chapter 12 QUALITY ASSURANCE OF EXTERNAL BEAM RADIOTHERAPY D.I. THWAITES Department of Oncology Physics, Edinburgh Cancer Centre, University of Edinburgh, Western General Hospital, Edinburgh, United Kingdom
Quantitative Imaging In Clinical Trials Using PET/CT: Update Paul Kinahan, Robert Doot Imaging Research Laboratory Department of Radiology University of Washington, Seattle, WA Supported by RSNA Quantitative
DISCLAIMER: TO THE EXTENT ALLOWED BY LOCAL LAW, THIS INFORMATION IS PROVIDED TO YOU BY THE AMERICAN ASSOCIATION OF PHYSICISTS IN MEDICINE, A NON-PROFIT ORGANIZATION ORGANIZED TO PROMOTE THE APPLICATION
Today s agenda Computed Tomography Presented by: Dina Hernandez, BSRT, RT (R), CT, QM Krista Bush, RT, MBA Leonard Lucey, JD ACR Quality & Safety MIPPA overview and CMS requirements CT accreditation How
Scan for mobile link. Proton Therapy Proton therapy delivers radiation to tumor tissue in a much more confined way than conventional photon therapy thus allowing the radiation oncologist to use a greater
www.neurologica.com Critical Care Clinic Operating Room Radiology ED Trauma OR Neuro/Spine Pediatric Plastics Thoracic General Interventional Radiology ENT ED Trauma Trauma Stroke Critical Care Neuro Pediatric
CT QC Under the ACR QC Manual Douglas Pfeiffer, MS, DABR Boulder Community Hospital What? There s a Manual??! YES!!! Almost! Doesn t have a pretty cover yet! Should be out by RSNA! Really seriously I mean
ACR MRI Accreditation Program: The Technologist s Role Geoffrey D. Clarke, Ph.D. University of Texas Health Science Center at San Antonio Radiological Sciences Division Overview ACR MRI Accreditation Program
Scan for mobile link. X-ray (Radiography) - Chest What is a Chest X-ray (Chest Radiography)? The chest x-ray is the most commonly performed diagnostic x-ray examination. A chest x-ray produces images of
On the Use of a Diagnostic X-Ray Machine for Calibrating Personal Dosimeters A. T. Baptista Neto, T. A. Da Silva Centro de Desenvolvimento da Tecnologia Nuclear, Comissão Nacional de Energia Nuclear, Rua
The Ultimate... User Friendly... Intuitive Captus 3000 Thyroid Uptake And Well System Performing Thyroid Uptake measurements has never been easier. The Captus 3000 combines a totally new software program
Proceedings of the Ninth EGS4 Users' Meeting in Japan, KEK Proceedings 200-22, p.5-8 MONTE CARLO SIMULATION ANALYSIS OF BACKSCATTER FACTOR FOR LOW-ENERGY X-RAY K. Shimizu, K. Koshida and T. Miyati Department
Principles of Imaging Science I (RAD 119) X-ray Tube & Equipment X-ray Imaging Systems Medical X-ray Equipment Classified by purpose or energy/current levels kvp, ma Radiographic Non-dynamic procedures
AAPM COMP 011 Radiography: D and 3D Metrics of Performance Towards Quality Index Ehsan Samei, Sam Richard Duke University Learning objectives Outlook 1. To understand methods for D and 3D resolution measurements..
GE Healthcare Evolution for Bone Collimator-Detector Response Compensation in Iterative SPECT Imaging Reconstruction Algorithm Version 1.0 Introduction The quality and quantitative accuracy of SPECT reconstructed
MDCT Technology Kalpana M. Kanal, Ph.D., DABR Assistant Professor Department of Radiology University of Washington Seattle, Washington ACMP Annual Meeting 2008 - Seattle, WA Educational Objectives Historical
Spiral CT: Single and Multiple Detector Systems AAPM Refresher Course Nashville, TN July 28,1999 Mike McNitt-Gray, PhD, DABR Assistant Professor UCLA Radiological Sciences email@example.com X-Ray
Whitepapers on Imaging Infrastructure for Research Paper 1. General Workflow Considerations Bradley J Erickson, Tony Pan, Daniel J Marcus, CTSA Imaging Informatics Working Group Introduction The use of
Quality Control of Full Field Digital Mammography Units Melissa C. Martin, M.S., FACMP, FACR, FAAPM Melissa@TherapyPhysics.com 310-612-8127 ACMP Annual Meeting Virginia Beach, VA May 2, 2009 History of
Department of Health and Human services Population Health Radiation Protection Act 2005 Section 17 CERTIFICATE OF COMPLIANCE: STANDARD FOR RADIATION APPARATUS - X-RAY MEDICAL DIAGNOSTIC (MAMMOGRAPHY) SECTION
SITE IMAGING MANUAL ACRIN 6698 Diffusion Weighted MR Imaging Biomarkers for Assessment of Breast Cancer Response to Neoadjuvant Treatment: A sub-study of the I-SPY 2 TRIAL Version: 1.0 Date: May 28, 2012
Surveying and QC of Stereotactic Breast Biopsy Units for ACR Accreditation LORAD Stereotactic Breast Biopsy System AAPM Spring Clinical Meeting Phoenix, AZ March 17, 2013 Melissa C. Martin, M.S., FACR,
Automotive Applications of 3D Laser Scanning Kyle Johnston, Ph.D., Metron Systems, Inc. 34935 SE Douglas Street, Suite 110, Snoqualmie, WA 98065 425-396-5577, www.metronsys.com 2002 Metron Systems, Inc
Australasian Physical & Engineering Sciences in Medicine Volume 28 Number 2, 2005 ACPSEM POSITION PAPER Recommendations for a technical quality control program for diagnostic X-ray equipment D. A. Causer