TO THE EDITOR: In the September issue of The Journal of Nuclear Medicine, Kramer et al. comprehensively investigated the repeatability (R) of various quantitative 18F-FDG uptake metrics in lung cancer patients, including SUVmax, SUVmean, SUVpeak, metabolically active tumor volume (MATV), and total lesion glycolysis (TLG) (1). A test–retest study was performed within 3 d involving double-baseline whole-body 18F-FDG PET/CT at 60 and 90 min after injection. The results were compared with those of some previously published studies on various oncologic diseases, such as ovarian cancer, non–small cell lung cancer, esophageal cancer, and liver metastases for TLG in particular.
However, a previous study on TLG variability in lung cancer patients—a study involving SUVmean and MATV—was not included in the comparison of Kramer et al. (2). That study assessed R differently; instead of a test–retest acquisition within 3 d over a total of 60 lesions, as used by Kramer et al., a dynamic acquisition involving 10 frames within 60–110 min after injection over 13 lesions was used. Moreover, SUVpeak R was obtained from a further dynamic study involving 20 lung cancer lesions (3). We thought it would of interest to compare the two methods—that is, dynamic and test–retest.
The values for dynamic R obtained within 60–110 min after injection versus test–retest R obtained at 90 min after injection for all lesions in the study of Kramer et al. (R is reproducibility coefficient in their Table 3) are as follows: 19.6 versus 23.3%, 14.1 versus 17.8%, 13.2 versus 15.8%, 31.6 versus 23.7%, and 36.4 versus 30.7% for SUVmax, SUVmean, SUVpeak, TLG, and MATV (95% reliability), respectively (1–3). Although the parameter range and the “50% of SUVmax” thresholding method (used for assessing SUVmean, MATV, and hence TLG) were not exactly alike, the dynamic R percentages are consistent with the test–retest R percentages, as much as can be determined considering that R estimate uncertainty was not provided.
Let us further compare the two designs. The dynamic design involved ten 2.5-min frames leading to a ±12.5-min time window around a mean uptake time, in comparison with the test–retest design, which reported a maximal range of 7 min for uptake time (scan 1 in Table 1 (1)). We suggest that reducing the acquisition time for bed positions—that is, less than the 2.5 min that is possible with modern PET/CT systems—and reducing the number of dynamic frames may bridge the two designs. In the framework of assessing response to treatment, it is noteworthy that the dynamic design does not take into account some origins of parameter variability such as changes in plasma glucose level (within the reference range), injected dose, and differences in uptake time. However, Kramer et al. reported that glucose correction does not improve R performance (and even deteriorates it) and that the relative uncertainty about the injected dose is usually very low (1). Regarding differences in uptake time, Kramer et al. showed that the correction proposed by van den Hoff et al. significantly reduced differences between 60-min and 90-min data (4). We would like to emphasize, first, that the time window of the dynamic design includes a variability of ±12.5 min around a mean uptake time, which may take into account differences in uptake time usually met in current clinical practice, and second, that simply removing the 18F physical decay correction can reduce differences between 60-min and 90-min SUVs (5). Finally, in comparison with the test–retest design, the dynamic design involving several frames reduces the number of lesions to be investigated (and hence the number of patients to be recruited) in order to reliably determine R, since for the same number of lesions, the greater the number of dynamic frames the lower the R estimate uncertainty. We thus suggest that the dynamic design takes into consideration both the patient radiation dose and a busy clinical practice.
In conclusion, the comprehensive study of Kramer et al. about the R value for various 18F-FDG uptake measures will be useful to nuclear physicians in their current practice (1). That study was achieved using a test–retest design, and we would like to emphasize that the tool box for assessing measurement uncertainty in quantitative PET imaging fortunately offers various designs, each with its own pros and cons. An alternative dynamic design is available that may be particularly suitable when the role of technical parameters in this uncertainty is investigated in an arbitrary PET/CT system. Nevertheless, whatever the design, it clearly appears that guidelines should recommend that any quantitative outcome be accompanied by its measurement uncertainty, which should be specifically determined for each PET/CT system as soon as commissioning.
Footnotes
Published online Dec. 8, 2017.
- © 2017 by the Society of Nuclear Medicine and Molecular Imaging.