Ongoing discussion within the nuclear medicine community suggests that all radionuclide therapies (RNTs) should include posttreatment quantitative dosimetry as part of standard clinical care. The hypothesis is that fixed administered activities limit the potential efficacy of RNT and increase the risk of side effects; therefore, patient-specific dosimetry should be leveraged to improve patient outcomes. Furthermore, the development of new radionuclides is often constrained by dosimetry-defined limits to normal organs extrapolated from external-beam radiation therapy (EBRT).
At the same time, with few exceptions, nonradioactive oncologic therapies are administered as fixed or calculated doses based on patient weight or body surface area. Although personalized dosing schemes based on tumor burden, pharmacokinetics, and pharmacodynamics can potentially improve the therapeutic index of cancer treatments, very few such regimens have been adopted in clinical practice (1–3). Large, randomized clinical trials are required to validate personalized treatment regimens compared with conventional ones, and few such trials have been conducted.
RNTs differ from nonradioactive systemic cancer treatments, as absorbed doses to tumors and normal organs can be quantified directly. Many experts argue that the similarities between RNT and EBRT mandate dosimetry: no radiation oncologist would conceive of treating a patient without a precise dose calculation to target tumor and surrounding tissues. Once a threshold of radioactivity administered to a field is exceeded, toxicity can be irreversible. By analogy, no systemic RNT should be administered without analyzing the absorbed dose to at-risk organs and tumors (4,5).
However, in many other ways, this comparison fails. EBRT is administered in a prescriptive fashion to a specific region. Effective dose ranges (typically measured in grays) for particular cancers have been well established, and the radiation sensitivities of surrounding tissues are known. Precise radiation doses to tumors and adjacent organs can be calibrated using increasingly sophisticated techniques to maximize response and minimize toxicity (6). None of these features of EBRT can be translated to systemically administered RNT. Instead, RNT dosimetry estimates absorbed dose to tumors and organs using imaging after the administration of the therapy. Additionally, absorbed doses and their biologic effects vary substantially on the basis of radionuclide properties, including pathlength and linear energy transfer (7).
In EBRT, dosimetry is calculated for tissues within a radiation field. With RNT, the usual organs of concern are typically the kidneys (for renally excreted radiopharmaceuticals) and the bone marrow, the organ most sensitive to the effects of systemic radiation. Renal doses are more straightforward to estimate than marrow doses, and therefore, the kidneys are commonly treated as the target organ. Traditionally, a threshold absorbed dose of 23 Gy to the kidneys has been considered a maximum tolerable dose, guided by the International Commission on Radiological Protection recommendations or QUANTEC (Quantitative Analyses of Normal Tissue Effects in the Clinic) (7–9). However, there are obvious pitfalls when trying to correlate the biologic effects of EBRT on the entire kidney with the effects of radionuclides excreted through renal tubules. Even among β-emitting radionuclides, the differences in nephrotoxicity between 90Y (12-mm pathlength) and 177Lu (2-mm pathlength) are substantial when administered at a similar estimated absorbed dose to the kidney (10).
Let us assume that we could accurately measure normal-organ–absorbed doses. With respect to kidney-based dosing, it is increasingly apparent that the kidney is not a dose-limiting organ. Although single-arm studies have suggested that 177Lu-DOTATATE causes a modest annual decrease in glomerular filtration rate, the randomized phase III NETTER 1 trial showed no difference in creatinine clearance over time between the 177Lu-DOTATATE and control arms (11). Studies that calibrate administered activity on the basis of absorbed renal doses nearly uniformly lead to an increase in total administered activity (12). However, if the kidney is not a dose-limiting organ, administration up to an artificial renal dose threshold should not be considered a personalized form of treatment but rather a simple dose escalation.
The bone marrow is a dose-limiting organ for many patients, and an absorbed dose of 2 Gy to the red marrow is considered a maximum threshold, extrapolated from 131I therapy (13). However, marrow dosimetric calculations are imprecise. Uptake on posttherapy imaging may be reduced because of partial-volume effects or may be overestimated because of background noise. Even with the addition of plasma sampling, calculated marrow doses may vary depending on the technique, and different radionuclides can produce substantial variations in marrow toxicity (14). Moreover, patient-specific factors (age, genomic predisposition, prior treatments, etc.) influence sensitivity to radiation (as is the case with chemotherapy) (15–17). Some studies have shown no correlation between bone marrow dose and cytopenias; others have shown weak correlations (10,18,19). However, a more straightforward method of assessing bone marrow toxicity and adjusting administered activity is the complete blood count. There is no evidence that dosimetric calculations are superior to a simple complete blood count for personalizing treatment. Additionally, there is no evidence that dosimetry can predict the most catastrophic long-term complications of treatment: myelodysplastic syndrome or acute leukemia (20,21).
Dosimetry can also be used to calculate absorbed tumor doses. Although dose–response relationships are expected, our understanding of tumor dosimetry in RNT lags far behind our knowledge of optimal dosing in EBRT. Traditional approaches using several manually identified index tumor lesions, typically with the highest activity, fail to correlate with survival outcomes or lead to actionable changes in management (22). This is not surprising given tumor heterogeneity and the varying doses delivered to sites within an individual. Newer approaches using whole-body tumor dosimetry may be superior. For example, among 11 patients who received less than a 10-Gy median whole-body tumor dose after 177Lu-PSMA-617 treatment, only one achieved a prostate-specific antigen response (23). However, it is not yet clear how tumor dosimetry data can be leveraged to improve patient outcomes. For example, should a low absorbed tumor dose prompt additional cycles of treatment or early discontinuation for futility?
Dosimetry has evolved enormously in the last decade. There has been a transition from planar imaging to quantitative SPECT/CT imaging. This has enabled a shift from dosimetry modeling based on a standard human with assumed organ masses and shapes to direct measurements using voxel-based techniques (24). New PET-like ring-designed SPECT/CT devices using solid-state detectors enable more accurate dose estimates with better resolution and speed. Contouring of normal organs has moved from a manual process at multiple time points to semiautomated techniques using defined thresholds or fully automated techniques using deep learning algorithms assisted by the CT (25). The number of time points required for accurate dosimetry is decreasing, with even single time points feasible, using either patient-specific parameters from cycle 1 or population-based databases (26).
Dosimetry sits at a crossroads. It is time to move away from extrapolating external-beam–defined normal-organ constraints to RNT. Direct observation of adverse effects is simpler and superior. We must still monitor for longer-term adverse effects, especially within organs of interest. Advances in quantitative SPECT/CT and software open new opportunities to redefine the use of dosimetry to improve patient outcomes. Undoubtedly, there are superior personalized administration schedules that modulate the amount and frequency of the administered activity. However, only well-designed randomized clinical trials with long-term follow-up can accurately evaluate whether novel dosimetry-based prescriptions are superior to fixed schedules. As with any other oncologic therapy, the burden of proof is on us to demonstrate that these strategies yield superior efficacy or safety outcomes. We hope that improved evidence-based strategies will be developed to improve patient care.
DISCLOSURE
Michael Hofman acknowledges philanthropic/government grant support from the Prostate Cancer Foundation (PCF), including CANICA Oslo Norway, the Peter MacCallum Foundation, the Medical Research Future Fund (MRFF), an NHMRC Investigator Grant, Movember, and the Prostate Cancer Foundation of Australia (PCFA); other funding (in the last 10 y) from the U.S. Department of Defense; research grant support (to the Peter MacCallum Cancer Centre) from Novartis (including AAA and Endocyte), ANSTO, Bayer, Isotopia, and MIM; and consulting fees for lectures or advisory boards from Astellas and AstraZeneca (in the last 2 y) and from Janssen, Merck/MSD, and Mundipharma (in the last 5 y). Thomas Hope has received grant funding to the institution from Clovis Oncology, GE Healthcare, Lantheus, Janssen, the Prostate Cancer Foundation, Telix, and the National Cancer Institute (R01CA235741 and R01CA212148); personal fees from Bayer, BlueEarth Diagnostics, and Lantheus; and fees from and has an equity interest in RayzeBio and Curium. Jonathan Strosberg has received grant funding to the institution from Rayzebio, Radiomedix, Novartis, and ITM. No other potential conflict of interest relevant to this article was reported.
Footnotes
Published online Jan. 11, 2024.
- © 2024 by the Society of Nuclear Medicine and Molecular Imaging.
REFERENCES
- Received for publication November 28, 2023.
- Accepted for publication December 8, 2023.