Almost daily, and from many media sources in the US, one is confronted with the disquieting demographic fact of the early 21st century that the baby ‘boomers’ are getting older. This fact introduces many of the grants reviewed at the National Institutes of Health (NIH), which emphasize the growing population at risk for cancer development. Of course, this fact seems to inform almost every disease analysis that could possibly be age-dependent, as well as some lesser known facts such as that as people age, they are less likely to relocate, which is illustrated by the statistic that the moving rate drops from 70% in those who are 20–29 years old, to 9% in 45–65 year-olds. Perhaps as the boomers grow older, they will have their cancer treated in more non-urban settings, further stressing the supportive care networks. Unfortunately, not all cancer treatments may be applied uniformly throughout the US, making this relocation statistic even more interesting. To quote an important finding from the Surveillance, Epidemiology, and End Results (SEER) data provided in 2004:
“The continued measurable declines for overall cancer death rates and for many of the top 15 cancers, along with improved survival rates, reflect progress in the prevention, early detection, and treatment of cancer. However, racial and ethnic disparities in survival and the risk of death from cancer, and geographic variation in stage distributions suggest that not all segments of the US population have benefited equally from such advances.”1
The diversity of the 76 million baby boomers born between 1946 and 1964, and their projected longer life span, create challenges for all those examining these facts. These demographic issues also inform the discussion of the use of myeloid colony stimulating factors (CSFs) in cancer. Their use is projected to increase and is not uniformly applied across the US, and the new molecular advances in translational cancer medicine may yet alter their use as well.
There are four general ways that CSFs are used in general cancer treatment in a non-transplant setting. The first is as a primary prophylactic measure to prevent severe neutropenia with its attendant morbidity and mortality risks and cost consequences. The ASCO guidelines point out that cost analyses have shown that CSFs save money when the risk of febrile neutropenia (FN) is greater than 40%.2 These guidelines also point out that most moderate-dose-intensity solid tumor cancer regimens have FN rates of approximately 15%, and that primary prophylaxis is not a regular approach for these chemotherapies. The exception to this rule is the older patient where the 15% assumption is not correct. In nine clinical trials of older patients with large cell non-Hodgkin’s lymphoma, the risk of FN was between 21% and 47%, and many authors indicate that primary prophylaxis is appropriate for patients receiving moderate intensity chemotherapy who are aged 70 years or older.3 Furthermore, advances in treatment have made the combination of monoclonal antibody therapy, using Rituxan™, with chemotherapy the standard of care, or at least the benchmark of care, in the lymphoid malignancies, lymphoma, and chronic lymphocytic leukaemia (CLL). This has resulted unexpectedly in increased myelosuppression in some of these combination regimens. Thus, the on-coming aging boomers will require increased use of CSFs in these clinical situations. Clinical practice for primary prophylaxis is also quite varied in the US, as suggested by the aforementioned SEER results statistics, indicating disparity of outcomes. Several studies have indicated that compliance with ASCO guidelines for primary prophylaxis with CSFs is not much better than 50%,4 with more use rather than less being the most common.
The second use of myeloid CSFs is as secondary prophylaxis, which means incorporating their administration after a patient has had an episode of FN or severe neutropenia. Here, too, there is controversy regarding the alternative choice of dose reduction of the chemotherapy as a means of preventing another event, or keeping the same chemotherapy and initiating the use of myeloid CSFs. In practice, CSFs are routinely given at this point whether dose modification is made or not.
The third use of CSFs in cancer is for myeloid recovery after dose intense chemotherapy such as in the acute leukemias. In this setting, all patients will have severe neutropenia, and the majority will have fever and require hospitalization.The aim in this clinical situation is to accelerate myeloid recovery after treatment, reduce the duration of antibiotic therapy, attempt to reduce the complications of infection, and shorten the hospital stay. This use has been endorsed by the ASCO guidelines,2 especially in the older patient, is economically sound, and is almost uniformly adopted in US practice.5 Again, demographic facts indicate the expected increase in this use with the aging boomers, as numbers of patients with acute myeloid leukaemia (AML) will rise with the graying tide. In addition, the hematologic disorder termed myelodysplastic syndrome (MDS) is more common than AML in older individuals and it too is on the rise.The first and only FDA-approved chemotherapy (Vidaza™) for MDS was approved this year. Its use in MDS will undoubtedly increase the use of CSFs. The use of CSFs during Vidaza™ treatment will be predominantly in line with both primary and secondary prophylaxis guidelines from ASCO. Published reports indicate granulocytopenia occurred in 58% of patients treated with this agent, suggesting that this treatment will meet ASCO guidelines for primary prophylaxis.6
The fourth use of CSFs in cancer has been its use as a ‘priming’ agent given before chemotherapy to increase the effect of chemotherapy. This has been studied in the myeloid leukemias with mixed results. In general, this approach has not been successful in most trials; however, a recent report of improved disease-free survival (DFS) has produced a resurgence of interest in this approach.7 After an average follow-up of 55 months, patients in complete remission who received their induction chemotherapy plus granulocyte (G)-CSF had a higher rate of DFS than patients who did not receive G-CSF. G-CSF did not improve the outcome in the subgroup of patients with an unfavorable prognosis based on their cytogenetic findings. In the 72% of patients with standard-risk AML, there was an overall survival at four years of 45%, compared with 35% in the group that did not receive G-CSF. This intriguing observation may spur more research.
So what factors may influence the continued use of the myeloid CSFs in cancer therapy in the future? One of the most touted advances in our treatment approach to cancer is the development of molecularly targeted agents. These treatments in general will not produce the myelosuppression that we have been discussing with chemotherapy. The expected result is already apparent, and diseases that respond to these drugs can be treated without chemotherapy. It is hoped that more such agents will arrive in the future, especially for the severe diseases such as leukemia. Already, second-line therapy clinical trials are being designed with these agents, especially in the older patient. Obviously, this will decrease the use of CSFs in cancer in general, as more agents come on-line. Finally, one can only imagine that molecular agents may mimic the CSFs themselves. The CSFs produce their myeloid stimulatory effect at a cellular level. This effect is the result of the transmission of signals from the cell receptor site for the CSF (G-CSF or granulocyte-macrophage (GM)-CSF) through signal transduction pathways in the cell to the nuclear processes that regulate the cell cycle, division, and maturation. These intracellular pathways themselves can be the target of molecular therapies that mimic the action of CSFs. This promise may be on the horizon in any laboratory at this moment. In the future this option may increase choices and provide new agents beyond today’s CSFs. ■