Daily aspirin might cut risk of fibrosis progression

Aspirin may be an option in NAFLD
Article Type
Changed
Tue, 11/19/2019 - 10:46

Taking daily aspirin may help keep nonalcoholic fatty liver disease (NAFLD) from progressing to liver fibrosis and nonalcoholic steatohepatitis (NASH), suggest the results of a prospective study of 361 adults.

Previously, preclinical evidence had linked aspirin to fibrogenesis prevention in fatty liver disease, but this is the first report of a prospective study to do so. Daily aspirin use “was associated with less severe histologic features of NAFLD (nonalcoholic fatty liver disease) at study enrollment and with significantly lower risk for advanced fibrosis over time in a duration-dependent manner,” wrote Tracey G. Simon, MD, MPH, and her associates. Their report is in Clinical Gastroenterology and Hepatology.

The study comprised 361 adults with biopsy-confirmed NAFLD who were enrolled in the Massachusetts General Hospital NAFLD Repository between 2006 and 2015. At baseline, 151 individuals were already on daily aspirin, usually to reduce the primary (54%) or secondary (30%) risk of cardiovascular disease. Median duration of aspirin use was 2.5 years. After a median 7.4 years of follow-up (which was similar between aspirin users and nonusers), daily aspirin use was associated with significantly lower odds of NASH (adjusted odds ratio, 0.68; 95% confidence interval, 0.37-0.89) and fibrosis (aOR, 0.54; 95% CI, 0.31-0.82).

The researchers did not find a similar protective effect for nonsteroidal anti-inflammatory drugs (NSAIDs) other than aspirin (adjusted hazard ratio for advanced fibrosis, 0.93; 95% CI, 0.81–1.05). This might be because of differences between how aspirin and nonaspirin NSAIDs affect COX isoforms – aspirin does so irreversibly, while other NSAIDs have a reversible effect, they added. “Nonaspirin NSAIDs also disrupt the intestinal barrier, increasing delivery of proinflammatory cytokines to the liver,” they wrote. “Finally, aspirin uniquely modulates bioactive lipids by stimulating the biosynthesis of pro-resolving mediators and inhibiting proinflammatory lipids, which in turn may prevent progressive liver damage.”

In this study, a single blinded hepatopathologist interpreted baseline liver biopsy specimens, and patients were followed every 3-6 months with clinical examinations and serial calculations of FIB-4, NFS, and APRI scores. All patients were followed for at least a year. Patients were classified as users of nonaspirin NSAIDs if they reported using an NSAID besides aspirin at least twice weekly, or if they had been prescribed drugs such as ibuprofen, naproxen, ketoprofen, diclofenac, or indomethacin.

In a longitudinal analysis of the 317 patients who had early-stage (F0-2) fibrosis at baseline, 86 developed new-onset advanced fibrosis over a median of 3,692 person-years, the researchers said. In all, 26 individuals developed hepatic decompensation and 18 patients died, including eight from liver-related causes. Importantly, the link between aspirin and decreased risk of fibrosis progression seemed to depend on duration of use (adjusted P trend = .026), with the greatest benefit seen with 4 years or more of use (aHR, 0.50; 95% CI, 0.35-0.73). Although subgroup analyses were limited by lack of power, daily aspirin use was associated with a 36% lower odds of incident advanced fibrosis among the 72 study participants who had paired biopsy samples, even after accounting for the effect of age, sex, baseline fibrosis stage, and time between biopsies (aOR, 0.64; 95% CI, 0.50-0.80).

“Our findings add to the growing literature supporting the potential hepatoprotective effects of aspirin in NAFLD,” the researchers concluded. “Research to uncover the mechanisms by which aspirin might prevent fibrogenesis could help develop urgently needed antifibrotic therapies for NAFLD.”

Funders included the National Institutes of Health and the AASLD Foundation. The investigators reported having no conflicts of interest.

SOURCE: Simon TG et al. Clin Gastroenterol Hepatol. 2019 May 8.

Body

Slowing, preventing, or reversing fibrogenesis in patients with NAFLD remains an unmet need. Lifestyle interventions are beneficial to this population but challenging because of concerns with adherence and sustainability, thus, favoring pharmacologic interventions.

The study by Simon et al. provides initial prospective evidence of the role of aspirin in reducing progression of fibrosis. In a thoughtful design, authors showed both cross-sectional and longitudinal associations of reduced risk for progressed fibrosis among aspirin users, all with biological coherence and while accounting for various confounding factors. Although the accuracy of blood-based noninvasive assessment of liver fibrosis (by FIB-4, NFS, and APRI) to determine progression of fibrosis in NAFLD has moderate accuracy at its best, the relatively high FIB-4 cutoff value used by the authors and their sensitivity analyses (including liver biopsy and combinations of blood-based markers combined endpoints) bring certainty to their results.

However, before we can start prescribing aspirin to halt progression of fibrosis in NAFLD, larger and adequately powered studies are needed. Caution with the use of aspirin as prophylaxis for atherosclerotic cardiovascular disease (ASCVD) is now advised, based on results from large clinical trials (i.e., ASCEND). NAFLD patients represent a particular population with both a high ASCVD risk and a high risk for gastrointestinal bleeding, and it is unclear what the number needed to treat or to harm would be without confirmatory studies.

Dr. Andres Duarte-Rojo
An “NAFLD polypill” including a combination of drugs addressing multiple metabolic pathways (e.g. aspirin, a statin, and vitamin E) might well tip the scale in favor of improved clinical outcomes, a concept recently shown as beneficial for ASCVD prevention in the PolyIran study. Until then, properly weighing the use of prophylactic aspirin in patients with NAFLD and adhering to standard recommendations is advised.

Andres Duarte-Rojo, MD, PhD, is associate professor of medicine, division of gastroenterology, hepatology, and nutrition at the University of Pittsburgh Medical Center, and Pittsburgh Liver Research Center. He received research support from Echosens, USA.

Publications
Topics
Sections
Body

Slowing, preventing, or reversing fibrogenesis in patients with NAFLD remains an unmet need. Lifestyle interventions are beneficial to this population but challenging because of concerns with adherence and sustainability, thus, favoring pharmacologic interventions.

The study by Simon et al. provides initial prospective evidence of the role of aspirin in reducing progression of fibrosis. In a thoughtful design, authors showed both cross-sectional and longitudinal associations of reduced risk for progressed fibrosis among aspirin users, all with biological coherence and while accounting for various confounding factors. Although the accuracy of blood-based noninvasive assessment of liver fibrosis (by FIB-4, NFS, and APRI) to determine progression of fibrosis in NAFLD has moderate accuracy at its best, the relatively high FIB-4 cutoff value used by the authors and their sensitivity analyses (including liver biopsy and combinations of blood-based markers combined endpoints) bring certainty to their results.

However, before we can start prescribing aspirin to halt progression of fibrosis in NAFLD, larger and adequately powered studies are needed. Caution with the use of aspirin as prophylaxis for atherosclerotic cardiovascular disease (ASCVD) is now advised, based on results from large clinical trials (i.e., ASCEND). NAFLD patients represent a particular population with both a high ASCVD risk and a high risk for gastrointestinal bleeding, and it is unclear what the number needed to treat or to harm would be without confirmatory studies.

Dr. Andres Duarte-Rojo
An “NAFLD polypill” including a combination of drugs addressing multiple metabolic pathways (e.g. aspirin, a statin, and vitamin E) might well tip the scale in favor of improved clinical outcomes, a concept recently shown as beneficial for ASCVD prevention in the PolyIran study. Until then, properly weighing the use of prophylactic aspirin in patients with NAFLD and adhering to standard recommendations is advised.

Andres Duarte-Rojo, MD, PhD, is associate professor of medicine, division of gastroenterology, hepatology, and nutrition at the University of Pittsburgh Medical Center, and Pittsburgh Liver Research Center. He received research support from Echosens, USA.

Body

Slowing, preventing, or reversing fibrogenesis in patients with NAFLD remains an unmet need. Lifestyle interventions are beneficial to this population but challenging because of concerns with adherence and sustainability, thus, favoring pharmacologic interventions.

The study by Simon et al. provides initial prospective evidence of the role of aspirin in reducing progression of fibrosis. In a thoughtful design, authors showed both cross-sectional and longitudinal associations of reduced risk for progressed fibrosis among aspirin users, all with biological coherence and while accounting for various confounding factors. Although the accuracy of blood-based noninvasive assessment of liver fibrosis (by FIB-4, NFS, and APRI) to determine progression of fibrosis in NAFLD has moderate accuracy at its best, the relatively high FIB-4 cutoff value used by the authors and their sensitivity analyses (including liver biopsy and combinations of blood-based markers combined endpoints) bring certainty to their results.

However, before we can start prescribing aspirin to halt progression of fibrosis in NAFLD, larger and adequately powered studies are needed. Caution with the use of aspirin as prophylaxis for atherosclerotic cardiovascular disease (ASCVD) is now advised, based on results from large clinical trials (i.e., ASCEND). NAFLD patients represent a particular population with both a high ASCVD risk and a high risk for gastrointestinal bleeding, and it is unclear what the number needed to treat or to harm would be without confirmatory studies.

Dr. Andres Duarte-Rojo
An “NAFLD polypill” including a combination of drugs addressing multiple metabolic pathways (e.g. aspirin, a statin, and vitamin E) might well tip the scale in favor of improved clinical outcomes, a concept recently shown as beneficial for ASCVD prevention in the PolyIran study. Until then, properly weighing the use of prophylactic aspirin in patients with NAFLD and adhering to standard recommendations is advised.

Andres Duarte-Rojo, MD, PhD, is associate professor of medicine, division of gastroenterology, hepatology, and nutrition at the University of Pittsburgh Medical Center, and Pittsburgh Liver Research Center. He received research support from Echosens, USA.

Title
Aspirin may be an option in NAFLD
Aspirin may be an option in NAFLD

Taking daily aspirin may help keep nonalcoholic fatty liver disease (NAFLD) from progressing to liver fibrosis and nonalcoholic steatohepatitis (NASH), suggest the results of a prospective study of 361 adults.

Previously, preclinical evidence had linked aspirin to fibrogenesis prevention in fatty liver disease, but this is the first report of a prospective study to do so. Daily aspirin use “was associated with less severe histologic features of NAFLD (nonalcoholic fatty liver disease) at study enrollment and with significantly lower risk for advanced fibrosis over time in a duration-dependent manner,” wrote Tracey G. Simon, MD, MPH, and her associates. Their report is in Clinical Gastroenterology and Hepatology.

The study comprised 361 adults with biopsy-confirmed NAFLD who were enrolled in the Massachusetts General Hospital NAFLD Repository between 2006 and 2015. At baseline, 151 individuals were already on daily aspirin, usually to reduce the primary (54%) or secondary (30%) risk of cardiovascular disease. Median duration of aspirin use was 2.5 years. After a median 7.4 years of follow-up (which was similar between aspirin users and nonusers), daily aspirin use was associated with significantly lower odds of NASH (adjusted odds ratio, 0.68; 95% confidence interval, 0.37-0.89) and fibrosis (aOR, 0.54; 95% CI, 0.31-0.82).

The researchers did not find a similar protective effect for nonsteroidal anti-inflammatory drugs (NSAIDs) other than aspirin (adjusted hazard ratio for advanced fibrosis, 0.93; 95% CI, 0.81–1.05). This might be because of differences between how aspirin and nonaspirin NSAIDs affect COX isoforms – aspirin does so irreversibly, while other NSAIDs have a reversible effect, they added. “Nonaspirin NSAIDs also disrupt the intestinal barrier, increasing delivery of proinflammatory cytokines to the liver,” they wrote. “Finally, aspirin uniquely modulates bioactive lipids by stimulating the biosynthesis of pro-resolving mediators and inhibiting proinflammatory lipids, which in turn may prevent progressive liver damage.”

In this study, a single blinded hepatopathologist interpreted baseline liver biopsy specimens, and patients were followed every 3-6 months with clinical examinations and serial calculations of FIB-4, NFS, and APRI scores. All patients were followed for at least a year. Patients were classified as users of nonaspirin NSAIDs if they reported using an NSAID besides aspirin at least twice weekly, or if they had been prescribed drugs such as ibuprofen, naproxen, ketoprofen, diclofenac, or indomethacin.

In a longitudinal analysis of the 317 patients who had early-stage (F0-2) fibrosis at baseline, 86 developed new-onset advanced fibrosis over a median of 3,692 person-years, the researchers said. In all, 26 individuals developed hepatic decompensation and 18 patients died, including eight from liver-related causes. Importantly, the link between aspirin and decreased risk of fibrosis progression seemed to depend on duration of use (adjusted P trend = .026), with the greatest benefit seen with 4 years or more of use (aHR, 0.50; 95% CI, 0.35-0.73). Although subgroup analyses were limited by lack of power, daily aspirin use was associated with a 36% lower odds of incident advanced fibrosis among the 72 study participants who had paired biopsy samples, even after accounting for the effect of age, sex, baseline fibrosis stage, and time between biopsies (aOR, 0.64; 95% CI, 0.50-0.80).

“Our findings add to the growing literature supporting the potential hepatoprotective effects of aspirin in NAFLD,” the researchers concluded. “Research to uncover the mechanisms by which aspirin might prevent fibrogenesis could help develop urgently needed antifibrotic therapies for NAFLD.”

Funders included the National Institutes of Health and the AASLD Foundation. The investigators reported having no conflicts of interest.

SOURCE: Simon TG et al. Clin Gastroenterol Hepatol. 2019 May 8.

Taking daily aspirin may help keep nonalcoholic fatty liver disease (NAFLD) from progressing to liver fibrosis and nonalcoholic steatohepatitis (NASH), suggest the results of a prospective study of 361 adults.

Previously, preclinical evidence had linked aspirin to fibrogenesis prevention in fatty liver disease, but this is the first report of a prospective study to do so. Daily aspirin use “was associated with less severe histologic features of NAFLD (nonalcoholic fatty liver disease) at study enrollment and with significantly lower risk for advanced fibrosis over time in a duration-dependent manner,” wrote Tracey G. Simon, MD, MPH, and her associates. Their report is in Clinical Gastroenterology and Hepatology.

The study comprised 361 adults with biopsy-confirmed NAFLD who were enrolled in the Massachusetts General Hospital NAFLD Repository between 2006 and 2015. At baseline, 151 individuals were already on daily aspirin, usually to reduce the primary (54%) or secondary (30%) risk of cardiovascular disease. Median duration of aspirin use was 2.5 years. After a median 7.4 years of follow-up (which was similar between aspirin users and nonusers), daily aspirin use was associated with significantly lower odds of NASH (adjusted odds ratio, 0.68; 95% confidence interval, 0.37-0.89) and fibrosis (aOR, 0.54; 95% CI, 0.31-0.82).

The researchers did not find a similar protective effect for nonsteroidal anti-inflammatory drugs (NSAIDs) other than aspirin (adjusted hazard ratio for advanced fibrosis, 0.93; 95% CI, 0.81–1.05). This might be because of differences between how aspirin and nonaspirin NSAIDs affect COX isoforms – aspirin does so irreversibly, while other NSAIDs have a reversible effect, they added. “Nonaspirin NSAIDs also disrupt the intestinal barrier, increasing delivery of proinflammatory cytokines to the liver,” they wrote. “Finally, aspirin uniquely modulates bioactive lipids by stimulating the biosynthesis of pro-resolving mediators and inhibiting proinflammatory lipids, which in turn may prevent progressive liver damage.”

In this study, a single blinded hepatopathologist interpreted baseline liver biopsy specimens, and patients were followed every 3-6 months with clinical examinations and serial calculations of FIB-4, NFS, and APRI scores. All patients were followed for at least a year. Patients were classified as users of nonaspirin NSAIDs if they reported using an NSAID besides aspirin at least twice weekly, or if they had been prescribed drugs such as ibuprofen, naproxen, ketoprofen, diclofenac, or indomethacin.

In a longitudinal analysis of the 317 patients who had early-stage (F0-2) fibrosis at baseline, 86 developed new-onset advanced fibrosis over a median of 3,692 person-years, the researchers said. In all, 26 individuals developed hepatic decompensation and 18 patients died, including eight from liver-related causes. Importantly, the link between aspirin and decreased risk of fibrosis progression seemed to depend on duration of use (adjusted P trend = .026), with the greatest benefit seen with 4 years or more of use (aHR, 0.50; 95% CI, 0.35-0.73). Although subgroup analyses were limited by lack of power, daily aspirin use was associated with a 36% lower odds of incident advanced fibrosis among the 72 study participants who had paired biopsy samples, even after accounting for the effect of age, sex, baseline fibrosis stage, and time between biopsies (aOR, 0.64; 95% CI, 0.50-0.80).

“Our findings add to the growing literature supporting the potential hepatoprotective effects of aspirin in NAFLD,” the researchers concluded. “Research to uncover the mechanisms by which aspirin might prevent fibrogenesis could help develop urgently needed antifibrotic therapies for NAFLD.”

Funders included the National Institutes of Health and the AASLD Foundation. The investigators reported having no conflicts of interest.

SOURCE: Simon TG et al. Clin Gastroenterol Hepatol. 2019 May 8.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CLINICAL GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Managing psychosis in youth

Article Type
Changed
Fri, 11/08/2019 - 14:52

Strong feelings – such as intense anxiety, irritability, or depressed mood – may affect every child for brief periods of time during their development. Parents and pediatricians are wise to not treat them as psychiatric disorders unless they persist for weeks, impair functioning, or are dramatically severe. Psychosis – marked by hallucinations, perceptual distortions, or profoundly disorganized thinking and behavior – typically looks dramatically severe. Even when psychotic symptoms are mild or brief, they can cause very serious distress for parents and clinicians. The worry is that they may represent a “first break,” a psychotic episode that requires much work for recovery, or the beginning of a lifelong struggle with schizophrenia or other chronic psychotic illness.

KatarzynaBialasiewicz/iStock/Getty Images Plus

While it is important to recognize schizophrenia early – because early interventions are thought to improve the course of the disease – schizophrenia in childhood is rare. It is not commonly recognized that psychotic or psychoticlike symptoms are much more common than schizophrenia. In childhood, psychosis is three times more common than in adults, and can indicate a number of different psychiatric or medical problems. While it is important to begin a thoughtful evaluation when a child or teenager presents with psychosis, it also is important to know that the majority of young people who experience psychotic symptoms do not have schizophrenia or other psychotic illness.

Psychosis describes symptoms in which there has been some “break with reality,” often in the form of hallucinations (seeing or hearing things which are not objectively present) or of distorted perceptions (such as paranoia or grandiosity). “Subsyndromal psychotic symptoms” occur when a person experiences these perceptual disturbances but has doubt about whether or not they are real. In frank psychosis, patients have a “fixed and firm” belief in the truth or accuracy of their perceptions, no matter the evidence against them. The voices they hear or hallucinations they see are “real” and there is a wholehearted belief that what the voice says or what they are seeing is as true as what you or I see and hear.

Schizophrenia is a diagnosis that requires the presence of both these “positive” psychotic symptoms and “negative” symptoms of flat affect; loss of motivation, social, or motor abilities; and cognitive impairment. These symptoms typically emerge in late adolescence (median age, 18 years) in males and early adulthood (median age, 25 years) in females, with another (smaller) peak in incidence in middle age. Importantly, the negative symptoms often emerge first so there often is a history of subtle cognitive decline and social withdrawal, one of the most common patterns in children, before psychosis emerges. Schizophrenia is quite rare, with a prevalence of slightly under 1% of the global population, an annual incidence of approximately 15 people per 100,000, and 1 in 40,000 in children under 13 years old, according to the National Institute of Mental Health. Psychotic symptoms are much more common than schizophrenia, affecting approximately 5% of the adult population at any point in time. They are even more common in children and adolescents. A meta-analysis of population-based studies of psychotic symptoms in youth demonstrated a median prevalence of 17% in children aged 9-12 years and 7.5% in adolescents aged 13-18 years.1 Of course, as with all statistics, much depends on the definitions used to identify this high prevalence rate.

Dr. Susan D. Swick, physician in chief at Ohana,Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula.
Dr. Susan D. Swick

Children and adolescents who report psychotic symptoms are at increased risk for developing schizophrenia, compared with the general population, but most youth with psychotic symptoms will not go on to develop schizophrenia. They are more likely to indicate other, nonpsychotic psychiatric illnesses, such as anxiety or mood disorders, including depression, obsessive compulsive disorder (OCD), and PTSD. In younger children, these symptoms may prove to be benign, but in adolescents they usually indicate the presence of a psychiatric illness. In one study, 57% of children aged 11-13 years with psychotic symptoms were found to have a nonpsychotic psychiatric illness, but the rate jumped to 80% for those aged 13-15 years with psychotic symptoms.2 So while psychosis in teenagers only rarely indicates schizophrenia, these symptoms usually indicate the presence of a psychiatric illness, and a psychiatric evaluation should be initiated.

If a child in your practice presents with psychotic symptoms, it is appropriate to assess their safety and then start a medical work-up. Find out from your patient or their parents if their behavior has been affected by their perceptual disturbances. Are they frightened and avoiding school? Are they withdrawing from social relationships? Is their sleep disrupted? Have they been more impulsive or unpredictable? If their behavior has been affected, you should refer to a child psychiatrist to perform a full diagnostic evaluation and help with management of these symptoms.

Your medical work-up should include a drug screen, blood count, metabolic panel, and thyroid function test. Medications, particularly stimulants, steroids, and anticholinergics can cause psychotic symptoms in high doses or vulnerable patients (such as those with a developmental disorder or traumatic brain injury). If the physical or neurologic exam are suggestive, further investigation of the many potential medical sources of psychotic symptoms in youth can be pursued to rule out autoimmune illnesses, endocrine disorders, metabolic illnesses, heavy metal poisoning, neurologic diseases, infectious diseases, and nutritional deficits. It is worth noting that childhood sleep disorders also can present with psychosis. Persistent psychotic symptoms in children are very hard to evaluate and may be the harbinger of a serious psychiatric disorder, so even if the medical work-up is negative and the persistent symptoms are mild and not causing a safety concern, a referral to a child psychiatrist for a full mental health evaluation is appropriate.

Dr. Michael S. Jellinek, professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston
Dr. Michael S. Jellinek

Psychotic symptoms in an adolescent sometimes are easier to assess, more worrisome for serious mental illness, and are a high-risk category for self-destructive behavior and substance use. Before you begin a medical work-up, you always should assess for safety, including suicide risk, if your adolescent patient presents with psychotic symptoms. Screening for symptoms of mood, anxiety, and substance use disorders also can help reveal the nature of their presenting problem. If your adolescent patient is using drugs, that does not rule out the possibility of an underlying mood, anxiety, or thought disorder. While intoxication with many drugs may precipitate psychotic symptoms (including stimulants, hallucinogens, and marijuana), others may precipitate psychosis in withdrawal states (alcohol, benzodiazepines, and other CNS depressants). It also is important to note that adolescents with emerging schizophrenia have very high rates of comorbid substance abuse (as high as 60%), so their drug use may not be the cause of their psychotic symptoms. There also is emerging evidence that use of certain drugs during sensitive developmental periods can significantly increase the likelihood of developing schizophrenia in vulnerable populations, such as with regular marijuana use in adolescents who have a family history of schizophrenia.

For those rare pediatric patients who present with both negative and positive symptoms of emerging schizophrenia, early diagnosis and treatment has shown promise in improving the course of the disease. Cognitive-behavioral therapy (CBT) for psychosis has shown promise in lowering the rates of conversion to schizophrenia in select patient populations. This therapy teaches strategies for improving reality testing, cognitive flexibility, and social skills. The social skills appear to be especially important for improving adaptive function, even in those patients who progress to schizophrenia. Family therapies, focused on improving family cohesion, communication, and adaptive functioning, appear to improve family well-being and the course of the patient’s illness (such as fewer and less severe psychotic episodes and improved mood and adaptive function). Early use of antipsychotic medications also appears to improve the course of the illness.

While schizophrenia is not curable, early detection (perhaps by a pediatrician), referral, and treatment can be powerfully protective for patients and their families.
 

Dr. Swick is physician in chief at Ohana, Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula. Dr. Jellinek is professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston. Email them at pdnews@mdedge.com.

References

1. Psychol Med. 2012 Sep;42(9):1857-63.

2. Br J Psychiatry. 2012 Jul;201(1):26-32.

Publications
Topics
Sections

Strong feelings – such as intense anxiety, irritability, or depressed mood – may affect every child for brief periods of time during their development. Parents and pediatricians are wise to not treat them as psychiatric disorders unless they persist for weeks, impair functioning, or are dramatically severe. Psychosis – marked by hallucinations, perceptual distortions, or profoundly disorganized thinking and behavior – typically looks dramatically severe. Even when psychotic symptoms are mild or brief, they can cause very serious distress for parents and clinicians. The worry is that they may represent a “first break,” a psychotic episode that requires much work for recovery, or the beginning of a lifelong struggle with schizophrenia or other chronic psychotic illness.

KatarzynaBialasiewicz/iStock/Getty Images Plus

While it is important to recognize schizophrenia early – because early interventions are thought to improve the course of the disease – schizophrenia in childhood is rare. It is not commonly recognized that psychotic or psychoticlike symptoms are much more common than schizophrenia. In childhood, psychosis is three times more common than in adults, and can indicate a number of different psychiatric or medical problems. While it is important to begin a thoughtful evaluation when a child or teenager presents with psychosis, it also is important to know that the majority of young people who experience psychotic symptoms do not have schizophrenia or other psychotic illness.

Psychosis describes symptoms in which there has been some “break with reality,” often in the form of hallucinations (seeing or hearing things which are not objectively present) or of distorted perceptions (such as paranoia or grandiosity). “Subsyndromal psychotic symptoms” occur when a person experiences these perceptual disturbances but has doubt about whether or not they are real. In frank psychosis, patients have a “fixed and firm” belief in the truth or accuracy of their perceptions, no matter the evidence against them. The voices they hear or hallucinations they see are “real” and there is a wholehearted belief that what the voice says or what they are seeing is as true as what you or I see and hear.

Schizophrenia is a diagnosis that requires the presence of both these “positive” psychotic symptoms and “negative” symptoms of flat affect; loss of motivation, social, or motor abilities; and cognitive impairment. These symptoms typically emerge in late adolescence (median age, 18 years) in males and early adulthood (median age, 25 years) in females, with another (smaller) peak in incidence in middle age. Importantly, the negative symptoms often emerge first so there often is a history of subtle cognitive decline and social withdrawal, one of the most common patterns in children, before psychosis emerges. Schizophrenia is quite rare, with a prevalence of slightly under 1% of the global population, an annual incidence of approximately 15 people per 100,000, and 1 in 40,000 in children under 13 years old, according to the National Institute of Mental Health. Psychotic symptoms are much more common than schizophrenia, affecting approximately 5% of the adult population at any point in time. They are even more common in children and adolescents. A meta-analysis of population-based studies of psychotic symptoms in youth demonstrated a median prevalence of 17% in children aged 9-12 years and 7.5% in adolescents aged 13-18 years.1 Of course, as with all statistics, much depends on the definitions used to identify this high prevalence rate.

Dr. Susan D. Swick, physician in chief at Ohana,Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula.
Dr. Susan D. Swick

Children and adolescents who report psychotic symptoms are at increased risk for developing schizophrenia, compared with the general population, but most youth with psychotic symptoms will not go on to develop schizophrenia. They are more likely to indicate other, nonpsychotic psychiatric illnesses, such as anxiety or mood disorders, including depression, obsessive compulsive disorder (OCD), and PTSD. In younger children, these symptoms may prove to be benign, but in adolescents they usually indicate the presence of a psychiatric illness. In one study, 57% of children aged 11-13 years with psychotic symptoms were found to have a nonpsychotic psychiatric illness, but the rate jumped to 80% for those aged 13-15 years with psychotic symptoms.2 So while psychosis in teenagers only rarely indicates schizophrenia, these symptoms usually indicate the presence of a psychiatric illness, and a psychiatric evaluation should be initiated.

If a child in your practice presents with psychotic symptoms, it is appropriate to assess their safety and then start a medical work-up. Find out from your patient or their parents if their behavior has been affected by their perceptual disturbances. Are they frightened and avoiding school? Are they withdrawing from social relationships? Is their sleep disrupted? Have they been more impulsive or unpredictable? If their behavior has been affected, you should refer to a child psychiatrist to perform a full diagnostic evaluation and help with management of these symptoms.

Your medical work-up should include a drug screen, blood count, metabolic panel, and thyroid function test. Medications, particularly stimulants, steroids, and anticholinergics can cause psychotic symptoms in high doses or vulnerable patients (such as those with a developmental disorder or traumatic brain injury). If the physical or neurologic exam are suggestive, further investigation of the many potential medical sources of psychotic symptoms in youth can be pursued to rule out autoimmune illnesses, endocrine disorders, metabolic illnesses, heavy metal poisoning, neurologic diseases, infectious diseases, and nutritional deficits. It is worth noting that childhood sleep disorders also can present with psychosis. Persistent psychotic symptoms in children are very hard to evaluate and may be the harbinger of a serious psychiatric disorder, so even if the medical work-up is negative and the persistent symptoms are mild and not causing a safety concern, a referral to a child psychiatrist for a full mental health evaluation is appropriate.

Dr. Michael S. Jellinek, professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston
Dr. Michael S. Jellinek

Psychotic symptoms in an adolescent sometimes are easier to assess, more worrisome for serious mental illness, and are a high-risk category for self-destructive behavior and substance use. Before you begin a medical work-up, you always should assess for safety, including suicide risk, if your adolescent patient presents with psychotic symptoms. Screening for symptoms of mood, anxiety, and substance use disorders also can help reveal the nature of their presenting problem. If your adolescent patient is using drugs, that does not rule out the possibility of an underlying mood, anxiety, or thought disorder. While intoxication with many drugs may precipitate psychotic symptoms (including stimulants, hallucinogens, and marijuana), others may precipitate psychosis in withdrawal states (alcohol, benzodiazepines, and other CNS depressants). It also is important to note that adolescents with emerging schizophrenia have very high rates of comorbid substance abuse (as high as 60%), so their drug use may not be the cause of their psychotic symptoms. There also is emerging evidence that use of certain drugs during sensitive developmental periods can significantly increase the likelihood of developing schizophrenia in vulnerable populations, such as with regular marijuana use in adolescents who have a family history of schizophrenia.

For those rare pediatric patients who present with both negative and positive symptoms of emerging schizophrenia, early diagnosis and treatment has shown promise in improving the course of the disease. Cognitive-behavioral therapy (CBT) for psychosis has shown promise in lowering the rates of conversion to schizophrenia in select patient populations. This therapy teaches strategies for improving reality testing, cognitive flexibility, and social skills. The social skills appear to be especially important for improving adaptive function, even in those patients who progress to schizophrenia. Family therapies, focused on improving family cohesion, communication, and adaptive functioning, appear to improve family well-being and the course of the patient’s illness (such as fewer and less severe psychotic episodes and improved mood and adaptive function). Early use of antipsychotic medications also appears to improve the course of the illness.

While schizophrenia is not curable, early detection (perhaps by a pediatrician), referral, and treatment can be powerfully protective for patients and their families.
 

Dr. Swick is physician in chief at Ohana, Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula. Dr. Jellinek is professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston. Email them at pdnews@mdedge.com.

References

1. Psychol Med. 2012 Sep;42(9):1857-63.

2. Br J Psychiatry. 2012 Jul;201(1):26-32.

Strong feelings – such as intense anxiety, irritability, or depressed mood – may affect every child for brief periods of time during their development. Parents and pediatricians are wise to not treat them as psychiatric disorders unless they persist for weeks, impair functioning, or are dramatically severe. Psychosis – marked by hallucinations, perceptual distortions, or profoundly disorganized thinking and behavior – typically looks dramatically severe. Even when psychotic symptoms are mild or brief, they can cause very serious distress for parents and clinicians. The worry is that they may represent a “first break,” a psychotic episode that requires much work for recovery, or the beginning of a lifelong struggle with schizophrenia or other chronic psychotic illness.

KatarzynaBialasiewicz/iStock/Getty Images Plus

While it is important to recognize schizophrenia early – because early interventions are thought to improve the course of the disease – schizophrenia in childhood is rare. It is not commonly recognized that psychotic or psychoticlike symptoms are much more common than schizophrenia. In childhood, psychosis is three times more common than in adults, and can indicate a number of different psychiatric or medical problems. While it is important to begin a thoughtful evaluation when a child or teenager presents with psychosis, it also is important to know that the majority of young people who experience psychotic symptoms do not have schizophrenia or other psychotic illness.

Psychosis describes symptoms in which there has been some “break with reality,” often in the form of hallucinations (seeing or hearing things which are not objectively present) or of distorted perceptions (such as paranoia or grandiosity). “Subsyndromal psychotic symptoms” occur when a person experiences these perceptual disturbances but has doubt about whether or not they are real. In frank psychosis, patients have a “fixed and firm” belief in the truth or accuracy of their perceptions, no matter the evidence against them. The voices they hear or hallucinations they see are “real” and there is a wholehearted belief that what the voice says or what they are seeing is as true as what you or I see and hear.

Schizophrenia is a diagnosis that requires the presence of both these “positive” psychotic symptoms and “negative” symptoms of flat affect; loss of motivation, social, or motor abilities; and cognitive impairment. These symptoms typically emerge in late adolescence (median age, 18 years) in males and early adulthood (median age, 25 years) in females, with another (smaller) peak in incidence in middle age. Importantly, the negative symptoms often emerge first so there often is a history of subtle cognitive decline and social withdrawal, one of the most common patterns in children, before psychosis emerges. Schizophrenia is quite rare, with a prevalence of slightly under 1% of the global population, an annual incidence of approximately 15 people per 100,000, and 1 in 40,000 in children under 13 years old, according to the National Institute of Mental Health. Psychotic symptoms are much more common than schizophrenia, affecting approximately 5% of the adult population at any point in time. They are even more common in children and adolescents. A meta-analysis of population-based studies of psychotic symptoms in youth demonstrated a median prevalence of 17% in children aged 9-12 years and 7.5% in adolescents aged 13-18 years.1 Of course, as with all statistics, much depends on the definitions used to identify this high prevalence rate.

Dr. Susan D. Swick, physician in chief at Ohana,Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula.
Dr. Susan D. Swick

Children and adolescents who report psychotic symptoms are at increased risk for developing schizophrenia, compared with the general population, but most youth with psychotic symptoms will not go on to develop schizophrenia. They are more likely to indicate other, nonpsychotic psychiatric illnesses, such as anxiety or mood disorders, including depression, obsessive compulsive disorder (OCD), and PTSD. In younger children, these symptoms may prove to be benign, but in adolescents they usually indicate the presence of a psychiatric illness. In one study, 57% of children aged 11-13 years with psychotic symptoms were found to have a nonpsychotic psychiatric illness, but the rate jumped to 80% for those aged 13-15 years with psychotic symptoms.2 So while psychosis in teenagers only rarely indicates schizophrenia, these symptoms usually indicate the presence of a psychiatric illness, and a psychiatric evaluation should be initiated.

If a child in your practice presents with psychotic symptoms, it is appropriate to assess their safety and then start a medical work-up. Find out from your patient or their parents if their behavior has been affected by their perceptual disturbances. Are they frightened and avoiding school? Are they withdrawing from social relationships? Is their sleep disrupted? Have they been more impulsive or unpredictable? If their behavior has been affected, you should refer to a child psychiatrist to perform a full diagnostic evaluation and help with management of these symptoms.

Your medical work-up should include a drug screen, blood count, metabolic panel, and thyroid function test. Medications, particularly stimulants, steroids, and anticholinergics can cause psychotic symptoms in high doses or vulnerable patients (such as those with a developmental disorder or traumatic brain injury). If the physical or neurologic exam are suggestive, further investigation of the many potential medical sources of psychotic symptoms in youth can be pursued to rule out autoimmune illnesses, endocrine disorders, metabolic illnesses, heavy metal poisoning, neurologic diseases, infectious diseases, and nutritional deficits. It is worth noting that childhood sleep disorders also can present with psychosis. Persistent psychotic symptoms in children are very hard to evaluate and may be the harbinger of a serious psychiatric disorder, so even if the medical work-up is negative and the persistent symptoms are mild and not causing a safety concern, a referral to a child psychiatrist for a full mental health evaluation is appropriate.

Dr. Michael S. Jellinek, professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston
Dr. Michael S. Jellinek

Psychotic symptoms in an adolescent sometimes are easier to assess, more worrisome for serious mental illness, and are a high-risk category for self-destructive behavior and substance use. Before you begin a medical work-up, you always should assess for safety, including suicide risk, if your adolescent patient presents with psychotic symptoms. Screening for symptoms of mood, anxiety, and substance use disorders also can help reveal the nature of their presenting problem. If your adolescent patient is using drugs, that does not rule out the possibility of an underlying mood, anxiety, or thought disorder. While intoxication with many drugs may precipitate psychotic symptoms (including stimulants, hallucinogens, and marijuana), others may precipitate psychosis in withdrawal states (alcohol, benzodiazepines, and other CNS depressants). It also is important to note that adolescents with emerging schizophrenia have very high rates of comorbid substance abuse (as high as 60%), so their drug use may not be the cause of their psychotic symptoms. There also is emerging evidence that use of certain drugs during sensitive developmental periods can significantly increase the likelihood of developing schizophrenia in vulnerable populations, such as with regular marijuana use in adolescents who have a family history of schizophrenia.

For those rare pediatric patients who present with both negative and positive symptoms of emerging schizophrenia, early diagnosis and treatment has shown promise in improving the course of the disease. Cognitive-behavioral therapy (CBT) for psychosis has shown promise in lowering the rates of conversion to schizophrenia in select patient populations. This therapy teaches strategies for improving reality testing, cognitive flexibility, and social skills. The social skills appear to be especially important for improving adaptive function, even in those patients who progress to schizophrenia. Family therapies, focused on improving family cohesion, communication, and adaptive functioning, appear to improve family well-being and the course of the patient’s illness (such as fewer and less severe psychotic episodes and improved mood and adaptive function). Early use of antipsychotic medications also appears to improve the course of the illness.

While schizophrenia is not curable, early detection (perhaps by a pediatrician), referral, and treatment can be powerfully protective for patients and their families.
 

Dr. Swick is physician in chief at Ohana, Center for Child and Adolescent Behavioral Health, Community Hospital of the Monterey (Calif.) Peninsula. Dr. Jellinek is professor emeritus of psychiatry and pediatrics, Harvard Medical School, Boston. Email them at pdnews@mdedge.com.

References

1. Psychol Med. 2012 Sep;42(9):1857-63.

2. Br J Psychiatry. 2012 Jul;201(1):26-32.

Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

A progressive exercise intervention improved AGFR in breast cancer survivors

Article Type
Changed
Wed, 01/04/2023 - 16:43

 

– A progressive aerobic and resistance exercise intervention improved the android:gynoid fat ratio (AGFR) in breast cancer survivors, which could provide important health benefits.

AGFR is associated with increased risk for cardiovascular disease and type 2 diabetes in breast cancer survivors, therefore exercise-induced AGFR improvement may reduce the risk for such comorbid conditions, Christina Dieli-Conwright, PhD, of the University of Southern California Norris Comprehensive Cancer Center, Los Angeles, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

A significant decrease in AGFR from baseline was noted in 50 survivors of stage I-III breast cancer who participated in the exercise intervention, compared with 50 such survivors randomized to a usual care group (P less than .001), and strong correlations were found between AGFR and homeostatic model assessment of insulin resistance (HOMA-IR; r = 0.95; P less than .01), the investigators found.

Study participants had a mean age of 53 years, 54% were overweight (body mass index greater than 25 kg/m2), 63% were Hispanic, 90% had undergone a mastectomy, and 76% received chemotherapy and radiation therapy. Adherence to the intervention, which involved three weekly sessions of supervised, progressive, moderate-to-vigorous aerobic and resistance exercise for 16 weeks, was 95%.

AGFR was calculated using whole-body dual-energy x-ray absorptiometry and HOMA-IR was calculated using fasting insulin and glucose levels.

“Exercise reduces fat mass in breast cancer survivors, however, few studies have focused on AGFR,” the investigators wrote.

The findings of the current study suggest that a progressive aerobic and resistance exercise intervention is an effective strategy for decreasing AGFR in breast cancer survivors, they concluded.

The National Cancer Institute funded the study. The authors reported having no disclosures.

SOURCE: Dieli-Conwright C et al. ESMO 2019, Abstract 228P.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

 

– A progressive aerobic and resistance exercise intervention improved the android:gynoid fat ratio (AGFR) in breast cancer survivors, which could provide important health benefits.

AGFR is associated with increased risk for cardiovascular disease and type 2 diabetes in breast cancer survivors, therefore exercise-induced AGFR improvement may reduce the risk for such comorbid conditions, Christina Dieli-Conwright, PhD, of the University of Southern California Norris Comprehensive Cancer Center, Los Angeles, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

A significant decrease in AGFR from baseline was noted in 50 survivors of stage I-III breast cancer who participated in the exercise intervention, compared with 50 such survivors randomized to a usual care group (P less than .001), and strong correlations were found between AGFR and homeostatic model assessment of insulin resistance (HOMA-IR; r = 0.95; P less than .01), the investigators found.

Study participants had a mean age of 53 years, 54% were overweight (body mass index greater than 25 kg/m2), 63% were Hispanic, 90% had undergone a mastectomy, and 76% received chemotherapy and radiation therapy. Adherence to the intervention, which involved three weekly sessions of supervised, progressive, moderate-to-vigorous aerobic and resistance exercise for 16 weeks, was 95%.

AGFR was calculated using whole-body dual-energy x-ray absorptiometry and HOMA-IR was calculated using fasting insulin and glucose levels.

“Exercise reduces fat mass in breast cancer survivors, however, few studies have focused on AGFR,” the investigators wrote.

The findings of the current study suggest that a progressive aerobic and resistance exercise intervention is an effective strategy for decreasing AGFR in breast cancer survivors, they concluded.

The National Cancer Institute funded the study. The authors reported having no disclosures.

SOURCE: Dieli-Conwright C et al. ESMO 2019, Abstract 228P.

 

– A progressive aerobic and resistance exercise intervention improved the android:gynoid fat ratio (AGFR) in breast cancer survivors, which could provide important health benefits.

AGFR is associated with increased risk for cardiovascular disease and type 2 diabetes in breast cancer survivors, therefore exercise-induced AGFR improvement may reduce the risk for such comorbid conditions, Christina Dieli-Conwright, PhD, of the University of Southern California Norris Comprehensive Cancer Center, Los Angeles, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

A significant decrease in AGFR from baseline was noted in 50 survivors of stage I-III breast cancer who participated in the exercise intervention, compared with 50 such survivors randomized to a usual care group (P less than .001), and strong correlations were found between AGFR and homeostatic model assessment of insulin resistance (HOMA-IR; r = 0.95; P less than .01), the investigators found.

Study participants had a mean age of 53 years, 54% were overweight (body mass index greater than 25 kg/m2), 63% were Hispanic, 90% had undergone a mastectomy, and 76% received chemotherapy and radiation therapy. Adherence to the intervention, which involved three weekly sessions of supervised, progressive, moderate-to-vigorous aerobic and resistance exercise for 16 weeks, was 95%.

AGFR was calculated using whole-body dual-energy x-ray absorptiometry and HOMA-IR was calculated using fasting insulin and glucose levels.

“Exercise reduces fat mass in breast cancer survivors, however, few studies have focused on AGFR,” the investigators wrote.

The findings of the current study suggest that a progressive aerobic and resistance exercise intervention is an effective strategy for decreasing AGFR in breast cancer survivors, they concluded.

The National Cancer Institute funded the study. The authors reported having no disclosures.

SOURCE: Dieli-Conwright C et al. ESMO 2019, Abstract 228P.

Publications
Publications
Topics
Article Type
Sections
Article Source

REPORTING FROM ESMO 2019

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Was the success of hospital medicine inevitable?

Article Type
Changed
Fri, 11/08/2019 - 14:28

Early on, SHM defined the specialty

 

When I started at the Society of Hospital Medicine – known then as the National Association of Inpatient Physicians (NAIP) – in January 2000, Bill Clinton was still president. There were probably 500 hospitalists in the United States, and SHM had about 200-250 members.

Dr. Larry Wellikson
Dr. Larry Wellikson

It was so long ago that the iPhone hadn’t been invented, Twitter wasn’t even an idea, and Amazon was an online book store. SHM’s national offices were a cubicle at the American College of Physicians headquarters in Philadelphia, and our entire staff was me and a part-time assistant.

We have certainly come a long way in my 20 years as CEO of SHM.

When I first became involved with NAIP, it was to help the board with their strategic planning in 1998. At that time, the national thought leaders for the hospitalist movement (the term hospital medicine had not been invented yet) predicted that hospitalists would eventually do the inpatient work for about 25% of family doctors and for 15% of internists. Hospitalists were considered to be a form of “general medicine” without an office-based practice.

One of the first things we set about doing was to define the new specialty of hospital medicine before anyone else (e.g., American Medical Association, ACP, American Academy of Family Physicians, American Academy of Pediatrics, the government) defined us.

Most specialties were defined by a body organ (e.g., cardiology, renal), a population (e.g., pediatrics, geriatrics), or a disease (e.g., oncology), and there were a few other site-specific specialties (e.g., ED medicine, critical care). We felt that, to be a specialty, we needed certain key elements:

  • Separate group consciousness
  • Professional society
  • Distinct residency and fellowship programs
  • Separate CME
  • Distinct educational materials (e.g., textbooks)
  • Definable and distinct competencies
  • Separate credentials – certification and/or hospital insurance driven

Early on, SHM defined the Core Competencies for Hospital Medicine for adults in patient care and, eventually, for pediatric patients. We rebranded our specialty as hospital medicine to be more than just inpatient physicians, and to broadly encompass the growing “big tent” of SHM that included those trained in internal medicine, family medicine, pediatrics, med-peds, as well as nurse practitioners, physician assistants, pharmacists, and others.

We were the first and only specialty society to set the standard for hospitalist compensation (how much you are paid) and productivity (what you are expected to do) with our unique State of Hospital Medicine (SOHM) Report. Other specialties left this work to the Medical Group Management Association, the AMA, or commercial companies.

Our specialty was soon being asked to do things that no other group of clinicians was ever asked to do.

Hospitalists were expected to Save Money by reducing length of stay and the use of resources on the sickest patients. Hospitalists were asked to Improve Measurable Quality at a time when most other physicians or even hospitals weren’t even being measured.

We were expected to form and Lead Teams of other clinicians when health care was still seen as a solo enterprise. Hospitalists were expected to Improve Efficiency and to create a Seamless Continuity, both during the hospital stay and in the transitions out of the hospital.

Hospitalists were asked to do things no one else wanted to do, such as taking on the uncompensated patients and extra hospital committee work and just about any new project their hospital wanted to be involved in. Along the way, we were expected to Make Other Physicians’ Lives Better by taking on their inpatients, inpatient calls, comanagement with specialists, and unloading the ED.

And both at medical schools and in the community, hospitalists became the Major Educators of medical students, residents, nurses, and other hospital staff.

At the same time, SHM was focusing on becoming a very unique medical professional society.

SHM built on the energy of our young and innovative hospitalists to forge a different path. We had no reputation to protect. We were not bound like most other specialty societies to over 100 years of “the way it’s always been done.”

While other professional societies thought their role in quality improvement was to pontificate and publish clinical guidelines that often were little used, SHM embarked on an aggressive, hands-on, frontline approach by starting SHM’s Center for Quality Improvement. Over the last 15 years, the center has raised millions of dollars to deliver real change and improvement at hundreds of hospitals nationwide, many times bringing work plans and mentors to support and train local clinicians in quality improvement skills and data collection. This approach was recognized by the National Quality Forum and the Joint Commission with their prestigious John Eisenberg Award for Quality Improvement.

When we went to Washington to help shape the future of health care, we did not ask for more money for hospitalists. We did not ask for more power or to use regulations to protect our new specialty. Instead, we went with ideas of how to make acute medical care more effective and efficient. We could show the politicians and the regulators how we could reduce incidence of deep vein thrombosis and pulmonary emboli, how we could make the hospital discharge process work better, how we could help chart a smoother medication reconciliation process, and so many other ways the system could be improved.

And even the way SHM generated our new ideas was uniquely different than other specialties. Way back in 2000 – long before Twitter and other social media were able to crowdsource and use the Internet to percolate new ideas – SHM relied on our members’ conversations on the SHM electronic mail discussion list to see what hospitalists were worried about, and what everyone was being asked to do, and SHM provided the resources and initiatives to support our nation’s hospitalists.

From these early conversations, SHM heard that hospitalists were being asked to Lead Change without much of an idea of the skills they would need. And so, the SHM leadership academies were born, which have now educated more than 2,700 hospitalist leaders.

Early on, we learned that hospitalists and even their bosses had no idea of how to start or run a successful hospital medicine group. SHM started our practice management courses and webinars and we developed the groundbreaking document, Key Characteristics of Effective Hospital Medicine Groups. In a typical SHM manner, we challenged most of our members to improve and get better rather trying to defend the status quo. At SHM, we have constantly felt that hospital medicine was a “work in progress.” We may not be perfect today, but we will be better in 90 days and even better in a year.

I have more to say about how we got this far and even more to say about where we might go. So, stay tuned and keep contributing to the future and success of SHM and hospital medicine.

Dr. Wellikson is the CEO of SHM. He has announced his plan to retire from SHM in late 2020. This article is the first in a series celebrating Dr. Wellikson’s tenure as CEO.

Publications
Topics
Sections

Early on, SHM defined the specialty

Early on, SHM defined the specialty

 

When I started at the Society of Hospital Medicine – known then as the National Association of Inpatient Physicians (NAIP) – in January 2000, Bill Clinton was still president. There were probably 500 hospitalists in the United States, and SHM had about 200-250 members.

Dr. Larry Wellikson
Dr. Larry Wellikson

It was so long ago that the iPhone hadn’t been invented, Twitter wasn’t even an idea, and Amazon was an online book store. SHM’s national offices were a cubicle at the American College of Physicians headquarters in Philadelphia, and our entire staff was me and a part-time assistant.

We have certainly come a long way in my 20 years as CEO of SHM.

When I first became involved with NAIP, it was to help the board with their strategic planning in 1998. At that time, the national thought leaders for the hospitalist movement (the term hospital medicine had not been invented yet) predicted that hospitalists would eventually do the inpatient work for about 25% of family doctors and for 15% of internists. Hospitalists were considered to be a form of “general medicine” without an office-based practice.

One of the first things we set about doing was to define the new specialty of hospital medicine before anyone else (e.g., American Medical Association, ACP, American Academy of Family Physicians, American Academy of Pediatrics, the government) defined us.

Most specialties were defined by a body organ (e.g., cardiology, renal), a population (e.g., pediatrics, geriatrics), or a disease (e.g., oncology), and there were a few other site-specific specialties (e.g., ED medicine, critical care). We felt that, to be a specialty, we needed certain key elements:

  • Separate group consciousness
  • Professional society
  • Distinct residency and fellowship programs
  • Separate CME
  • Distinct educational materials (e.g., textbooks)
  • Definable and distinct competencies
  • Separate credentials – certification and/or hospital insurance driven

Early on, SHM defined the Core Competencies for Hospital Medicine for adults in patient care and, eventually, for pediatric patients. We rebranded our specialty as hospital medicine to be more than just inpatient physicians, and to broadly encompass the growing “big tent” of SHM that included those trained in internal medicine, family medicine, pediatrics, med-peds, as well as nurse practitioners, physician assistants, pharmacists, and others.

We were the first and only specialty society to set the standard for hospitalist compensation (how much you are paid) and productivity (what you are expected to do) with our unique State of Hospital Medicine (SOHM) Report. Other specialties left this work to the Medical Group Management Association, the AMA, or commercial companies.

Our specialty was soon being asked to do things that no other group of clinicians was ever asked to do.

Hospitalists were expected to Save Money by reducing length of stay and the use of resources on the sickest patients. Hospitalists were asked to Improve Measurable Quality at a time when most other physicians or even hospitals weren’t even being measured.

We were expected to form and Lead Teams of other clinicians when health care was still seen as a solo enterprise. Hospitalists were expected to Improve Efficiency and to create a Seamless Continuity, both during the hospital stay and in the transitions out of the hospital.

Hospitalists were asked to do things no one else wanted to do, such as taking on the uncompensated patients and extra hospital committee work and just about any new project their hospital wanted to be involved in. Along the way, we were expected to Make Other Physicians’ Lives Better by taking on their inpatients, inpatient calls, comanagement with specialists, and unloading the ED.

And both at medical schools and in the community, hospitalists became the Major Educators of medical students, residents, nurses, and other hospital staff.

At the same time, SHM was focusing on becoming a very unique medical professional society.

SHM built on the energy of our young and innovative hospitalists to forge a different path. We had no reputation to protect. We were not bound like most other specialty societies to over 100 years of “the way it’s always been done.”

While other professional societies thought their role in quality improvement was to pontificate and publish clinical guidelines that often were little used, SHM embarked on an aggressive, hands-on, frontline approach by starting SHM’s Center for Quality Improvement. Over the last 15 years, the center has raised millions of dollars to deliver real change and improvement at hundreds of hospitals nationwide, many times bringing work plans and mentors to support and train local clinicians in quality improvement skills and data collection. This approach was recognized by the National Quality Forum and the Joint Commission with their prestigious John Eisenberg Award for Quality Improvement.

When we went to Washington to help shape the future of health care, we did not ask for more money for hospitalists. We did not ask for more power or to use regulations to protect our new specialty. Instead, we went with ideas of how to make acute medical care more effective and efficient. We could show the politicians and the regulators how we could reduce incidence of deep vein thrombosis and pulmonary emboli, how we could make the hospital discharge process work better, how we could help chart a smoother medication reconciliation process, and so many other ways the system could be improved.

And even the way SHM generated our new ideas was uniquely different than other specialties. Way back in 2000 – long before Twitter and other social media were able to crowdsource and use the Internet to percolate new ideas – SHM relied on our members’ conversations on the SHM electronic mail discussion list to see what hospitalists were worried about, and what everyone was being asked to do, and SHM provided the resources and initiatives to support our nation’s hospitalists.

From these early conversations, SHM heard that hospitalists were being asked to Lead Change without much of an idea of the skills they would need. And so, the SHM leadership academies were born, which have now educated more than 2,700 hospitalist leaders.

Early on, we learned that hospitalists and even their bosses had no idea of how to start or run a successful hospital medicine group. SHM started our practice management courses and webinars and we developed the groundbreaking document, Key Characteristics of Effective Hospital Medicine Groups. In a typical SHM manner, we challenged most of our members to improve and get better rather trying to defend the status quo. At SHM, we have constantly felt that hospital medicine was a “work in progress.” We may not be perfect today, but we will be better in 90 days and even better in a year.

I have more to say about how we got this far and even more to say about where we might go. So, stay tuned and keep contributing to the future and success of SHM and hospital medicine.

Dr. Wellikson is the CEO of SHM. He has announced his plan to retire from SHM in late 2020. This article is the first in a series celebrating Dr. Wellikson’s tenure as CEO.

 

When I started at the Society of Hospital Medicine – known then as the National Association of Inpatient Physicians (NAIP) – in January 2000, Bill Clinton was still president. There were probably 500 hospitalists in the United States, and SHM had about 200-250 members.

Dr. Larry Wellikson
Dr. Larry Wellikson

It was so long ago that the iPhone hadn’t been invented, Twitter wasn’t even an idea, and Amazon was an online book store. SHM’s national offices were a cubicle at the American College of Physicians headquarters in Philadelphia, and our entire staff was me and a part-time assistant.

We have certainly come a long way in my 20 years as CEO of SHM.

When I first became involved with NAIP, it was to help the board with their strategic planning in 1998. At that time, the national thought leaders for the hospitalist movement (the term hospital medicine had not been invented yet) predicted that hospitalists would eventually do the inpatient work for about 25% of family doctors and for 15% of internists. Hospitalists were considered to be a form of “general medicine” without an office-based practice.

One of the first things we set about doing was to define the new specialty of hospital medicine before anyone else (e.g., American Medical Association, ACP, American Academy of Family Physicians, American Academy of Pediatrics, the government) defined us.

Most specialties were defined by a body organ (e.g., cardiology, renal), a population (e.g., pediatrics, geriatrics), or a disease (e.g., oncology), and there were a few other site-specific specialties (e.g., ED medicine, critical care). We felt that, to be a specialty, we needed certain key elements:

  • Separate group consciousness
  • Professional society
  • Distinct residency and fellowship programs
  • Separate CME
  • Distinct educational materials (e.g., textbooks)
  • Definable and distinct competencies
  • Separate credentials – certification and/or hospital insurance driven

Early on, SHM defined the Core Competencies for Hospital Medicine for adults in patient care and, eventually, for pediatric patients. We rebranded our specialty as hospital medicine to be more than just inpatient physicians, and to broadly encompass the growing “big tent” of SHM that included those trained in internal medicine, family medicine, pediatrics, med-peds, as well as nurse practitioners, physician assistants, pharmacists, and others.

We were the first and only specialty society to set the standard for hospitalist compensation (how much you are paid) and productivity (what you are expected to do) with our unique State of Hospital Medicine (SOHM) Report. Other specialties left this work to the Medical Group Management Association, the AMA, or commercial companies.

Our specialty was soon being asked to do things that no other group of clinicians was ever asked to do.

Hospitalists were expected to Save Money by reducing length of stay and the use of resources on the sickest patients. Hospitalists were asked to Improve Measurable Quality at a time when most other physicians or even hospitals weren’t even being measured.

We were expected to form and Lead Teams of other clinicians when health care was still seen as a solo enterprise. Hospitalists were expected to Improve Efficiency and to create a Seamless Continuity, both during the hospital stay and in the transitions out of the hospital.

Hospitalists were asked to do things no one else wanted to do, such as taking on the uncompensated patients and extra hospital committee work and just about any new project their hospital wanted to be involved in. Along the way, we were expected to Make Other Physicians’ Lives Better by taking on their inpatients, inpatient calls, comanagement with specialists, and unloading the ED.

And both at medical schools and in the community, hospitalists became the Major Educators of medical students, residents, nurses, and other hospital staff.

At the same time, SHM was focusing on becoming a very unique medical professional society.

SHM built on the energy of our young and innovative hospitalists to forge a different path. We had no reputation to protect. We were not bound like most other specialty societies to over 100 years of “the way it’s always been done.”

While other professional societies thought their role in quality improvement was to pontificate and publish clinical guidelines that often were little used, SHM embarked on an aggressive, hands-on, frontline approach by starting SHM’s Center for Quality Improvement. Over the last 15 years, the center has raised millions of dollars to deliver real change and improvement at hundreds of hospitals nationwide, many times bringing work plans and mentors to support and train local clinicians in quality improvement skills and data collection. This approach was recognized by the National Quality Forum and the Joint Commission with their prestigious John Eisenberg Award for Quality Improvement.

When we went to Washington to help shape the future of health care, we did not ask for more money for hospitalists. We did not ask for more power or to use regulations to protect our new specialty. Instead, we went with ideas of how to make acute medical care more effective and efficient. We could show the politicians and the regulators how we could reduce incidence of deep vein thrombosis and pulmonary emboli, how we could make the hospital discharge process work better, how we could help chart a smoother medication reconciliation process, and so many other ways the system could be improved.

And even the way SHM generated our new ideas was uniquely different than other specialties. Way back in 2000 – long before Twitter and other social media were able to crowdsource and use the Internet to percolate new ideas – SHM relied on our members’ conversations on the SHM electronic mail discussion list to see what hospitalists were worried about, and what everyone was being asked to do, and SHM provided the resources and initiatives to support our nation’s hospitalists.

From these early conversations, SHM heard that hospitalists were being asked to Lead Change without much of an idea of the skills they would need. And so, the SHM leadership academies were born, which have now educated more than 2,700 hospitalist leaders.

Early on, we learned that hospitalists and even their bosses had no idea of how to start or run a successful hospital medicine group. SHM started our practice management courses and webinars and we developed the groundbreaking document, Key Characteristics of Effective Hospital Medicine Groups. In a typical SHM manner, we challenged most of our members to improve and get better rather trying to defend the status quo. At SHM, we have constantly felt that hospital medicine was a “work in progress.” We may not be perfect today, but we will be better in 90 days and even better in a year.

I have more to say about how we got this far and even more to say about where we might go. So, stay tuned and keep contributing to the future and success of SHM and hospital medicine.

Dr. Wellikson is the CEO of SHM. He has announced his plan to retire from SHM in late 2020. This article is the first in a series celebrating Dr. Wellikson’s tenure as CEO.

Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Bisphosphonates turn 50

Article Type
Changed
Wed, 01/04/2023 - 16:43

– Cumulative evidence over the years has shown that bisphosphonates reduce distant metastases in postmenopausal women with early-stage breast cancer but that effect does not seem to extend to younger, premenopausal women who may experience adverse effects with bisphosphonates, nor has it been replicated in other types of cancer, such as lung or prostate, Robert Coleman MBBS, MD, said in a presentation marking the 50th anniversary of the publication of the first papers describing the then-new class of drugs.

“We have some science – at least in postmenopausal breast cancer – [showing] that we’re really making a difference, but we don’t understand why it’s not working in the other patients or in the other diseases,” said Dr. Coleman of the University of Sheffield (England) at the annual meeting of the American Society for Bone and Mineral Research.
 

Mixed findings in early studies

Early studies showing the metastases-prevention effects of bisphosphonates were with clodronate. In one study, Dr. Coleman said, researchers found that oral clodronate at a dose of 1,600 mg/day as a supplementary treatment to standard treatment for primary, operable breast cancer reduced the risk of bone metastases in a cohort of 1,069 patients with stage 1-3 breast cancer (hazard ratio, 0.692; P = .043) over 5 years (Breast Cancer Res. 2006 Mar 15. doi: 10.1186/bcr1384). Another trial, of 302 patients with breast cancer who received oral clodronate at 1,600 mg/day for 2 years, found that 20.4% of patients who received oral clodronate had died by 8.5 years of follow-up, compared with 40.7% of those who did not receive the intervention (Ann Oncol. 2008;19[12]:2007-11).

However, those results were in conflict with findings from an earlier study in which researchers followed patients with breast cancer who received 1,600 mg/day of oral clodronate or placebo for 3 years. They found a similar number of bone metastases in the clodronate and placebo groups (32% and 29%, respectively), as well as a lower overall disease-free survival rate at 10 years for the clodronate group, compared with the placebo group (45% vs. 58%; Acta Oncol. 2004;43[7]:650-6).“For various other reasons, clodronate did not gain traction as a therapeutic strategy in early-stage breast cancer,” said Dr. Coleman, although emerging evidence showed that other bisphosphonate agents were effective in some patients with breast cancer.

In a 2009 article, Gnant et al. reported that premenopausal patients, who underwent primary surgery for stage 1 or stage 2 breast cancer and received standard goserelin therapy for induced menopause and endocrine therapy (either tamoxifen or anastrozole) in addition to treatment with zoledronic acid, had a disease-free survival rate of 94.0% at a median 47.8 months of follow-up (N Eng J Med. 2009;360:679-91). After a median 94.4 months of follow-up, the investigators reported a lower risk of disease progression (HR, 0.77; 95% confidence interval, 0.60-0.99; P = .042) and death (HR, 0.66; 95% CI, 0.43-1.02; P = .064; Ann Oncol. 2015;26[2]:313-20).

In his own group, Dr. Coleman and colleagues recruited patients with early-stage breast cancer in the AZURE trial, a phase 3 study of 3,360 patients who received standard therapy with or without zoledronic acid 4 mg every 3-4 weeks for 6 doses, followed by 8 doses every 3 months then 5 doses every 6 months (N Eng J Med. 2011;365:1396-405). “We saw no effect,” said Dr. Coleman. “[That was] very different from what was shown by [Dr.] Gnant.”

For a subgroup of postmenopausal patients, Dr. Coleman and colleagues found that the adjusted HR for disease-free survival was 0.82 for women who were more than 5 years postmenopausal (95% CI, 0.67-1.00) at the time of breast cancer and bisphosphonate treatment, but women younger than 40 years had worse survival outcomes at 10 years (HR, 1.56; 95% CI, 1.09-2.22) and had a significantly higher risk of death from breast cancer (HR, 1.67; 95% CI, 1.16-2.40; J Bone Oncol. 2018 Sep 27. doi: 10.1016/j.jbo.2018.09.008).
 

 

 

Trying to reconcile disparate findings

Those findings left his group with a dilemma, said Dr. Coleman. “Do we start again and run trials, and wait another 10 years, or is there a shortcut to [understanding] what’s going on?” he asked.

In a meta-analysis of all trials from the Early Breast Cancer Trialists’ Collaborative Group in 2015 examining adjuvant bisphosphonate treatment and placebo in patients with early-stage breast cancer, intent-to-treat analyses did not show significant benefit after therapy, but postmenopausal women (11,767 women in 36 trials) saw a clear benefit in all recurrence (rate ratio, 0.86; 95% CI, 0.78-0.94), bone recurrence (RR, 0.72; 95% CI, 0.60-0.86), and breast cancer–related mortality (RR, 0.82; 95% CI, 0.73-0.93; Lancet. 2015 Jul 23. doi: 10.1016/S0140-6736[15]60908-4). The effect seemed to be similar, regardless of bisphosphonate type, with other results seen across trials that used clodronate, zoledronic acid, pamidronate, or ibandronate.

“Although those outcome differences might look quite small for a common disease, that’s a really big effect. Reducing one-sixth of breast cancer deaths at 10 years is the equivalent of saving 10,000 lives across the [European Union], and about half that in the United States,” said Dr. Coleman, noting that guidelines in North America from the American Society of Clinical Oncology and the European Society for Medical Oncology now support adjuvant bisphosphonates in postmenopausal patients with breast cancer.

However, bisphosphonates’ effect on breast cancer does not extend to other cancers, such as non–small cell lung cancer or prostate cancer. In other absorption inhibitors such as denosumab, there also seems to be no benefit for patients with breast cancer, including in postmenopausal patient subgroups, said Dr. Coleman. “In my view, osteoclast inhibition is only part of the story,” he noted.

In the AZURE trial, secondary outcomes examined how the transcription factor MAF interacted with menopausal status and treatment with zoledronic acid. The 79% of patients with tumors that were negative for MAF fluorescence in situ hybridization had improved overall survival (0.69; 95% CI, 0.50-0.94), regardless of menopause status (J Bone Oncol. 2018. doi: 10.1016/j.jbo.2018.09.008). “There’s probably a need to merge the treatment: in this case, the bisphosphonate, the biology of the cancer, and the environment the cancer finds itself in,” noted Dr. Coleman.

“From the cancer perspective, bisphosphonates have transformed the experience for many of our patients with advanced disease, and now we’re working to see great improvements, at least in common diseases such as postmenopausal breast cancer,” he concluded.

Dr. Coleman reports being a paid employee of prIME Oncology (until March 2019); is a consultant for Amgen, Astellas, Boehringer Ingelheim, Scandell, and Biocon; is on the speakers bureau for Amgen and Eisai; holds intellectual property rights for a biomarker being developed by Inbiomotion; and is on the scientific advisory board for Inbiomotion.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

– Cumulative evidence over the years has shown that bisphosphonates reduce distant metastases in postmenopausal women with early-stage breast cancer but that effect does not seem to extend to younger, premenopausal women who may experience adverse effects with bisphosphonates, nor has it been replicated in other types of cancer, such as lung or prostate, Robert Coleman MBBS, MD, said in a presentation marking the 50th anniversary of the publication of the first papers describing the then-new class of drugs.

“We have some science – at least in postmenopausal breast cancer – [showing] that we’re really making a difference, but we don’t understand why it’s not working in the other patients or in the other diseases,” said Dr. Coleman of the University of Sheffield (England) at the annual meeting of the American Society for Bone and Mineral Research.
 

Mixed findings in early studies

Early studies showing the metastases-prevention effects of bisphosphonates were with clodronate. In one study, Dr. Coleman said, researchers found that oral clodronate at a dose of 1,600 mg/day as a supplementary treatment to standard treatment for primary, operable breast cancer reduced the risk of bone metastases in a cohort of 1,069 patients with stage 1-3 breast cancer (hazard ratio, 0.692; P = .043) over 5 years (Breast Cancer Res. 2006 Mar 15. doi: 10.1186/bcr1384). Another trial, of 302 patients with breast cancer who received oral clodronate at 1,600 mg/day for 2 years, found that 20.4% of patients who received oral clodronate had died by 8.5 years of follow-up, compared with 40.7% of those who did not receive the intervention (Ann Oncol. 2008;19[12]:2007-11).

However, those results were in conflict with findings from an earlier study in which researchers followed patients with breast cancer who received 1,600 mg/day of oral clodronate or placebo for 3 years. They found a similar number of bone metastases in the clodronate and placebo groups (32% and 29%, respectively), as well as a lower overall disease-free survival rate at 10 years for the clodronate group, compared with the placebo group (45% vs. 58%; Acta Oncol. 2004;43[7]:650-6).“For various other reasons, clodronate did not gain traction as a therapeutic strategy in early-stage breast cancer,” said Dr. Coleman, although emerging evidence showed that other bisphosphonate agents were effective in some patients with breast cancer.

In a 2009 article, Gnant et al. reported that premenopausal patients, who underwent primary surgery for stage 1 or stage 2 breast cancer and received standard goserelin therapy for induced menopause and endocrine therapy (either tamoxifen or anastrozole) in addition to treatment with zoledronic acid, had a disease-free survival rate of 94.0% at a median 47.8 months of follow-up (N Eng J Med. 2009;360:679-91). After a median 94.4 months of follow-up, the investigators reported a lower risk of disease progression (HR, 0.77; 95% confidence interval, 0.60-0.99; P = .042) and death (HR, 0.66; 95% CI, 0.43-1.02; P = .064; Ann Oncol. 2015;26[2]:313-20).

In his own group, Dr. Coleman and colleagues recruited patients with early-stage breast cancer in the AZURE trial, a phase 3 study of 3,360 patients who received standard therapy with or without zoledronic acid 4 mg every 3-4 weeks for 6 doses, followed by 8 doses every 3 months then 5 doses every 6 months (N Eng J Med. 2011;365:1396-405). “We saw no effect,” said Dr. Coleman. “[That was] very different from what was shown by [Dr.] Gnant.”

For a subgroup of postmenopausal patients, Dr. Coleman and colleagues found that the adjusted HR for disease-free survival was 0.82 for women who were more than 5 years postmenopausal (95% CI, 0.67-1.00) at the time of breast cancer and bisphosphonate treatment, but women younger than 40 years had worse survival outcomes at 10 years (HR, 1.56; 95% CI, 1.09-2.22) and had a significantly higher risk of death from breast cancer (HR, 1.67; 95% CI, 1.16-2.40; J Bone Oncol. 2018 Sep 27. doi: 10.1016/j.jbo.2018.09.008).
 

 

 

Trying to reconcile disparate findings

Those findings left his group with a dilemma, said Dr. Coleman. “Do we start again and run trials, and wait another 10 years, or is there a shortcut to [understanding] what’s going on?” he asked.

In a meta-analysis of all trials from the Early Breast Cancer Trialists’ Collaborative Group in 2015 examining adjuvant bisphosphonate treatment and placebo in patients with early-stage breast cancer, intent-to-treat analyses did not show significant benefit after therapy, but postmenopausal women (11,767 women in 36 trials) saw a clear benefit in all recurrence (rate ratio, 0.86; 95% CI, 0.78-0.94), bone recurrence (RR, 0.72; 95% CI, 0.60-0.86), and breast cancer–related mortality (RR, 0.82; 95% CI, 0.73-0.93; Lancet. 2015 Jul 23. doi: 10.1016/S0140-6736[15]60908-4). The effect seemed to be similar, regardless of bisphosphonate type, with other results seen across trials that used clodronate, zoledronic acid, pamidronate, or ibandronate.

“Although those outcome differences might look quite small for a common disease, that’s a really big effect. Reducing one-sixth of breast cancer deaths at 10 years is the equivalent of saving 10,000 lives across the [European Union], and about half that in the United States,” said Dr. Coleman, noting that guidelines in North America from the American Society of Clinical Oncology and the European Society for Medical Oncology now support adjuvant bisphosphonates in postmenopausal patients with breast cancer.

However, bisphosphonates’ effect on breast cancer does not extend to other cancers, such as non–small cell lung cancer or prostate cancer. In other absorption inhibitors such as denosumab, there also seems to be no benefit for patients with breast cancer, including in postmenopausal patient subgroups, said Dr. Coleman. “In my view, osteoclast inhibition is only part of the story,” he noted.

In the AZURE trial, secondary outcomes examined how the transcription factor MAF interacted with menopausal status and treatment with zoledronic acid. The 79% of patients with tumors that were negative for MAF fluorescence in situ hybridization had improved overall survival (0.69; 95% CI, 0.50-0.94), regardless of menopause status (J Bone Oncol. 2018. doi: 10.1016/j.jbo.2018.09.008). “There’s probably a need to merge the treatment: in this case, the bisphosphonate, the biology of the cancer, and the environment the cancer finds itself in,” noted Dr. Coleman.

“From the cancer perspective, bisphosphonates have transformed the experience for many of our patients with advanced disease, and now we’re working to see great improvements, at least in common diseases such as postmenopausal breast cancer,” he concluded.

Dr. Coleman reports being a paid employee of prIME Oncology (until March 2019); is a consultant for Amgen, Astellas, Boehringer Ingelheim, Scandell, and Biocon; is on the speakers bureau for Amgen and Eisai; holds intellectual property rights for a biomarker being developed by Inbiomotion; and is on the scientific advisory board for Inbiomotion.

– Cumulative evidence over the years has shown that bisphosphonates reduce distant metastases in postmenopausal women with early-stage breast cancer but that effect does not seem to extend to younger, premenopausal women who may experience adverse effects with bisphosphonates, nor has it been replicated in other types of cancer, such as lung or prostate, Robert Coleman MBBS, MD, said in a presentation marking the 50th anniversary of the publication of the first papers describing the then-new class of drugs.

“We have some science – at least in postmenopausal breast cancer – [showing] that we’re really making a difference, but we don’t understand why it’s not working in the other patients or in the other diseases,” said Dr. Coleman of the University of Sheffield (England) at the annual meeting of the American Society for Bone and Mineral Research.
 

Mixed findings in early studies

Early studies showing the metastases-prevention effects of bisphosphonates were with clodronate. In one study, Dr. Coleman said, researchers found that oral clodronate at a dose of 1,600 mg/day as a supplementary treatment to standard treatment for primary, operable breast cancer reduced the risk of bone metastases in a cohort of 1,069 patients with stage 1-3 breast cancer (hazard ratio, 0.692; P = .043) over 5 years (Breast Cancer Res. 2006 Mar 15. doi: 10.1186/bcr1384). Another trial, of 302 patients with breast cancer who received oral clodronate at 1,600 mg/day for 2 years, found that 20.4% of patients who received oral clodronate had died by 8.5 years of follow-up, compared with 40.7% of those who did not receive the intervention (Ann Oncol. 2008;19[12]:2007-11).

However, those results were in conflict with findings from an earlier study in which researchers followed patients with breast cancer who received 1,600 mg/day of oral clodronate or placebo for 3 years. They found a similar number of bone metastases in the clodronate and placebo groups (32% and 29%, respectively), as well as a lower overall disease-free survival rate at 10 years for the clodronate group, compared with the placebo group (45% vs. 58%; Acta Oncol. 2004;43[7]:650-6).“For various other reasons, clodronate did not gain traction as a therapeutic strategy in early-stage breast cancer,” said Dr. Coleman, although emerging evidence showed that other bisphosphonate agents were effective in some patients with breast cancer.

In a 2009 article, Gnant et al. reported that premenopausal patients, who underwent primary surgery for stage 1 or stage 2 breast cancer and received standard goserelin therapy for induced menopause and endocrine therapy (either tamoxifen or anastrozole) in addition to treatment with zoledronic acid, had a disease-free survival rate of 94.0% at a median 47.8 months of follow-up (N Eng J Med. 2009;360:679-91). After a median 94.4 months of follow-up, the investigators reported a lower risk of disease progression (HR, 0.77; 95% confidence interval, 0.60-0.99; P = .042) and death (HR, 0.66; 95% CI, 0.43-1.02; P = .064; Ann Oncol. 2015;26[2]:313-20).

In his own group, Dr. Coleman and colleagues recruited patients with early-stage breast cancer in the AZURE trial, a phase 3 study of 3,360 patients who received standard therapy with or without zoledronic acid 4 mg every 3-4 weeks for 6 doses, followed by 8 doses every 3 months then 5 doses every 6 months (N Eng J Med. 2011;365:1396-405). “We saw no effect,” said Dr. Coleman. “[That was] very different from what was shown by [Dr.] Gnant.”

For a subgroup of postmenopausal patients, Dr. Coleman and colleagues found that the adjusted HR for disease-free survival was 0.82 for women who were more than 5 years postmenopausal (95% CI, 0.67-1.00) at the time of breast cancer and bisphosphonate treatment, but women younger than 40 years had worse survival outcomes at 10 years (HR, 1.56; 95% CI, 1.09-2.22) and had a significantly higher risk of death from breast cancer (HR, 1.67; 95% CI, 1.16-2.40; J Bone Oncol. 2018 Sep 27. doi: 10.1016/j.jbo.2018.09.008).
 

 

 

Trying to reconcile disparate findings

Those findings left his group with a dilemma, said Dr. Coleman. “Do we start again and run trials, and wait another 10 years, or is there a shortcut to [understanding] what’s going on?” he asked.

In a meta-analysis of all trials from the Early Breast Cancer Trialists’ Collaborative Group in 2015 examining adjuvant bisphosphonate treatment and placebo in patients with early-stage breast cancer, intent-to-treat analyses did not show significant benefit after therapy, but postmenopausal women (11,767 women in 36 trials) saw a clear benefit in all recurrence (rate ratio, 0.86; 95% CI, 0.78-0.94), bone recurrence (RR, 0.72; 95% CI, 0.60-0.86), and breast cancer–related mortality (RR, 0.82; 95% CI, 0.73-0.93; Lancet. 2015 Jul 23. doi: 10.1016/S0140-6736[15]60908-4). The effect seemed to be similar, regardless of bisphosphonate type, with other results seen across trials that used clodronate, zoledronic acid, pamidronate, or ibandronate.

“Although those outcome differences might look quite small for a common disease, that’s a really big effect. Reducing one-sixth of breast cancer deaths at 10 years is the equivalent of saving 10,000 lives across the [European Union], and about half that in the United States,” said Dr. Coleman, noting that guidelines in North America from the American Society of Clinical Oncology and the European Society for Medical Oncology now support adjuvant bisphosphonates in postmenopausal patients with breast cancer.

However, bisphosphonates’ effect on breast cancer does not extend to other cancers, such as non–small cell lung cancer or prostate cancer. In other absorption inhibitors such as denosumab, there also seems to be no benefit for patients with breast cancer, including in postmenopausal patient subgroups, said Dr. Coleman. “In my view, osteoclast inhibition is only part of the story,” he noted.

In the AZURE trial, secondary outcomes examined how the transcription factor MAF interacted with menopausal status and treatment with zoledronic acid. The 79% of patients with tumors that were negative for MAF fluorescence in situ hybridization had improved overall survival (0.69; 95% CI, 0.50-0.94), regardless of menopause status (J Bone Oncol. 2018. doi: 10.1016/j.jbo.2018.09.008). “There’s probably a need to merge the treatment: in this case, the bisphosphonate, the biology of the cancer, and the environment the cancer finds itself in,” noted Dr. Coleman.

“From the cancer perspective, bisphosphonates have transformed the experience for many of our patients with advanced disease, and now we’re working to see great improvements, at least in common diseases such as postmenopausal breast cancer,” he concluded.

Dr. Coleman reports being a paid employee of prIME Oncology (until March 2019); is a consultant for Amgen, Astellas, Boehringer Ingelheim, Scandell, and Biocon; is on the speakers bureau for Amgen and Eisai; holds intellectual property rights for a biomarker being developed by Inbiomotion; and is on the scientific advisory board for Inbiomotion.

Publications
Publications
Topics
Article Type
Sections
Article Source

EXPERT ANALYSIS FROM ASBMR 2019

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Expert reviews strategies for diagnosing, treating onychomycosis

Article Type
Changed
Wed, 11/13/2019 - 07:21

– The way Neal Bhatia, MD, sees it, there is no such thing as a classical presentation of onychomycosis.

Dr. Neal Bhatia, director of clinical dermatology research at  Therapeutics Clinical Research, San Diego.
Doug Brunk/MDedge News
Dr. Neal Bhatia

“This is where proving your diagnosis is half the battle, even though we are sometimes using empiric therapy in suspected cases,” he said at Skin Disease Education Foundation’s annual Las Vegas Dermatology Seminar. “Prove the diagnosis and get the extra tests necessary.”

According to Dr. Bhatia, director of clinical dermatology research at San Diego-based Therapeutics Clinical Research, tinea unguium, tinea corporis, tinea cruris, and tinea pedis account for more than half of all visits for dermatophytosis. With onychomycosis, the ultimate treatment goal from the standpoint of clinicians is no more fungus, he noted, while the desired endpoint from the standpoint of some patients is normal-looking nails.

“Endpoint failures in a research trial are not the same as what we tell patients in the clinic,” Dr. Bhatia said. “If you have a patient using something topical for 52 weeks and they see two-thirds of their nail improve, you’re not going to say, ‘Stop; you’re a failure now.’ You tell them, ‘Keep going and we’ll keep watching.’ But in the research world, it’s very different when you look at all of the different endpoints that have to be met at the finish line. It’s very important to measure success based on what that patient’s experiencing.”

According to Dr. Bhatia, diagnosing onychomycosis by visual assessment has a sensitivity of 77% and a specificity of 47%, while KOH has a sensitivity between 67% and 93% and specificity between 38% and 78%. PCR, meanwhile, “is quick, but it’s expensive and it has a high false-positive rate. So, in those difficult patients who aren’t responding [to treatment], maybe we can’t just trust our eyes alone [to make a diagnosis].”

Tests such as a KOH stain have low sensitivity, with high costs of more sensitive tests such as the periodic acid–Schiff (PAS) stain. In a recent study, researchers conducted a retrospective cohort analysis of 600 patients with toenail clippings sent for PAS stain during January 2000–December 2013 (J Am Acad Dermatol. 2015 May; 72(5):AB116). They reviewed records to determine which PAS stains were performed to confirm probable clinical diagnosis of onychomycosis.

The researchers found that 30% of toenail clippings were sent for confirmatory PAS staining by dermatologists, compared with 37% by podiatrists and 34% by other clinicians. Of these tests, 75% ordered by dermatologists were positive for fungus, compared with 81% ordered by podiatrists, and 66% ordered by other physicians. “The positive predictive value of clinical suspicion for true onychomycosis was high, and the findings question whether or not a confirmatory test is really necessary,” Dr. Bhatia said.



Preventative strategies to control recurrence of onychomycosis include using maintenance regimens of the recommended antifungal agent, discarding old shoes, alternating wearing different pairs of shoes, periodically disinfecting shoes, washing feet regularly, and alerting the physician to the first sign of infection.

In an effort to investigate strategies to minimize recurrence of onychomycosis, Dr. Bhatia and colleagues evaluated 73 patients over the course of 7 years who were taking either terbinafine or itraconazole (J Drugs Dermatol. 2016;15[3]:279-82). Thirty-six months later, the overall mean recurrence rate among patients was 14%. Prognostic factors influencing recurrence included patient’s family history; lifestyle; underlying physiology (presence of a very thick nail, extensive involvement of the entire nail unit, lateral nail disease and yellow spikes); physical trauma, especially in the elderly; concomitant disease, such as peripheral artery disease and/or diabetes; immunocompromised or immunosuppressed patients, and the presence of tinea pedis.

Based on their analysis, they recommended the following strategies to prevent or limit recurrence: prophylactic use of a topical antifungal twice-weekly for 2-3 years; periodic application of a topical antifungal to plantar surface and/or interdigital spaces; treatment of any coexisting tinea pedis; treating immediate family members; footwear and sock decontamination with antifungal powder; ultraviolet light or ozone; avoidance of activities known to risk spread of disease, such as communal swimming pools.

Dr. Bhatia concluded his presentation by noting that the ideal treatment for onychomycosis would not pose a systemic risk to the liver, heart, or other organs; would not require monitoring of labs; would not require debridement; and would not interact with other drugs. It would also penetrate the nail plate – especially the diseased nail – and would be quick drying.

SDEF and this news organization are owned by the same parent company.

Dr. Bhatia disclosed having affiliations with Actavis, Allergan, Anacor, Aqua, Bayer, Biofrontera, BiopharmX, Cipher, Dermira, Dusa, Exeltis, Ferndale, Foamix, Galderma, Intraderm, ISDIN, LaRoche-Posay, Leo, Novan, Novartis, PharmaDerm, Promius, Regeneron, Sanofi, Sun Pharma, and Valeant.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

– The way Neal Bhatia, MD, sees it, there is no such thing as a classical presentation of onychomycosis.

Dr. Neal Bhatia, director of clinical dermatology research at  Therapeutics Clinical Research, San Diego.
Doug Brunk/MDedge News
Dr. Neal Bhatia

“This is where proving your diagnosis is half the battle, even though we are sometimes using empiric therapy in suspected cases,” he said at Skin Disease Education Foundation’s annual Las Vegas Dermatology Seminar. “Prove the diagnosis and get the extra tests necessary.”

According to Dr. Bhatia, director of clinical dermatology research at San Diego-based Therapeutics Clinical Research, tinea unguium, tinea corporis, tinea cruris, and tinea pedis account for more than half of all visits for dermatophytosis. With onychomycosis, the ultimate treatment goal from the standpoint of clinicians is no more fungus, he noted, while the desired endpoint from the standpoint of some patients is normal-looking nails.

“Endpoint failures in a research trial are not the same as what we tell patients in the clinic,” Dr. Bhatia said. “If you have a patient using something topical for 52 weeks and they see two-thirds of their nail improve, you’re not going to say, ‘Stop; you’re a failure now.’ You tell them, ‘Keep going and we’ll keep watching.’ But in the research world, it’s very different when you look at all of the different endpoints that have to be met at the finish line. It’s very important to measure success based on what that patient’s experiencing.”

According to Dr. Bhatia, diagnosing onychomycosis by visual assessment has a sensitivity of 77% and a specificity of 47%, while KOH has a sensitivity between 67% and 93% and specificity between 38% and 78%. PCR, meanwhile, “is quick, but it’s expensive and it has a high false-positive rate. So, in those difficult patients who aren’t responding [to treatment], maybe we can’t just trust our eyes alone [to make a diagnosis].”

Tests such as a KOH stain have low sensitivity, with high costs of more sensitive tests such as the periodic acid–Schiff (PAS) stain. In a recent study, researchers conducted a retrospective cohort analysis of 600 patients with toenail clippings sent for PAS stain during January 2000–December 2013 (J Am Acad Dermatol. 2015 May; 72(5):AB116). They reviewed records to determine which PAS stains were performed to confirm probable clinical diagnosis of onychomycosis.

The researchers found that 30% of toenail clippings were sent for confirmatory PAS staining by dermatologists, compared with 37% by podiatrists and 34% by other clinicians. Of these tests, 75% ordered by dermatologists were positive for fungus, compared with 81% ordered by podiatrists, and 66% ordered by other physicians. “The positive predictive value of clinical suspicion for true onychomycosis was high, and the findings question whether or not a confirmatory test is really necessary,” Dr. Bhatia said.



Preventative strategies to control recurrence of onychomycosis include using maintenance regimens of the recommended antifungal agent, discarding old shoes, alternating wearing different pairs of shoes, periodically disinfecting shoes, washing feet regularly, and alerting the physician to the first sign of infection.

In an effort to investigate strategies to minimize recurrence of onychomycosis, Dr. Bhatia and colleagues evaluated 73 patients over the course of 7 years who were taking either terbinafine or itraconazole (J Drugs Dermatol. 2016;15[3]:279-82). Thirty-six months later, the overall mean recurrence rate among patients was 14%. Prognostic factors influencing recurrence included patient’s family history; lifestyle; underlying physiology (presence of a very thick nail, extensive involvement of the entire nail unit, lateral nail disease and yellow spikes); physical trauma, especially in the elderly; concomitant disease, such as peripheral artery disease and/or diabetes; immunocompromised or immunosuppressed patients, and the presence of tinea pedis.

Based on their analysis, they recommended the following strategies to prevent or limit recurrence: prophylactic use of a topical antifungal twice-weekly for 2-3 years; periodic application of a topical antifungal to plantar surface and/or interdigital spaces; treatment of any coexisting tinea pedis; treating immediate family members; footwear and sock decontamination with antifungal powder; ultraviolet light or ozone; avoidance of activities known to risk spread of disease, such as communal swimming pools.

Dr. Bhatia concluded his presentation by noting that the ideal treatment for onychomycosis would not pose a systemic risk to the liver, heart, or other organs; would not require monitoring of labs; would not require debridement; and would not interact with other drugs. It would also penetrate the nail plate – especially the diseased nail – and would be quick drying.

SDEF and this news organization are owned by the same parent company.

Dr. Bhatia disclosed having affiliations with Actavis, Allergan, Anacor, Aqua, Bayer, Biofrontera, BiopharmX, Cipher, Dermira, Dusa, Exeltis, Ferndale, Foamix, Galderma, Intraderm, ISDIN, LaRoche-Posay, Leo, Novan, Novartis, PharmaDerm, Promius, Regeneron, Sanofi, Sun Pharma, and Valeant.

– The way Neal Bhatia, MD, sees it, there is no such thing as a classical presentation of onychomycosis.

Dr. Neal Bhatia, director of clinical dermatology research at  Therapeutics Clinical Research, San Diego.
Doug Brunk/MDedge News
Dr. Neal Bhatia

“This is where proving your diagnosis is half the battle, even though we are sometimes using empiric therapy in suspected cases,” he said at Skin Disease Education Foundation’s annual Las Vegas Dermatology Seminar. “Prove the diagnosis and get the extra tests necessary.”

According to Dr. Bhatia, director of clinical dermatology research at San Diego-based Therapeutics Clinical Research, tinea unguium, tinea corporis, tinea cruris, and tinea pedis account for more than half of all visits for dermatophytosis. With onychomycosis, the ultimate treatment goal from the standpoint of clinicians is no more fungus, he noted, while the desired endpoint from the standpoint of some patients is normal-looking nails.

“Endpoint failures in a research trial are not the same as what we tell patients in the clinic,” Dr. Bhatia said. “If you have a patient using something topical for 52 weeks and they see two-thirds of their nail improve, you’re not going to say, ‘Stop; you’re a failure now.’ You tell them, ‘Keep going and we’ll keep watching.’ But in the research world, it’s very different when you look at all of the different endpoints that have to be met at the finish line. It’s very important to measure success based on what that patient’s experiencing.”

According to Dr. Bhatia, diagnosing onychomycosis by visual assessment has a sensitivity of 77% and a specificity of 47%, while KOH has a sensitivity between 67% and 93% and specificity between 38% and 78%. PCR, meanwhile, “is quick, but it’s expensive and it has a high false-positive rate. So, in those difficult patients who aren’t responding [to treatment], maybe we can’t just trust our eyes alone [to make a diagnosis].”

Tests such as a KOH stain have low sensitivity, with high costs of more sensitive tests such as the periodic acid–Schiff (PAS) stain. In a recent study, researchers conducted a retrospective cohort analysis of 600 patients with toenail clippings sent for PAS stain during January 2000–December 2013 (J Am Acad Dermatol. 2015 May; 72(5):AB116). They reviewed records to determine which PAS stains were performed to confirm probable clinical diagnosis of onychomycosis.

The researchers found that 30% of toenail clippings were sent for confirmatory PAS staining by dermatologists, compared with 37% by podiatrists and 34% by other clinicians. Of these tests, 75% ordered by dermatologists were positive for fungus, compared with 81% ordered by podiatrists, and 66% ordered by other physicians. “The positive predictive value of clinical suspicion for true onychomycosis was high, and the findings question whether or not a confirmatory test is really necessary,” Dr. Bhatia said.



Preventative strategies to control recurrence of onychomycosis include using maintenance regimens of the recommended antifungal agent, discarding old shoes, alternating wearing different pairs of shoes, periodically disinfecting shoes, washing feet regularly, and alerting the physician to the first sign of infection.

In an effort to investigate strategies to minimize recurrence of onychomycosis, Dr. Bhatia and colleagues evaluated 73 patients over the course of 7 years who were taking either terbinafine or itraconazole (J Drugs Dermatol. 2016;15[3]:279-82). Thirty-six months later, the overall mean recurrence rate among patients was 14%. Prognostic factors influencing recurrence included patient’s family history; lifestyle; underlying physiology (presence of a very thick nail, extensive involvement of the entire nail unit, lateral nail disease and yellow spikes); physical trauma, especially in the elderly; concomitant disease, such as peripheral artery disease and/or diabetes; immunocompromised or immunosuppressed patients, and the presence of tinea pedis.

Based on their analysis, they recommended the following strategies to prevent or limit recurrence: prophylactic use of a topical antifungal twice-weekly for 2-3 years; periodic application of a topical antifungal to plantar surface and/or interdigital spaces; treatment of any coexisting tinea pedis; treating immediate family members; footwear and sock decontamination with antifungal powder; ultraviolet light or ozone; avoidance of activities known to risk spread of disease, such as communal swimming pools.

Dr. Bhatia concluded his presentation by noting that the ideal treatment for onychomycosis would not pose a systemic risk to the liver, heart, or other organs; would not require monitoring of labs; would not require debridement; and would not interact with other drugs. It would also penetrate the nail plate – especially the diseased nail – and would be quick drying.

SDEF and this news organization are owned by the same parent company.

Dr. Bhatia disclosed having affiliations with Actavis, Allergan, Anacor, Aqua, Bayer, Biofrontera, BiopharmX, Cipher, Dermira, Dusa, Exeltis, Ferndale, Foamix, Galderma, Intraderm, ISDIN, LaRoche-Posay, Leo, Novan, Novartis, PharmaDerm, Promius, Regeneron, Sanofi, Sun Pharma, and Valeant.

Publications
Publications
Topics
Article Type
Sections
Article Source

AT THE SDEF LAS VEGAS DERMATOLOGY SEMINAR

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

FDA approves anemia treatment for transfusion-dependent beta thalassemia patients

Article Type
Changed
Fri, 11/08/2019 - 15:54

 

The Food and Drug Administration has approved the first treatment for anemia in adults with transfusion-dependent beta thalassemia.

A stamp saying "FDA approved."
Olivier Le Moal/Getty Images

Luspatercept-aamt (Reblozyl) is an erythroid maturation agent that reduced the transfusion burden for patients with beta thalassemia in the BELIEVE trial of 336 patients. In total, 21% of patients who received luspatercept-aamt achieved at least a 33% reduction in red blood cell transfusions, compared with 4.5% of patients who received placebo, according to the FDA.

Common side effects associated with luspatercept-aamt were headache, bone pain, arthralgia, fatigue, cough, abdominal pain, diarrhea, and dizziness. Patients taking the agent should be monitored for thrombosis, the FDA advised.



Celgene, which makes luspatercept-aamt, said the agent would be available about 1 week following the FDA approval.

The FDA is also evaluating luspatercept-aamt as an anemia treatment in adults with very-low– to intermediate-risk myelodysplastic syndromes who have ring sideroblasts and require red blood cell transfusions. The agency is expected to take action on that application in April 2020.

Publications
Topics
Sections

 

The Food and Drug Administration has approved the first treatment for anemia in adults with transfusion-dependent beta thalassemia.

A stamp saying "FDA approved."
Olivier Le Moal/Getty Images

Luspatercept-aamt (Reblozyl) is an erythroid maturation agent that reduced the transfusion burden for patients with beta thalassemia in the BELIEVE trial of 336 patients. In total, 21% of patients who received luspatercept-aamt achieved at least a 33% reduction in red blood cell transfusions, compared with 4.5% of patients who received placebo, according to the FDA.

Common side effects associated with luspatercept-aamt were headache, bone pain, arthralgia, fatigue, cough, abdominal pain, diarrhea, and dizziness. Patients taking the agent should be monitored for thrombosis, the FDA advised.



Celgene, which makes luspatercept-aamt, said the agent would be available about 1 week following the FDA approval.

The FDA is also evaluating luspatercept-aamt as an anemia treatment in adults with very-low– to intermediate-risk myelodysplastic syndromes who have ring sideroblasts and require red blood cell transfusions. The agency is expected to take action on that application in April 2020.

 

The Food and Drug Administration has approved the first treatment for anemia in adults with transfusion-dependent beta thalassemia.

A stamp saying "FDA approved."
Olivier Le Moal/Getty Images

Luspatercept-aamt (Reblozyl) is an erythroid maturation agent that reduced the transfusion burden for patients with beta thalassemia in the BELIEVE trial of 336 patients. In total, 21% of patients who received luspatercept-aamt achieved at least a 33% reduction in red blood cell transfusions, compared with 4.5% of patients who received placebo, according to the FDA.

Common side effects associated with luspatercept-aamt were headache, bone pain, arthralgia, fatigue, cough, abdominal pain, diarrhea, and dizziness. Patients taking the agent should be monitored for thrombosis, the FDA advised.



Celgene, which makes luspatercept-aamt, said the agent would be available about 1 week following the FDA approval.

The FDA is also evaluating luspatercept-aamt as an anemia treatment in adults with very-low– to intermediate-risk myelodysplastic syndromes who have ring sideroblasts and require red blood cell transfusions. The agency is expected to take action on that application in April 2020.

Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Exercise improved QoL, functioning in breast cancer survivors

Article Type
Changed
Wed, 01/04/2023 - 16:43

– A supervised and adapted exercise program improved quality of life, physical functioning, and strength in breast cancer survivors participating in the MAMA MOVE Gaia study.

Of 19 women who initiated participation in the program, which included a 16-week control phase followed by a 16-week exercise training intervention phase, 15 completed the program, and, after the training intervention, they experienced a significant increase in handgrip strength and sit-to-stand repetitions, Ana Joaquim, MD, of Centro Hospitalar de Vila Nova de Gaia/Espinho, Portugal, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

During the control phase of the prospective nonrandomized study, participants experienced no significant changes over time in any domain of quality of life as measured by the EORTC QLQ-C30 questionnaire, although a trend toward improved physical functioning was noted at an evaluation performed 8 weeks after the control phase, compared with one performed just prior to the intervention phase (77.3 to 85.3 points, P = .051), the investigators said.

After the intervention phase, however, handgrip strength improved significantly at both the limb where surgery was performed and at the nonoperated limb (from 22.2 to 25.6 kg.f and from 22.6 to 26.9 kg.f). Similar results were observed for a sit-to-stand test (improvement from 12 to 17 repetitions).

Participants in the single-arm clinical trial were assessed after 8 weeks of the control phase, immediately prior to the intervention period, 8 weeks after the control phase, and 16 weeks into the invention phase.

The intervention phase consisted of 3 60-minute sessions per week of combined moderate to vigorous aerobic and strength exercise, defined as exercise at 65%-85% of maximum heart rate or at 6-8 points on the OMNI scale. Mean compliance among the participants was 63.6%.

The participants had a median age of 59 and 15 of the 19 were diagnosed with invasive carcinoma. Following surgery, 13 underwent radiotherapy, 15 received chemotherapy, and 18 received hormone therapy.

“Treatments for early breast cancer have side effects that affect quality of life and cause deconditioning,” the investigators wrote, adding that “physical exercise might have a supportive and coadjuvant role in the rehabilitation of breast cancer survivors.”

The MAMA MOVE trial aimed to assess the potential benefits of a community-based supervised exercise training program, and the findings suggest such programs could help improve quality of life, particularly with respect to physical functioning, they concluded.

The MAMA MOVE Gaia study was funded by Liga Portuguesa Contra o Cancro. The investigators reported having no disclosures.
 

SOURCE: Joaquim A et al. ESMO 2019, Abstract 234P.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

– A supervised and adapted exercise program improved quality of life, physical functioning, and strength in breast cancer survivors participating in the MAMA MOVE Gaia study.

Of 19 women who initiated participation in the program, which included a 16-week control phase followed by a 16-week exercise training intervention phase, 15 completed the program, and, after the training intervention, they experienced a significant increase in handgrip strength and sit-to-stand repetitions, Ana Joaquim, MD, of Centro Hospitalar de Vila Nova de Gaia/Espinho, Portugal, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

During the control phase of the prospective nonrandomized study, participants experienced no significant changes over time in any domain of quality of life as measured by the EORTC QLQ-C30 questionnaire, although a trend toward improved physical functioning was noted at an evaluation performed 8 weeks after the control phase, compared with one performed just prior to the intervention phase (77.3 to 85.3 points, P = .051), the investigators said.

After the intervention phase, however, handgrip strength improved significantly at both the limb where surgery was performed and at the nonoperated limb (from 22.2 to 25.6 kg.f and from 22.6 to 26.9 kg.f). Similar results were observed for a sit-to-stand test (improvement from 12 to 17 repetitions).

Participants in the single-arm clinical trial were assessed after 8 weeks of the control phase, immediately prior to the intervention period, 8 weeks after the control phase, and 16 weeks into the invention phase.

The intervention phase consisted of 3 60-minute sessions per week of combined moderate to vigorous aerobic and strength exercise, defined as exercise at 65%-85% of maximum heart rate or at 6-8 points on the OMNI scale. Mean compliance among the participants was 63.6%.

The participants had a median age of 59 and 15 of the 19 were diagnosed with invasive carcinoma. Following surgery, 13 underwent radiotherapy, 15 received chemotherapy, and 18 received hormone therapy.

“Treatments for early breast cancer have side effects that affect quality of life and cause deconditioning,” the investigators wrote, adding that “physical exercise might have a supportive and coadjuvant role in the rehabilitation of breast cancer survivors.”

The MAMA MOVE trial aimed to assess the potential benefits of a community-based supervised exercise training program, and the findings suggest such programs could help improve quality of life, particularly with respect to physical functioning, they concluded.

The MAMA MOVE Gaia study was funded by Liga Portuguesa Contra o Cancro. The investigators reported having no disclosures.
 

SOURCE: Joaquim A et al. ESMO 2019, Abstract 234P.

– A supervised and adapted exercise program improved quality of life, physical functioning, and strength in breast cancer survivors participating in the MAMA MOVE Gaia study.

Of 19 women who initiated participation in the program, which included a 16-week control phase followed by a 16-week exercise training intervention phase, 15 completed the program, and, after the training intervention, they experienced a significant increase in handgrip strength and sit-to-stand repetitions, Ana Joaquim, MD, of Centro Hospitalar de Vila Nova de Gaia/Espinho, Portugal, and colleagues reported in a poster at the European Society for Medical Oncology Congress.

During the control phase of the prospective nonrandomized study, participants experienced no significant changes over time in any domain of quality of life as measured by the EORTC QLQ-C30 questionnaire, although a trend toward improved physical functioning was noted at an evaluation performed 8 weeks after the control phase, compared with one performed just prior to the intervention phase (77.3 to 85.3 points, P = .051), the investigators said.

After the intervention phase, however, handgrip strength improved significantly at both the limb where surgery was performed and at the nonoperated limb (from 22.2 to 25.6 kg.f and from 22.6 to 26.9 kg.f). Similar results were observed for a sit-to-stand test (improvement from 12 to 17 repetitions).

Participants in the single-arm clinical trial were assessed after 8 weeks of the control phase, immediately prior to the intervention period, 8 weeks after the control phase, and 16 weeks into the invention phase.

The intervention phase consisted of 3 60-minute sessions per week of combined moderate to vigorous aerobic and strength exercise, defined as exercise at 65%-85% of maximum heart rate or at 6-8 points on the OMNI scale. Mean compliance among the participants was 63.6%.

The participants had a median age of 59 and 15 of the 19 were diagnosed with invasive carcinoma. Following surgery, 13 underwent radiotherapy, 15 received chemotherapy, and 18 received hormone therapy.

“Treatments for early breast cancer have side effects that affect quality of life and cause deconditioning,” the investigators wrote, adding that “physical exercise might have a supportive and coadjuvant role in the rehabilitation of breast cancer survivors.”

The MAMA MOVE trial aimed to assess the potential benefits of a community-based supervised exercise training program, and the findings suggest such programs could help improve quality of life, particularly with respect to physical functioning, they concluded.

The MAMA MOVE Gaia study was funded by Liga Portuguesa Contra o Cancro. The investigators reported having no disclosures.
 

SOURCE: Joaquim A et al. ESMO 2019, Abstract 234P.

Publications
Publications
Topics
Article Type
Sections
Article Source

REPORTING FROM ESMO 2019

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

New model for CKD risk draws on clinical, demographic factors

Article Type
Changed
Tue, 05/03/2022 - 15:12

 

Data from more than 5 million individuals has been used to develop an equation for predicting the risk of incident chronic kidney disease (CKD) in people with or without diabetes, according to a presentation at Kidney Week 2019, sponsored by the American Society of Nephrology.

In a paper published simultaneously online in JAMA, researchers reported the outcome of an individual-level data analysis of 34 multinational cohorts involving 5,222,711 individuals – including 781,627 with diabetes – from 28 countries as part of the Chronic Kidney Disease Prognosis Consortium.

“An equation for kidney failure risk may help improve care for patients with established CKD, but relatively little work has been performed to develop predictive tools to identify those at increased risk of developing CKD – defined by reduced eGFR [estimated glomerular filtration rate], despite the high lifetime risk of CKD – which is estimated to be 59.1% in the United States,” wrote Robert G. Nelson, MD, PhD, from the National Institute of Diabetes and Digestive and Kidney Diseases in Phoenix and colleagues.

Over a mean follow-up of 4 years, 15% of individuals without diabetes and 40% of individuals with diabetes developed incident chronic kidney disease, defined as an eGFR below 60 mL/min per 1.73m2.

The key risk factors were older age, female sex, black race, hypertension, history of cardiovascular disease, lower eGFR values, and higher urine albumin to creatinine ratio. Smoking was also significantly associated with reduced eGFR but only in cohorts without diabetes. In cohorts with diabetes, elevated hemoglobin A1c and the presence and type of diabetes medication were also significantly associated with reduced eGFR.

Using this information, the researchers developed a prediction model built from weighted-average hazard ratios and validated it in nine external validation cohorts of 18 study populations involving a total of 2,253,540 individuals. They found that in 16 of the 18 study populations, the slope of observed to predicted risk ranged from 0.80 to 1.25.

Moreover, in the cohorts without diabetes, the risk equations had a median C-statistic for the 5-year predicted probability of 0.845 (interquartile range, 0.789-0.890) and of 0.801 (IQR, 0.750-0.819) in the cohorts with diabetes, the investigators reported.

“Several models have been developed for estimating the risk of prevalent and incident CKD and end-stage kidney disease, but even those with good discriminative performance have not always performed well for cohorts of people outside the original derivation cohort,” the authors wrote. They argued that their model “demonstrated high discrimination and variable calibration in diverse populations.”

However, they stressed that further study was needed to determine if use of the equations would actually lead to improvements in clinical care and patient outcomes. In an accompanying editorial, Sri Lekha Tummalapalli, MD, and Michelle M. Estrella, MD, of the Kidney Health Research Collaborative at the University of California, San Francisco, said the study and its focus on primary, rather than secondary, prevention of kidney disease is a critical step toward reducing the burden of that disease, especially given that an estimated 37 million people in the United States have chronic kidney disease.

It is also important, they added, that primary prevention of kidney disease is tailored to the individual patient’s risk because risk prediction and screening strategies are unlikely to improve outcomes if they are not paired with effective individualized interventions, such as lifestyle modification or management of blood pressure.

These risk equations could be more holistic by integrating the prediction of both elevated albuminuria and reduced eGFR because more than 40% of individuals with chronic kidney disease have increased albuminuria without reduced eGFR, they noted (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17378).

The study and CKD Prognosis Consortium were supported by the U.S. National Kidney Foundation and the National Institute of Diabetes and Digestive and Kidney Diseases. One author was supported by a grant from the German Research Foundation. Nine authors declared grants, consultancies, and other support from the private sector and research organizations. No other conflicts of interest were declared. Dr. Tummalapalli and Dr. Estrella reported no conflicts of interest.

SOURCE: Nelson R et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17379.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

 

Data from more than 5 million individuals has been used to develop an equation for predicting the risk of incident chronic kidney disease (CKD) in people with or without diabetes, according to a presentation at Kidney Week 2019, sponsored by the American Society of Nephrology.

In a paper published simultaneously online in JAMA, researchers reported the outcome of an individual-level data analysis of 34 multinational cohorts involving 5,222,711 individuals – including 781,627 with diabetes – from 28 countries as part of the Chronic Kidney Disease Prognosis Consortium.

“An equation for kidney failure risk may help improve care for patients with established CKD, but relatively little work has been performed to develop predictive tools to identify those at increased risk of developing CKD – defined by reduced eGFR [estimated glomerular filtration rate], despite the high lifetime risk of CKD – which is estimated to be 59.1% in the United States,” wrote Robert G. Nelson, MD, PhD, from the National Institute of Diabetes and Digestive and Kidney Diseases in Phoenix and colleagues.

Over a mean follow-up of 4 years, 15% of individuals without diabetes and 40% of individuals with diabetes developed incident chronic kidney disease, defined as an eGFR below 60 mL/min per 1.73m2.

The key risk factors were older age, female sex, black race, hypertension, history of cardiovascular disease, lower eGFR values, and higher urine albumin to creatinine ratio. Smoking was also significantly associated with reduced eGFR but only in cohorts without diabetes. In cohorts with diabetes, elevated hemoglobin A1c and the presence and type of diabetes medication were also significantly associated with reduced eGFR.

Using this information, the researchers developed a prediction model built from weighted-average hazard ratios and validated it in nine external validation cohorts of 18 study populations involving a total of 2,253,540 individuals. They found that in 16 of the 18 study populations, the slope of observed to predicted risk ranged from 0.80 to 1.25.

Moreover, in the cohorts without diabetes, the risk equations had a median C-statistic for the 5-year predicted probability of 0.845 (interquartile range, 0.789-0.890) and of 0.801 (IQR, 0.750-0.819) in the cohorts with diabetes, the investigators reported.

“Several models have been developed for estimating the risk of prevalent and incident CKD and end-stage kidney disease, but even those with good discriminative performance have not always performed well for cohorts of people outside the original derivation cohort,” the authors wrote. They argued that their model “demonstrated high discrimination and variable calibration in diverse populations.”

However, they stressed that further study was needed to determine if use of the equations would actually lead to improvements in clinical care and patient outcomes. In an accompanying editorial, Sri Lekha Tummalapalli, MD, and Michelle M. Estrella, MD, of the Kidney Health Research Collaborative at the University of California, San Francisco, said the study and its focus on primary, rather than secondary, prevention of kidney disease is a critical step toward reducing the burden of that disease, especially given that an estimated 37 million people in the United States have chronic kidney disease.

It is also important, they added, that primary prevention of kidney disease is tailored to the individual patient’s risk because risk prediction and screening strategies are unlikely to improve outcomes if they are not paired with effective individualized interventions, such as lifestyle modification or management of blood pressure.

These risk equations could be more holistic by integrating the prediction of both elevated albuminuria and reduced eGFR because more than 40% of individuals with chronic kidney disease have increased albuminuria without reduced eGFR, they noted (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17378).

The study and CKD Prognosis Consortium were supported by the U.S. National Kidney Foundation and the National Institute of Diabetes and Digestive and Kidney Diseases. One author was supported by a grant from the German Research Foundation. Nine authors declared grants, consultancies, and other support from the private sector and research organizations. No other conflicts of interest were declared. Dr. Tummalapalli and Dr. Estrella reported no conflicts of interest.

SOURCE: Nelson R et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17379.

 

Data from more than 5 million individuals has been used to develop an equation for predicting the risk of incident chronic kidney disease (CKD) in people with or without diabetes, according to a presentation at Kidney Week 2019, sponsored by the American Society of Nephrology.

In a paper published simultaneously online in JAMA, researchers reported the outcome of an individual-level data analysis of 34 multinational cohorts involving 5,222,711 individuals – including 781,627 with diabetes – from 28 countries as part of the Chronic Kidney Disease Prognosis Consortium.

“An equation for kidney failure risk may help improve care for patients with established CKD, but relatively little work has been performed to develop predictive tools to identify those at increased risk of developing CKD – defined by reduced eGFR [estimated glomerular filtration rate], despite the high lifetime risk of CKD – which is estimated to be 59.1% in the United States,” wrote Robert G. Nelson, MD, PhD, from the National Institute of Diabetes and Digestive and Kidney Diseases in Phoenix and colleagues.

Over a mean follow-up of 4 years, 15% of individuals without diabetes and 40% of individuals with diabetes developed incident chronic kidney disease, defined as an eGFR below 60 mL/min per 1.73m2.

The key risk factors were older age, female sex, black race, hypertension, history of cardiovascular disease, lower eGFR values, and higher urine albumin to creatinine ratio. Smoking was also significantly associated with reduced eGFR but only in cohorts without diabetes. In cohorts with diabetes, elevated hemoglobin A1c and the presence and type of diabetes medication were also significantly associated with reduced eGFR.

Using this information, the researchers developed a prediction model built from weighted-average hazard ratios and validated it in nine external validation cohorts of 18 study populations involving a total of 2,253,540 individuals. They found that in 16 of the 18 study populations, the slope of observed to predicted risk ranged from 0.80 to 1.25.

Moreover, in the cohorts without diabetes, the risk equations had a median C-statistic for the 5-year predicted probability of 0.845 (interquartile range, 0.789-0.890) and of 0.801 (IQR, 0.750-0.819) in the cohorts with diabetes, the investigators reported.

“Several models have been developed for estimating the risk of prevalent and incident CKD and end-stage kidney disease, but even those with good discriminative performance have not always performed well for cohorts of people outside the original derivation cohort,” the authors wrote. They argued that their model “demonstrated high discrimination and variable calibration in diverse populations.”

However, they stressed that further study was needed to determine if use of the equations would actually lead to improvements in clinical care and patient outcomes. In an accompanying editorial, Sri Lekha Tummalapalli, MD, and Michelle M. Estrella, MD, of the Kidney Health Research Collaborative at the University of California, San Francisco, said the study and its focus on primary, rather than secondary, prevention of kidney disease is a critical step toward reducing the burden of that disease, especially given that an estimated 37 million people in the United States have chronic kidney disease.

It is also important, they added, that primary prevention of kidney disease is tailored to the individual patient’s risk because risk prediction and screening strategies are unlikely to improve outcomes if they are not paired with effective individualized interventions, such as lifestyle modification or management of blood pressure.

These risk equations could be more holistic by integrating the prediction of both elevated albuminuria and reduced eGFR because more than 40% of individuals with chronic kidney disease have increased albuminuria without reduced eGFR, they noted (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17378).

The study and CKD Prognosis Consortium were supported by the U.S. National Kidney Foundation and the National Institute of Diabetes and Digestive and Kidney Diseases. One author was supported by a grant from the German Research Foundation. Nine authors declared grants, consultancies, and other support from the private sector and research organizations. No other conflicts of interest were declared. Dr. Tummalapalli and Dr. Estrella reported no conflicts of interest.

SOURCE: Nelson R et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17379.

Publications
Publications
Topics
Article Type
Sections
Article Source

REPORTING FROM KIDNEY WEEK 2019

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Vitals

 

Key clinical point: Demographic and clinical factors, including age, sex, ethnicity, cardiovascular risk factors, and albuminuria can be used to predict risk of chronic kidney disease.

Major finding: In the cohorts without diabetes, the risk equations had a median C-statistic for the 5-year predicted probability of 0.845 (interquartile range, 0.789-0.890), and of 0.801 (IQR, 0.750-0.819) in the cohorts with diabetes,

Study details: Analysis of cohort data from 5,222,711 individuals, including 781,627 with diabetes.

Disclosures: The study and CKD Prognosis Consortium were supported by the U.S. National Kidney Foundation and the National Institute of Diabetes and Digestive and Kidney Diseases. One author was supported by a grant from the German Research Foundation. Nine authors declared grants, consultancies, and other support from the private sector and research organizations. No other conflicts of interest were declared. Dr. Tummalapalli and Dr. Estrella reported no conflicts of interest.

Source: Nelson R et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17379.

Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.

Vitamin D, omega-3 fatty acids do not preserve kidney function in type 2 diabetes

Article Type
Changed
Tue, 05/03/2022 - 15:12

 

A new study has found that neither vitamin D nor omega-3 fatty acids are significantly more beneficial than placebo for prevention and treatment of chronic kidney disease in patients with type 2 diabetes, according to Ian H. de Boer, MD, of the University of Washington, Seattle, and coauthors.

Findings of the study were presented at Kidney Week 2019, sponsored the American Society of Nephrology, and published simultaneously in JAMA.

To determine the benefits of either vitamin D or omega-3 fatty acids in regard to kidney function, the researchers conducted a randomized clinical trial of 1,312 patients with type 2 diabetes. The trial was designed to accompany the Vitamin D and Omega-3 Trial (VITAL), which enrolled 25,871 patients to test the two supplements in the prevention of cardiovascular disease and cancer.

Participants in this study – known as VITAL–Diabetic Kidney Disease, designed as an ancillary to VITAL – were assigned to one of four groups: vitamin D plus omega-3 fatty acids (n = 370), vitamin D plus placebo (n = 333), omega-3 fatty acids plus placebo (n = 289), or both placebos (n = 320). The goal was to assess changes in in glomerular filtration rate estimated from serum creatinine and cystatin C (eGFR) after 5 years.

Of the initial 1,312 participants, 934 (71%) finished the study. At 5-year follow-up, patients taking vitamin D had a mean change in eGFR of −12.3 mL/min per 1.73 m2 (95% confidence interval, −13.4 to −11.2), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −11.9) with placebo. Patients taking omega-3 fatty acids had a mean eGFR change of −12.2 mL/min per 1.73 m2 (95% CI, −13.3 to −11.1), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −12.0) with placebo.

The authors noted that the modest number of measurements collected per participant limited the evaluation and analyses. In addition, the study focused broadly on the type 2 diabetes population and not on subgroups, “who may derive more benefit from the study interventions.”

In an accompanying editorial, authors Anika Lucas, MD and Myles Wolf, MD, of Duke University in Durham, N.C., said multiple clinical trials, including this latest study from de Boer and colleagues on kidney function, have failed to reinforce the previously reported benefits of vitamin D.

“The VITAL-DKD study population had nearly normal mean 25-hydroxyvitamin D levels at baseline, leaving open the question of whether the results would have differed had recruitment been restricted to patients with moderate or severe vitamin D deficiency,” they wrote (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17302).

Nevertheless, it seems safe to conclude that the previous associations between vitamin D deficiency and adverse health were “driven by unmeasured residual confounding or reverse causality.

“Without certainty about the ideal approach to vitamin D treatment in advanced CKD, a randomized trial that compared cholecalciferol, exogenous 25-hydroxyvitamin D, and an activated vitamin D analogue vs. placebo could definitively lay to rest multiple remaining questions in the area,” they suggested.

The National Institute of Diabetes and Digestive and Kidney Diseases funded the study. The authors reported numerous potential conflicts of interest, including receiving grants, consulting fees, and equipment and supplies from various pharmaceutical companies and the National Institutes of Health. Dr. Wolf reported having served as a consultant for Akebia, AMAG, Amgen, Ardelyx, Diasorin, and Pharmacosmos. No other disclosures were reported.

SOURCE: de Boer IH et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17380.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

 

A new study has found that neither vitamin D nor omega-3 fatty acids are significantly more beneficial than placebo for prevention and treatment of chronic kidney disease in patients with type 2 diabetes, according to Ian H. de Boer, MD, of the University of Washington, Seattle, and coauthors.

Findings of the study were presented at Kidney Week 2019, sponsored the American Society of Nephrology, and published simultaneously in JAMA.

To determine the benefits of either vitamin D or omega-3 fatty acids in regard to kidney function, the researchers conducted a randomized clinical trial of 1,312 patients with type 2 diabetes. The trial was designed to accompany the Vitamin D and Omega-3 Trial (VITAL), which enrolled 25,871 patients to test the two supplements in the prevention of cardiovascular disease and cancer.

Participants in this study – known as VITAL–Diabetic Kidney Disease, designed as an ancillary to VITAL – were assigned to one of four groups: vitamin D plus omega-3 fatty acids (n = 370), vitamin D plus placebo (n = 333), omega-3 fatty acids plus placebo (n = 289), or both placebos (n = 320). The goal was to assess changes in in glomerular filtration rate estimated from serum creatinine and cystatin C (eGFR) after 5 years.

Of the initial 1,312 participants, 934 (71%) finished the study. At 5-year follow-up, patients taking vitamin D had a mean change in eGFR of −12.3 mL/min per 1.73 m2 (95% confidence interval, −13.4 to −11.2), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −11.9) with placebo. Patients taking omega-3 fatty acids had a mean eGFR change of −12.2 mL/min per 1.73 m2 (95% CI, −13.3 to −11.1), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −12.0) with placebo.

The authors noted that the modest number of measurements collected per participant limited the evaluation and analyses. In addition, the study focused broadly on the type 2 diabetes population and not on subgroups, “who may derive more benefit from the study interventions.”

In an accompanying editorial, authors Anika Lucas, MD and Myles Wolf, MD, of Duke University in Durham, N.C., said multiple clinical trials, including this latest study from de Boer and colleagues on kidney function, have failed to reinforce the previously reported benefits of vitamin D.

“The VITAL-DKD study population had nearly normal mean 25-hydroxyvitamin D levels at baseline, leaving open the question of whether the results would have differed had recruitment been restricted to patients with moderate or severe vitamin D deficiency,” they wrote (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17302).

Nevertheless, it seems safe to conclude that the previous associations between vitamin D deficiency and adverse health were “driven by unmeasured residual confounding or reverse causality.

“Without certainty about the ideal approach to vitamin D treatment in advanced CKD, a randomized trial that compared cholecalciferol, exogenous 25-hydroxyvitamin D, and an activated vitamin D analogue vs. placebo could definitively lay to rest multiple remaining questions in the area,” they suggested.

The National Institute of Diabetes and Digestive and Kidney Diseases funded the study. The authors reported numerous potential conflicts of interest, including receiving grants, consulting fees, and equipment and supplies from various pharmaceutical companies and the National Institutes of Health. Dr. Wolf reported having served as a consultant for Akebia, AMAG, Amgen, Ardelyx, Diasorin, and Pharmacosmos. No other disclosures were reported.

SOURCE: de Boer IH et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17380.

 

A new study has found that neither vitamin D nor omega-3 fatty acids are significantly more beneficial than placebo for prevention and treatment of chronic kidney disease in patients with type 2 diabetes, according to Ian H. de Boer, MD, of the University of Washington, Seattle, and coauthors.

Findings of the study were presented at Kidney Week 2019, sponsored the American Society of Nephrology, and published simultaneously in JAMA.

To determine the benefits of either vitamin D or omega-3 fatty acids in regard to kidney function, the researchers conducted a randomized clinical trial of 1,312 patients with type 2 diabetes. The trial was designed to accompany the Vitamin D and Omega-3 Trial (VITAL), which enrolled 25,871 patients to test the two supplements in the prevention of cardiovascular disease and cancer.

Participants in this study – known as VITAL–Diabetic Kidney Disease, designed as an ancillary to VITAL – were assigned to one of four groups: vitamin D plus omega-3 fatty acids (n = 370), vitamin D plus placebo (n = 333), omega-3 fatty acids plus placebo (n = 289), or both placebos (n = 320). The goal was to assess changes in in glomerular filtration rate estimated from serum creatinine and cystatin C (eGFR) after 5 years.

Of the initial 1,312 participants, 934 (71%) finished the study. At 5-year follow-up, patients taking vitamin D had a mean change in eGFR of −12.3 mL/min per 1.73 m2 (95% confidence interval, −13.4 to −11.2), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −11.9) with placebo. Patients taking omega-3 fatty acids had a mean eGFR change of −12.2 mL/min per 1.73 m2 (95% CI, −13.3 to −11.1), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −12.0) with placebo.

The authors noted that the modest number of measurements collected per participant limited the evaluation and analyses. In addition, the study focused broadly on the type 2 diabetes population and not on subgroups, “who may derive more benefit from the study interventions.”

In an accompanying editorial, authors Anika Lucas, MD and Myles Wolf, MD, of Duke University in Durham, N.C., said multiple clinical trials, including this latest study from de Boer and colleagues on kidney function, have failed to reinforce the previously reported benefits of vitamin D.

“The VITAL-DKD study population had nearly normal mean 25-hydroxyvitamin D levels at baseline, leaving open the question of whether the results would have differed had recruitment been restricted to patients with moderate or severe vitamin D deficiency,” they wrote (JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17302).

Nevertheless, it seems safe to conclude that the previous associations between vitamin D deficiency and adverse health were “driven by unmeasured residual confounding or reverse causality.

“Without certainty about the ideal approach to vitamin D treatment in advanced CKD, a randomized trial that compared cholecalciferol, exogenous 25-hydroxyvitamin D, and an activated vitamin D analogue vs. placebo could definitively lay to rest multiple remaining questions in the area,” they suggested.

The National Institute of Diabetes and Digestive and Kidney Diseases funded the study. The authors reported numerous potential conflicts of interest, including receiving grants, consulting fees, and equipment and supplies from various pharmaceutical companies and the National Institutes of Health. Dr. Wolf reported having served as a consultant for Akebia, AMAG, Amgen, Ardelyx, Diasorin, and Pharmacosmos. No other disclosures were reported.

SOURCE: de Boer IH et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17380.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM KIDNEY WEEK 2019

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Vitals

 

Key clinical point: Vitamin D and omega-3 fatty acid supplementation is no better than placebo for preserving kidney function in patients with type 2 diabetes.

Major finding: At 5-year follow-up, patients taking vitamin D had a mean change in eGFR of −12.3 mL/min per 1.73 m2 (95% CI, −13.4 to −11.2), compared with −13.1 mL/min per 1.73 m2 (95% CI, −14.2 to −11.9) with placebo.

Study details: A randomized clinical trial of 1,312 adults with type 2 diabetes.

Disclosures: The National Institute of Diabetes and Digestive and Kidney Diseases funded the study. The authors reported numerous potential conflicts of interest, including receiving grants, consulting fees, and equipment and supplies from various pharmaceutical companies and the National Institutes of Health. Dr. Wolf reported having served as a consultant for Akebia, AMAG, Amgen, Ardelyx, Diasorin, and Pharmacosmos. No other disclosures were reported.

Source: de Boer IH et al. JAMA. 2019 Nov 8. doi: 10.1001/jama.2019.17380.

Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.