Real-world data suggest coprescribing PDE5 inhibitors and nitrates may be safe

Article Type
Changed
Thu, 04/21/2022 - 09:00

As coprescribing drugs for erectile dysfunction and oral organic nitrates for ischemic heart disease (IHD) surged, cardiovascular adverse events did not significantly increase, a new study finds.

The authors of the new research specifically examined how frequently phosphodiesterase type 5 (PDE5) inhibitors, such as Viagra, were prescribed. The U.S. Food and Drug Administration and the European Medicines Agency have warned that these drugs for erectile dysfunction are contraindicated for use with nitrates because of concerns about cardiovascular risks.

“Small, randomized, pharmacologic studies have reported an amplified decrease in blood pressure during controlled coexposure with nitrates and [phosphodiesterase type 5 inhibitors], both in healthy participants and in participants with IHD,” wrote lead author Anders Holt, MD, of Copenhagen University Hospital–Herlev and Gentofte and colleagues, in Annals of Internal Medicine. “Potentially, this increases the risk for vascular ischemic events including myocardial infarction and stroke.”

But there is a scarcity of real-world data showing that using both types of drugs together increase these risks, the researchers noted.

To address this knowledge gap, Dr. Holt and colleagues conducted a retrospective study involving 249,541 Danish men with IHD. In this overall population, from 2000 to 2018, prescriptions for PDE5 inhibitors increased 10-fold, from 3.1 to 30.9 prescriptions per 100 persons per year. Within a subgroup of 42,073 patients continuously prescribed oral organic nitrates, PDE5-inhibitor prescriptions jumped twice that magnitude, from 0.9 to 19.7 prescriptions per 100 persons per year.

Despite this surge in coprescribing, the investigators did not observe a significant increase in either of two composite measures of cardiovascular adverse events. The first composite included ischemic stroke, shock, cardiac arrest, myocardial infarction, or acute coronary arteriography (odds ratio, 0.58; 95% confidence interval, 0.28-1.13). The second composite included drug-related adverse events, angina pectoris, or syncope (OR, 0.73; CI, 0.40-1.32).
 

Lead author speculates on reasons for findings

“I propose several explanations [for these findings],” Dr. Holt said in an interview, “but I want to emphasize that our study does not contain any data to back it up. It is just speculation. First, the observed drop in blood pressure may not cause a condition for which patients seek a hospital. A drop in blood pressure has been shown in pharmacologic trials, but it might not translate to a real-life risk for cardiovascular outcomes. Second, patients could be well informed and adherent to guidance that the prescribing physician has provided. For example, patients are aware of the recommended pause in nitrate treatment before PDE5-inhibitor use and follow these recommendations. Third, nitrates are often taken in the morning, and with the careful assumption that most PDE5-inhibitor activities take place in the evening, the nitrates could be metabolized to a degree such that the synergistic interaction is negligible.”

Dr. Holt went on to suggest a novel clinical approach based on the new findings.

“Coadministration should still be contraindicated due to the proven drop in blood pressure,” he said. “However, perhaps physicians can allow for coprescription if patients are adequately informed.”

A qualitative study is needed to determine how patients and physicians discuss coprescription, including avoidance of coadministration, Dr. Holt added.
 

 

 

Findings call for a reassessment of whether the contraindication is warranted

Robert A. Kloner, MD, PhD, chief science officer at the Huntington Medical Research Institutes in Pasadena, Calif., and professor of medicine at University of Southern California, Los Angeles, previously conducted research exploring drug interactions with PDE5 inhibitors, and in 2018, coauthored a literature review that concluded that PDE5 inhibitors and nitrates are contraindicated.

But now, considering these new findings, Dr. Kloner is offering a fresh perspective.

“This study is reassuring,” Dr. Kloner said in an interview. “I think that it’s time to reassess whether there should be an absolute contraindication, or this should be more of like a warning.”

He noted that in controlled studies, like the ones he previously conducted, PDE5 inhibitors and nitrates were administered “very close to each other, on purpose,” yet this probably doesn’t reflect typical practice, in which clinicians can guide usage based on durations of drug metabolism.

“I think that physicians might be more comfortable now prescribing the drugs at the same time, but then telling patients that they shouldn’t take the two drugs simultaneously; they should wait and take the nitrate 24 hours after the last Viagra, or the nitrate 48 hours after the last Cialis,” Dr. Kloner said. “I suspect that that is happening. I suspect also the fact that people would be more likely to take the nitrate in the morning and the PDE5 inhibitor at night probably also contributes to the safety findings.”

Dr. Kloner noted that blood pressures vary throughout the day based on circadian rhythm, and that the body can adapt to some fluctuations without negative effects.

There could still be some people who experience a drop in blood pressure and get sick from it from the two drugs interacting, but that’s probably not that common, he said.

The study was supported by several grants. The investigators disclosed relationships with Merck, BMS, Bayer, and others. Dr. Kloner consults for Sanofi.

Publications
Topics
Sections

As coprescribing drugs for erectile dysfunction and oral organic nitrates for ischemic heart disease (IHD) surged, cardiovascular adverse events did not significantly increase, a new study finds.

The authors of the new research specifically examined how frequently phosphodiesterase type 5 (PDE5) inhibitors, such as Viagra, were prescribed. The U.S. Food and Drug Administration and the European Medicines Agency have warned that these drugs for erectile dysfunction are contraindicated for use with nitrates because of concerns about cardiovascular risks.

“Small, randomized, pharmacologic studies have reported an amplified decrease in blood pressure during controlled coexposure with nitrates and [phosphodiesterase type 5 inhibitors], both in healthy participants and in participants with IHD,” wrote lead author Anders Holt, MD, of Copenhagen University Hospital–Herlev and Gentofte and colleagues, in Annals of Internal Medicine. “Potentially, this increases the risk for vascular ischemic events including myocardial infarction and stroke.”

But there is a scarcity of real-world data showing that using both types of drugs together increase these risks, the researchers noted.

To address this knowledge gap, Dr. Holt and colleagues conducted a retrospective study involving 249,541 Danish men with IHD. In this overall population, from 2000 to 2018, prescriptions for PDE5 inhibitors increased 10-fold, from 3.1 to 30.9 prescriptions per 100 persons per year. Within a subgroup of 42,073 patients continuously prescribed oral organic nitrates, PDE5-inhibitor prescriptions jumped twice that magnitude, from 0.9 to 19.7 prescriptions per 100 persons per year.

Despite this surge in coprescribing, the investigators did not observe a significant increase in either of two composite measures of cardiovascular adverse events. The first composite included ischemic stroke, shock, cardiac arrest, myocardial infarction, or acute coronary arteriography (odds ratio, 0.58; 95% confidence interval, 0.28-1.13). The second composite included drug-related adverse events, angina pectoris, or syncope (OR, 0.73; CI, 0.40-1.32).
 

Lead author speculates on reasons for findings

“I propose several explanations [for these findings],” Dr. Holt said in an interview, “but I want to emphasize that our study does not contain any data to back it up. It is just speculation. First, the observed drop in blood pressure may not cause a condition for which patients seek a hospital. A drop in blood pressure has been shown in pharmacologic trials, but it might not translate to a real-life risk for cardiovascular outcomes. Second, patients could be well informed and adherent to guidance that the prescribing physician has provided. For example, patients are aware of the recommended pause in nitrate treatment before PDE5-inhibitor use and follow these recommendations. Third, nitrates are often taken in the morning, and with the careful assumption that most PDE5-inhibitor activities take place in the evening, the nitrates could be metabolized to a degree such that the synergistic interaction is negligible.”

Dr. Holt went on to suggest a novel clinical approach based on the new findings.

“Coadministration should still be contraindicated due to the proven drop in blood pressure,” he said. “However, perhaps physicians can allow for coprescription if patients are adequately informed.”

A qualitative study is needed to determine how patients and physicians discuss coprescription, including avoidance of coadministration, Dr. Holt added.
 

 

 

Findings call for a reassessment of whether the contraindication is warranted

Robert A. Kloner, MD, PhD, chief science officer at the Huntington Medical Research Institutes in Pasadena, Calif., and professor of medicine at University of Southern California, Los Angeles, previously conducted research exploring drug interactions with PDE5 inhibitors, and in 2018, coauthored a literature review that concluded that PDE5 inhibitors and nitrates are contraindicated.

But now, considering these new findings, Dr. Kloner is offering a fresh perspective.

“This study is reassuring,” Dr. Kloner said in an interview. “I think that it’s time to reassess whether there should be an absolute contraindication, or this should be more of like a warning.”

He noted that in controlled studies, like the ones he previously conducted, PDE5 inhibitors and nitrates were administered “very close to each other, on purpose,” yet this probably doesn’t reflect typical practice, in which clinicians can guide usage based on durations of drug metabolism.

“I think that physicians might be more comfortable now prescribing the drugs at the same time, but then telling patients that they shouldn’t take the two drugs simultaneously; they should wait and take the nitrate 24 hours after the last Viagra, or the nitrate 48 hours after the last Cialis,” Dr. Kloner said. “I suspect that that is happening. I suspect also the fact that people would be more likely to take the nitrate in the morning and the PDE5 inhibitor at night probably also contributes to the safety findings.”

Dr. Kloner noted that blood pressures vary throughout the day based on circadian rhythm, and that the body can adapt to some fluctuations without negative effects.

There could still be some people who experience a drop in blood pressure and get sick from it from the two drugs interacting, but that’s probably not that common, he said.

The study was supported by several grants. The investigators disclosed relationships with Merck, BMS, Bayer, and others. Dr. Kloner consults for Sanofi.

As coprescribing drugs for erectile dysfunction and oral organic nitrates for ischemic heart disease (IHD) surged, cardiovascular adverse events did not significantly increase, a new study finds.

The authors of the new research specifically examined how frequently phosphodiesterase type 5 (PDE5) inhibitors, such as Viagra, were prescribed. The U.S. Food and Drug Administration and the European Medicines Agency have warned that these drugs for erectile dysfunction are contraindicated for use with nitrates because of concerns about cardiovascular risks.

“Small, randomized, pharmacologic studies have reported an amplified decrease in blood pressure during controlled coexposure with nitrates and [phosphodiesterase type 5 inhibitors], both in healthy participants and in participants with IHD,” wrote lead author Anders Holt, MD, of Copenhagen University Hospital–Herlev and Gentofte and colleagues, in Annals of Internal Medicine. “Potentially, this increases the risk for vascular ischemic events including myocardial infarction and stroke.”

But there is a scarcity of real-world data showing that using both types of drugs together increase these risks, the researchers noted.

To address this knowledge gap, Dr. Holt and colleagues conducted a retrospective study involving 249,541 Danish men with IHD. In this overall population, from 2000 to 2018, prescriptions for PDE5 inhibitors increased 10-fold, from 3.1 to 30.9 prescriptions per 100 persons per year. Within a subgroup of 42,073 patients continuously prescribed oral organic nitrates, PDE5-inhibitor prescriptions jumped twice that magnitude, from 0.9 to 19.7 prescriptions per 100 persons per year.

Despite this surge in coprescribing, the investigators did not observe a significant increase in either of two composite measures of cardiovascular adverse events. The first composite included ischemic stroke, shock, cardiac arrest, myocardial infarction, or acute coronary arteriography (odds ratio, 0.58; 95% confidence interval, 0.28-1.13). The second composite included drug-related adverse events, angina pectoris, or syncope (OR, 0.73; CI, 0.40-1.32).
 

Lead author speculates on reasons for findings

“I propose several explanations [for these findings],” Dr. Holt said in an interview, “but I want to emphasize that our study does not contain any data to back it up. It is just speculation. First, the observed drop in blood pressure may not cause a condition for which patients seek a hospital. A drop in blood pressure has been shown in pharmacologic trials, but it might not translate to a real-life risk for cardiovascular outcomes. Second, patients could be well informed and adherent to guidance that the prescribing physician has provided. For example, patients are aware of the recommended pause in nitrate treatment before PDE5-inhibitor use and follow these recommendations. Third, nitrates are often taken in the morning, and with the careful assumption that most PDE5-inhibitor activities take place in the evening, the nitrates could be metabolized to a degree such that the synergistic interaction is negligible.”

Dr. Holt went on to suggest a novel clinical approach based on the new findings.

“Coadministration should still be contraindicated due to the proven drop in blood pressure,” he said. “However, perhaps physicians can allow for coprescription if patients are adequately informed.”

A qualitative study is needed to determine how patients and physicians discuss coprescription, including avoidance of coadministration, Dr. Holt added.
 

 

 

Findings call for a reassessment of whether the contraindication is warranted

Robert A. Kloner, MD, PhD, chief science officer at the Huntington Medical Research Institutes in Pasadena, Calif., and professor of medicine at University of Southern California, Los Angeles, previously conducted research exploring drug interactions with PDE5 inhibitors, and in 2018, coauthored a literature review that concluded that PDE5 inhibitors and nitrates are contraindicated.

But now, considering these new findings, Dr. Kloner is offering a fresh perspective.

“This study is reassuring,” Dr. Kloner said in an interview. “I think that it’s time to reassess whether there should be an absolute contraindication, or this should be more of like a warning.”

He noted that in controlled studies, like the ones he previously conducted, PDE5 inhibitors and nitrates were administered “very close to each other, on purpose,” yet this probably doesn’t reflect typical practice, in which clinicians can guide usage based on durations of drug metabolism.

“I think that physicians might be more comfortable now prescribing the drugs at the same time, but then telling patients that they shouldn’t take the two drugs simultaneously; they should wait and take the nitrate 24 hours after the last Viagra, or the nitrate 48 hours after the last Cialis,” Dr. Kloner said. “I suspect that that is happening. I suspect also the fact that people would be more likely to take the nitrate in the morning and the PDE5 inhibitor at night probably also contributes to the safety findings.”

Dr. Kloner noted that blood pressures vary throughout the day based on circadian rhythm, and that the body can adapt to some fluctuations without negative effects.

There could still be some people who experience a drop in blood pressure and get sick from it from the two drugs interacting, but that’s probably not that common, he said.

The study was supported by several grants. The investigators disclosed relationships with Merck, BMS, Bayer, and others. Dr. Kloner consults for Sanofi.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM ANNALS OF INTERNAL MEDICINE

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Study: Physical fitness in children linked with concentration, quality of life

Article Type
Changed
Mon, 04/11/2022 - 11:06

Physically fit children have a greater ability to concentrate and better health-related quality of life (HRQOL), according to a new study.

The findings of the German study involving more than 6,500 kids emphasize the importance of cardiorespiratory health in childhood, and support physical fitness initiatives in schools, according to lead author Katharina Köble, MSc, of the Technical University of Munich (Germany), and colleagues.

“Recent studies show that only a few children meet the recommendations of physical activity,” the investigators wrote in Journal of Clinical Medicine.

While the health benefits of physical activity are clearly documented, Ms. Köble and colleagues noted that typical measures of activity, such as accelerometers or self-reported questionnaires, are suboptimal research tools.

“Physical fitness is a more objective parameter to quantify when evaluating health promotion,” the investigators wrote. “Furthermore, cardiorespiratory fitness as part of physical fitness is more strongly related to risk factors of cardiovascular disease than physical activity.”

According to the investigators, physical fitness has also been linked with better concentration and HRQOL, but never in the same population of children.

The new study aimed to address this knowledge gap by assessing 6,533 healthy children aged 6-10 years, approximately half boys and half girls. Associations between physical fitness, concentration, and HRQOL were evaluated using multiple linear regression analysis in participants aged 9-10 years.

Physical fitness was measured using a series of challenges, including curl-ups (pull-ups with palms facing body), push-ups, standing long jump, handgrip strength measurement, and Progressive Aerobic Cardiovascular Endurance Run (PACER). Performing the multistage shuttle run, PACER, “requires participants to maintain the pace set by an audio signal, which progressively increases the intensity every minute.” Results of the PACER test were used to estimate VO2max.

Concentration was measured using the d2-R test, “a paper-pencil cancellation test, where subjects have to cross out all ‘d’ letters with two dashes under a time limit.”

HRQOL was evaluated with the KINDL questionnaire, which covers emotional well-being, physical well-being, everyday functioning (school), friends, family, and self-esteem.

Analysis showed that physical fitness improved with age (P < .001), except for VO2max in girls (P = .129). Concentration also improved with age (P < .001), while HRQOL did not (P = .179).

Among children aged 9-10 years, VO2max scores were strongly associated with both HRQOL (P < .001) and concentration (P < .001).

“VO2max was found to be one of the main factors influencing concentration levels and HRQOL dimensions in primary school children,” the investigators wrote. “Physical fitness, especially cardiorespiratory performance, should therefore be promoted more specifically in school settings to support the promotion of an overall healthy lifestyle in children and adolescents.”
 

Findings are having a real-word impact, according to researcher

In an interview, Ms. Köble noted that the findings are already having a real-world impact.

“We continued data assessment in the long-term and specifically adapted prevention programs in school to the needs of the school children we identified in our study,” she said. “Schools are partially offering specific movement and nutrition classes now.”

In addition, Ms. Köble and colleagues plan on educating teachers about the “urgent need for sufficient physical activity.”

“Academic performance should be considered as an additional health factor in future studies, as well as screen time and eating patterns, as all those variables showed interactions with physical fitness and concentration. In a subanalysis, we showed that children with better physical fitness and concentration values were those who usually went to higher education secondary schools,” they wrote.
 

 

 

VO2max did not correlate with BMI

Gregory Weaver, MD, a pediatrician at Cleveland Clinic Children’s, voiced some concerns about the reliability of the findings. He noted that VO2max did not correlate with body mass index or other measures of physical fitness, and that using the PACER test to estimate VO2max may have skewed the association between physical fitness and concentration.

“It is quite conceivable that children who can maintain the focus to perform maximally on this test will also do well on other tests of attention/concentration,” Dr. Weaver said. “Most children I know would have a very difficult time performing a physical fitness test which requires them to match a recorded pace that slowly increases overtime. I’m not an expert in the area, but it is my understanding that usually VO2max tests involve a treadmill which allows investigators to have complete control over pace.”

Dr. Weaver concluded that more work is needed to determine if physical fitness interventions can have a positive impact on HRQOL and concentration.

“I think the authors of this study attempted to ask an important question about the possible association between physical fitness and concentration among school aged children,” Dr. Weaver said in an interview. “But what is even more vital are studies demonstrating that a change in modifiable health factors like nutrition, physical fitness, or the built environment can improve quality of life. I was hoping the authors would show that an improvement in VO2max over time resulted in an improvement in concentration. Frustratingly, that is not what this article demonstrates.”

The investigators and Dr. Weaver reported no conflicts of interest.

Publications
Topics
Sections

Physically fit children have a greater ability to concentrate and better health-related quality of life (HRQOL), according to a new study.

The findings of the German study involving more than 6,500 kids emphasize the importance of cardiorespiratory health in childhood, and support physical fitness initiatives in schools, according to lead author Katharina Köble, MSc, of the Technical University of Munich (Germany), and colleagues.

“Recent studies show that only a few children meet the recommendations of physical activity,” the investigators wrote in Journal of Clinical Medicine.

While the health benefits of physical activity are clearly documented, Ms. Köble and colleagues noted that typical measures of activity, such as accelerometers or self-reported questionnaires, are suboptimal research tools.

“Physical fitness is a more objective parameter to quantify when evaluating health promotion,” the investigators wrote. “Furthermore, cardiorespiratory fitness as part of physical fitness is more strongly related to risk factors of cardiovascular disease than physical activity.”

According to the investigators, physical fitness has also been linked with better concentration and HRQOL, but never in the same population of children.

The new study aimed to address this knowledge gap by assessing 6,533 healthy children aged 6-10 years, approximately half boys and half girls. Associations between physical fitness, concentration, and HRQOL were evaluated using multiple linear regression analysis in participants aged 9-10 years.

Physical fitness was measured using a series of challenges, including curl-ups (pull-ups with palms facing body), push-ups, standing long jump, handgrip strength measurement, and Progressive Aerobic Cardiovascular Endurance Run (PACER). Performing the multistage shuttle run, PACER, “requires participants to maintain the pace set by an audio signal, which progressively increases the intensity every minute.” Results of the PACER test were used to estimate VO2max.

Concentration was measured using the d2-R test, “a paper-pencil cancellation test, where subjects have to cross out all ‘d’ letters with two dashes under a time limit.”

HRQOL was evaluated with the KINDL questionnaire, which covers emotional well-being, physical well-being, everyday functioning (school), friends, family, and self-esteem.

Analysis showed that physical fitness improved with age (P < .001), except for VO2max in girls (P = .129). Concentration also improved with age (P < .001), while HRQOL did not (P = .179).

Among children aged 9-10 years, VO2max scores were strongly associated with both HRQOL (P < .001) and concentration (P < .001).

“VO2max was found to be one of the main factors influencing concentration levels and HRQOL dimensions in primary school children,” the investigators wrote. “Physical fitness, especially cardiorespiratory performance, should therefore be promoted more specifically in school settings to support the promotion of an overall healthy lifestyle in children and adolescents.”
 

Findings are having a real-word impact, according to researcher

In an interview, Ms. Köble noted that the findings are already having a real-world impact.

“We continued data assessment in the long-term and specifically adapted prevention programs in school to the needs of the school children we identified in our study,” she said. “Schools are partially offering specific movement and nutrition classes now.”

In addition, Ms. Köble and colleagues plan on educating teachers about the “urgent need for sufficient physical activity.”

“Academic performance should be considered as an additional health factor in future studies, as well as screen time and eating patterns, as all those variables showed interactions with physical fitness and concentration. In a subanalysis, we showed that children with better physical fitness and concentration values were those who usually went to higher education secondary schools,” they wrote.
 

 

 

VO2max did not correlate with BMI

Gregory Weaver, MD, a pediatrician at Cleveland Clinic Children’s, voiced some concerns about the reliability of the findings. He noted that VO2max did not correlate with body mass index or other measures of physical fitness, and that using the PACER test to estimate VO2max may have skewed the association between physical fitness and concentration.

“It is quite conceivable that children who can maintain the focus to perform maximally on this test will also do well on other tests of attention/concentration,” Dr. Weaver said. “Most children I know would have a very difficult time performing a physical fitness test which requires them to match a recorded pace that slowly increases overtime. I’m not an expert in the area, but it is my understanding that usually VO2max tests involve a treadmill which allows investigators to have complete control over pace.”

Dr. Weaver concluded that more work is needed to determine if physical fitness interventions can have a positive impact on HRQOL and concentration.

“I think the authors of this study attempted to ask an important question about the possible association between physical fitness and concentration among school aged children,” Dr. Weaver said in an interview. “But what is even more vital are studies demonstrating that a change in modifiable health factors like nutrition, physical fitness, or the built environment can improve quality of life. I was hoping the authors would show that an improvement in VO2max over time resulted in an improvement in concentration. Frustratingly, that is not what this article demonstrates.”

The investigators and Dr. Weaver reported no conflicts of interest.

Physically fit children have a greater ability to concentrate and better health-related quality of life (HRQOL), according to a new study.

The findings of the German study involving more than 6,500 kids emphasize the importance of cardiorespiratory health in childhood, and support physical fitness initiatives in schools, according to lead author Katharina Köble, MSc, of the Technical University of Munich (Germany), and colleagues.

“Recent studies show that only a few children meet the recommendations of physical activity,” the investigators wrote in Journal of Clinical Medicine.

While the health benefits of physical activity are clearly documented, Ms. Köble and colleagues noted that typical measures of activity, such as accelerometers or self-reported questionnaires, are suboptimal research tools.

“Physical fitness is a more objective parameter to quantify when evaluating health promotion,” the investigators wrote. “Furthermore, cardiorespiratory fitness as part of physical fitness is more strongly related to risk factors of cardiovascular disease than physical activity.”

According to the investigators, physical fitness has also been linked with better concentration and HRQOL, but never in the same population of children.

The new study aimed to address this knowledge gap by assessing 6,533 healthy children aged 6-10 years, approximately half boys and half girls. Associations between physical fitness, concentration, and HRQOL were evaluated using multiple linear regression analysis in participants aged 9-10 years.

Physical fitness was measured using a series of challenges, including curl-ups (pull-ups with palms facing body), push-ups, standing long jump, handgrip strength measurement, and Progressive Aerobic Cardiovascular Endurance Run (PACER). Performing the multistage shuttle run, PACER, “requires participants to maintain the pace set by an audio signal, which progressively increases the intensity every minute.” Results of the PACER test were used to estimate VO2max.

Concentration was measured using the d2-R test, “a paper-pencil cancellation test, where subjects have to cross out all ‘d’ letters with two dashes under a time limit.”

HRQOL was evaluated with the KINDL questionnaire, which covers emotional well-being, physical well-being, everyday functioning (school), friends, family, and self-esteem.

Analysis showed that physical fitness improved with age (P < .001), except for VO2max in girls (P = .129). Concentration also improved with age (P < .001), while HRQOL did not (P = .179).

Among children aged 9-10 years, VO2max scores were strongly associated with both HRQOL (P < .001) and concentration (P < .001).

“VO2max was found to be one of the main factors influencing concentration levels and HRQOL dimensions in primary school children,” the investigators wrote. “Physical fitness, especially cardiorespiratory performance, should therefore be promoted more specifically in school settings to support the promotion of an overall healthy lifestyle in children and adolescents.”
 

Findings are having a real-word impact, according to researcher

In an interview, Ms. Köble noted that the findings are already having a real-world impact.

“We continued data assessment in the long-term and specifically adapted prevention programs in school to the needs of the school children we identified in our study,” she said. “Schools are partially offering specific movement and nutrition classes now.”

In addition, Ms. Köble and colleagues plan on educating teachers about the “urgent need for sufficient physical activity.”

“Academic performance should be considered as an additional health factor in future studies, as well as screen time and eating patterns, as all those variables showed interactions with physical fitness and concentration. In a subanalysis, we showed that children with better physical fitness and concentration values were those who usually went to higher education secondary schools,” they wrote.
 

 

 

VO2max did not correlate with BMI

Gregory Weaver, MD, a pediatrician at Cleveland Clinic Children’s, voiced some concerns about the reliability of the findings. He noted that VO2max did not correlate with body mass index or other measures of physical fitness, and that using the PACER test to estimate VO2max may have skewed the association between physical fitness and concentration.

“It is quite conceivable that children who can maintain the focus to perform maximally on this test will also do well on other tests of attention/concentration,” Dr. Weaver said. “Most children I know would have a very difficult time performing a physical fitness test which requires them to match a recorded pace that slowly increases overtime. I’m not an expert in the area, but it is my understanding that usually VO2max tests involve a treadmill which allows investigators to have complete control over pace.”

Dr. Weaver concluded that more work is needed to determine if physical fitness interventions can have a positive impact on HRQOL and concentration.

“I think the authors of this study attempted to ask an important question about the possible association between physical fitness and concentration among school aged children,” Dr. Weaver said in an interview. “But what is even more vital are studies demonstrating that a change in modifiable health factors like nutrition, physical fitness, or the built environment can improve quality of life. I was hoping the authors would show that an improvement in VO2max over time resulted in an improvement in concentration. Frustratingly, that is not what this article demonstrates.”

The investigators and Dr. Weaver reported no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM THE JOURNAL OF CLINICAL MEDICINE

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

New HBV model may open door to more effective antivirals

Long–sought-after breakthrough?
Article Type
Changed
Tue, 03/15/2022 - 17:03

A new mouse model that better represents chronic infection with hepatitis B virus (HBV) in humans may lead to more effective antiviral therapies for HBV, according to investigators.

During human infection, HBV genomes take the form of covalently closed circular DNA (cccDNA), a structure that has thwarted effective antiviral therapy and, until now, creation of an accurate mouse model, reported lead author Zaichao Xu, PhD, of Wuhan (China) University and colleagues.

“As the viral persistence reservoir plays a central role in HBV infection, HBV cccDNA is the key obstacle for a cure,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology.

Although several previous mouse models have approximated this phenomenon with recombinant cccDNA-like molecules (rcccDNA), the present model is the first to achieve genuine cccDNA, which does not naturally occur in mice.

“Although rcccDNA supports persistent viral replication and antigen expression, the nature of rcccDNA may differ from authentic cccDNA, as additional sequences, like LoxP or attR, were inserted into the HBV genome,” the investigators noted.

The new model was created by first constructing an adeno-associated virus vector carrying a replication-deficient HBV1.04-fold genome (AAV-HBV1.04). When injected into mice, the vector led to cccDNA formation via ataxia-telangiectasia and Rad3-related protein (ATR)–mediated DNA damage response, a finding that was confirmed by blocking the same process with ATR inhibitors.

Immediately after injection, mice tested positive for both hepatitis B e antigen (HBeAg) and hepatitis B surface antigen (HBsAg), with peak concentrations after either 4 or 8 weeks depending on dose. HBV DNA was also detected in serum after injection, and 50% of hepatocytes exhibited HBsAg and hepatitis B core protein (HBc) after 1 week. At week 66, HBsAg, HBeAg, and HBc were still detectable in the liver.

“The expression of HBc could only be observed in the liver, but not in other organs or tissues, suggesting that the AAV-HBV1.04 only targeted the mouse liver,” the investigators wrote.

Further experimentation involving known cccDNA-binding proteins supported the similarity between cccDNA in the mouse model and natural infection.

“These results suggested that the chromatinization and transcriptional activation of cccDNA formed in this model dose not differ from wild-type cccDNA formed through infection.”

Next, Dr. Xu and colleagues demonstrated that the infected mice could serve as a reliable model for antiviral research. One week after injection with the vector, mice were treated with entecavir, polyinosinic-polycytidylic acid (poly[I:C]), or phosphate-buffered saline (PBS; control). As anticipated, entecavir suppressed circulating HBV DNA, but not HBsAg, HBeAg, or HBV cccDNA, whereas treatment with poly(I:C) reduced all HBV markers.

“This novel mouse model will provide a unique platform for studying HBV cccDNA and developing novel antivirals to achieve HBV cure,” the investigators concluded.

The study was supported by the National Natural Science Foundation of China, the Fundamental Research Funds for the Central Universities, Hubei Province’s Outstanding Medical Academic Leader Program, and others. The investigators reported no conflicts of interest.

Body

 

On the heels of the wondrous development of curative antiviral agents for hepatitis C virus (HCV), renewed attention has been directed to efforts to bring about the cure of HBV. However, this task will hinge on successful elimination of covalently closed circular DNA (cccDNA), a highly stable form of viral DNA that is exceedingly difficult to eliminate. Efforts to develop successful curative strategies will in turn rely on development of small animal models that support HBV cccDNA formation and virus production, which has until recently proved elusive. In the past several years, several mouse HBV models supporting cccDNA formation have been constructed using adeno-associated vector (AAV)–mediated transduction of a linearized HBV genome. Both the AAV-HBV linear episome and cccDNA have been consistently replicated and detected in these models. While they recapitulate the key steps of the viral life cycle, these models do not, however, lend themselves to direct assessment of cccDNA, which have traditionally required detection of cccDNA in the liver.

Dr. Raymond T. Chung
Xu et al. have now developed a novel mouse model in which generation of HBsAg is directly dependent on generation of cccDNA. This dependence thus yields a simple marker for assessment of cccDNA status and allows monitoring of the therapeutic effects of novel agents targeting cccDNA by simply following HBsAg titers. More studies are required to explore the mechanisms underlying HBV cccDNA formation and elimination, but this work suggests a new way forward to tractably evaluate agents that specifically interrupt cccDNA metabolism, an important step in our systematic march toward HBV cure.
 

Raymond T. Chung, MD, is a professor of medicine at Harvard Medical School and director of the Hepatology and Liver Center at Massachusetts General Hospital, both in Boston. He has no conflicts to disclose.

Publications
Topics
Sections
Body

 

On the heels of the wondrous development of curative antiviral agents for hepatitis C virus (HCV), renewed attention has been directed to efforts to bring about the cure of HBV. However, this task will hinge on successful elimination of covalently closed circular DNA (cccDNA), a highly stable form of viral DNA that is exceedingly difficult to eliminate. Efforts to develop successful curative strategies will in turn rely on development of small animal models that support HBV cccDNA formation and virus production, which has until recently proved elusive. In the past several years, several mouse HBV models supporting cccDNA formation have been constructed using adeno-associated vector (AAV)–mediated transduction of a linearized HBV genome. Both the AAV-HBV linear episome and cccDNA have been consistently replicated and detected in these models. While they recapitulate the key steps of the viral life cycle, these models do not, however, lend themselves to direct assessment of cccDNA, which have traditionally required detection of cccDNA in the liver.

Dr. Raymond T. Chung
Xu et al. have now developed a novel mouse model in which generation of HBsAg is directly dependent on generation of cccDNA. This dependence thus yields a simple marker for assessment of cccDNA status and allows monitoring of the therapeutic effects of novel agents targeting cccDNA by simply following HBsAg titers. More studies are required to explore the mechanisms underlying HBV cccDNA formation and elimination, but this work suggests a new way forward to tractably evaluate agents that specifically interrupt cccDNA metabolism, an important step in our systematic march toward HBV cure.
 

Raymond T. Chung, MD, is a professor of medicine at Harvard Medical School and director of the Hepatology and Liver Center at Massachusetts General Hospital, both in Boston. He has no conflicts to disclose.

Body

 

On the heels of the wondrous development of curative antiviral agents for hepatitis C virus (HCV), renewed attention has been directed to efforts to bring about the cure of HBV. However, this task will hinge on successful elimination of covalently closed circular DNA (cccDNA), a highly stable form of viral DNA that is exceedingly difficult to eliminate. Efforts to develop successful curative strategies will in turn rely on development of small animal models that support HBV cccDNA formation and virus production, which has until recently proved elusive. In the past several years, several mouse HBV models supporting cccDNA formation have been constructed using adeno-associated vector (AAV)–mediated transduction of a linearized HBV genome. Both the AAV-HBV linear episome and cccDNA have been consistently replicated and detected in these models. While they recapitulate the key steps of the viral life cycle, these models do not, however, lend themselves to direct assessment of cccDNA, which have traditionally required detection of cccDNA in the liver.

Dr. Raymond T. Chung
Xu et al. have now developed a novel mouse model in which generation of HBsAg is directly dependent on generation of cccDNA. This dependence thus yields a simple marker for assessment of cccDNA status and allows monitoring of the therapeutic effects of novel agents targeting cccDNA by simply following HBsAg titers. More studies are required to explore the mechanisms underlying HBV cccDNA formation and elimination, but this work suggests a new way forward to tractably evaluate agents that specifically interrupt cccDNA metabolism, an important step in our systematic march toward HBV cure.
 

Raymond T. Chung, MD, is a professor of medicine at Harvard Medical School and director of the Hepatology and Liver Center at Massachusetts General Hospital, both in Boston. He has no conflicts to disclose.

Title
Long–sought-after breakthrough?
Long–sought-after breakthrough?

A new mouse model that better represents chronic infection with hepatitis B virus (HBV) in humans may lead to more effective antiviral therapies for HBV, according to investigators.

During human infection, HBV genomes take the form of covalently closed circular DNA (cccDNA), a structure that has thwarted effective antiviral therapy and, until now, creation of an accurate mouse model, reported lead author Zaichao Xu, PhD, of Wuhan (China) University and colleagues.

“As the viral persistence reservoir plays a central role in HBV infection, HBV cccDNA is the key obstacle for a cure,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology.

Although several previous mouse models have approximated this phenomenon with recombinant cccDNA-like molecules (rcccDNA), the present model is the first to achieve genuine cccDNA, which does not naturally occur in mice.

“Although rcccDNA supports persistent viral replication and antigen expression, the nature of rcccDNA may differ from authentic cccDNA, as additional sequences, like LoxP or attR, were inserted into the HBV genome,” the investigators noted.

The new model was created by first constructing an adeno-associated virus vector carrying a replication-deficient HBV1.04-fold genome (AAV-HBV1.04). When injected into mice, the vector led to cccDNA formation via ataxia-telangiectasia and Rad3-related protein (ATR)–mediated DNA damage response, a finding that was confirmed by blocking the same process with ATR inhibitors.

Immediately after injection, mice tested positive for both hepatitis B e antigen (HBeAg) and hepatitis B surface antigen (HBsAg), with peak concentrations after either 4 or 8 weeks depending on dose. HBV DNA was also detected in serum after injection, and 50% of hepatocytes exhibited HBsAg and hepatitis B core protein (HBc) after 1 week. At week 66, HBsAg, HBeAg, and HBc were still detectable in the liver.

“The expression of HBc could only be observed in the liver, but not in other organs or tissues, suggesting that the AAV-HBV1.04 only targeted the mouse liver,” the investigators wrote.

Further experimentation involving known cccDNA-binding proteins supported the similarity between cccDNA in the mouse model and natural infection.

“These results suggested that the chromatinization and transcriptional activation of cccDNA formed in this model dose not differ from wild-type cccDNA formed through infection.”

Next, Dr. Xu and colleagues demonstrated that the infected mice could serve as a reliable model for antiviral research. One week after injection with the vector, mice were treated with entecavir, polyinosinic-polycytidylic acid (poly[I:C]), or phosphate-buffered saline (PBS; control). As anticipated, entecavir suppressed circulating HBV DNA, but not HBsAg, HBeAg, or HBV cccDNA, whereas treatment with poly(I:C) reduced all HBV markers.

“This novel mouse model will provide a unique platform for studying HBV cccDNA and developing novel antivirals to achieve HBV cure,” the investigators concluded.

The study was supported by the National Natural Science Foundation of China, the Fundamental Research Funds for the Central Universities, Hubei Province’s Outstanding Medical Academic Leader Program, and others. The investigators reported no conflicts of interest.

A new mouse model that better represents chronic infection with hepatitis B virus (HBV) in humans may lead to more effective antiviral therapies for HBV, according to investigators.

During human infection, HBV genomes take the form of covalently closed circular DNA (cccDNA), a structure that has thwarted effective antiviral therapy and, until now, creation of an accurate mouse model, reported lead author Zaichao Xu, PhD, of Wuhan (China) University and colleagues.

“As the viral persistence reservoir plays a central role in HBV infection, HBV cccDNA is the key obstacle for a cure,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology.

Although several previous mouse models have approximated this phenomenon with recombinant cccDNA-like molecules (rcccDNA), the present model is the first to achieve genuine cccDNA, which does not naturally occur in mice.

“Although rcccDNA supports persistent viral replication and antigen expression, the nature of rcccDNA may differ from authentic cccDNA, as additional sequences, like LoxP or attR, were inserted into the HBV genome,” the investigators noted.

The new model was created by first constructing an adeno-associated virus vector carrying a replication-deficient HBV1.04-fold genome (AAV-HBV1.04). When injected into mice, the vector led to cccDNA formation via ataxia-telangiectasia and Rad3-related protein (ATR)–mediated DNA damage response, a finding that was confirmed by blocking the same process with ATR inhibitors.

Immediately after injection, mice tested positive for both hepatitis B e antigen (HBeAg) and hepatitis B surface antigen (HBsAg), with peak concentrations after either 4 or 8 weeks depending on dose. HBV DNA was also detected in serum after injection, and 50% of hepatocytes exhibited HBsAg and hepatitis B core protein (HBc) after 1 week. At week 66, HBsAg, HBeAg, and HBc were still detectable in the liver.

“The expression of HBc could only be observed in the liver, but not in other organs or tissues, suggesting that the AAV-HBV1.04 only targeted the mouse liver,” the investigators wrote.

Further experimentation involving known cccDNA-binding proteins supported the similarity between cccDNA in the mouse model and natural infection.

“These results suggested that the chromatinization and transcriptional activation of cccDNA formed in this model dose not differ from wild-type cccDNA formed through infection.”

Next, Dr. Xu and colleagues demonstrated that the infected mice could serve as a reliable model for antiviral research. One week after injection with the vector, mice were treated with entecavir, polyinosinic-polycytidylic acid (poly[I:C]), or phosphate-buffered saline (PBS; control). As anticipated, entecavir suppressed circulating HBV DNA, but not HBsAg, HBeAg, or HBV cccDNA, whereas treatment with poly(I:C) reduced all HBV markers.

“This novel mouse model will provide a unique platform for studying HBV cccDNA and developing novel antivirals to achieve HBV cure,” the investigators concluded.

The study was supported by the National Natural Science Foundation of China, the Fundamental Research Funds for the Central Universities, Hubei Province’s Outstanding Medical Academic Leader Program, and others. The investigators reported no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CELLULAR AND MOLECULAR GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Bowel ultrasound may overtake colonoscopy in Crohn’s

A 'significant financial burden' avoided
Article Type
Changed
Mon, 04/11/2022 - 16:16

Bowel ultrasound predicts the clinical course of Crohn’s disease for up to 1 year, according to results of a prospective trial involving 225 patients.

After additional confirmation in larger studies, ultrasound could serve as a noninvasive alternative to colonoscopy for monitoring and predicting disease course, reported lead author Mariangela Allocca, MD, PhD, of Humanitas University, Milan, and colleagues.

“Frequent colonoscopies are expensive, invasive, and not well tolerated by patients, thus noninvasive tools for assessment and monitoring are strongly needed,” the investigators wrote in Clinical Gastroenterology and Hepatology. “Bowel ultrasound accurately detects inflammatory bowel disease activity, extent, and complications, particularly in Crohn’s disease. Considering its low cost, minimal invasiveness ... and easy repeatability, bowel ultrasound may be a simple, readily available tool for assessing and monitoring Crohn’s disease.”

To test this hypothesis, Dr. Allocca and colleagues enrolled 225 consecutive patients with ileal and/or colonic Crohn’s disease diagnosed for at least 6 months and managed at a tertiary hospital in Italy. All patients underwent both colonoscopy and bowel ultrasound with no more than 3 months between each procedure.

Colonoscopy results were characterized by the Simplified Endoscopic Score for Crohn’s disease (SES-CD), whereas ultrasound was scored using a several parameters, including bowel wall pattern, bowel thickness, bowel wall flow, presence of complications (abscess, fistula, stricture), and characteristics of mesenteric lymph nodes and tissue. Ultrasound scores were considered high if they exceeded a cut-off of 3.52, which was determined by a receiver operating characteristic curve analysis.

Participants were followed for 12 months after baseline ultrasound. The primary objective was to determine the relationship between baseline ultrasound findings and negative disease course, defined by steroid usage, need for surgery, need for hospitalization, and/or change in therapy. The secondary objective was to understand the relationship between ultrasound findings and endoscopy activity.

Multivariable analysis revealed that ultrasound scores greater than 3.52 predicted a negative clinical disease course for up to one year (odds ratio, 6.97; 95% confidence interval, 2.87-16.93; P < .001), as did the presence of at least one disease complication at baseline (OR, 3.90; 95% CI, 1.21-12.53; P = 0.21). A worse clinical course at one-year was also predicted by a baseline fecal calprotectin value of at least 250 mcg/g (OR, 5.43; 95% CI, 2.25-13.11; P < .001) and male sex (OR, 2.60; 95% CI, 1.12-6.02; P = .025).

Investigators then assessed individual disease outcomes at 12 months and baseline results. For example, high ultrasound score and calprotectin at baseline each predicted the need for treatment escalation. In comparison, disease behavior (inflammatory, stricturing, penetrating) and C reactive protein predicted need for corticosteroids. The only significant predictor of hospitalization a year later was CRP.

“[B]owel ultrasound is able to predict disease course in Crohn’s disease patients,” they wrote. “It may identify patients at high risk of a negative course to adopt effective strategies to prevent any disease progression. Our data need to be confirmed and validated in further large studies.”

The investigators disclosed relationships with Janssen, AbbVie, Mundipharma, and others.

Body

Patients with Crohn’s disease (CD) undergo multiple colonoscopies during their lifetime. Endoscopic assessment is often necessary to determine extent and severity of inflammation to guide choice of therapy, assess mucosal healing on current therapy, and for surveillance examination for colorectal dysplasia. Multiple colonoscopies over a lifetime present a significant financial burden for patients. The invasive nature of the procedure, along with the small but potential risk of perforation and patient discomfort make for an undesirable experience. Cross-sectional imaging offers the advantage of noninvasive modality to assess bowel wall and extraluminal complications related to CD. Bowel ultrasound, performed as point of care imaging by gastroenterologists, is an emerging imaging alternative to visualize the bowel.

In the study by Allocca et al., the authors developed a bowel ultrasound–based score incorporating bowel wall thickness, pattern, flow, and presence of extraluminal complications. The score was developed by comparing ultrasound parameters with colonoscopy findings for each segment of the colon and terminal ileum. In a cohort of 225 patients, a bowel ultrasound score of >3.52 along with at least one extraluminal complication, baseline fecal calprotectin of >250 mcg/g, and male gender were linked with adverse outcomes within 12 months (defined as need for steroids, change of therapy, hospitalization, or surgery).

Dr. Manreet Kaur

While these observations need to be validated externally, this study further consolidates the role for bowel ultrasound as a viable imaging modality to monitor disease and response to therapy in CD. Prior studies have shown bowel ultrasound is a valid alternative to MR enterography – without the expense, limited availability, and need for gadolinium contrast. As the therapeutic targets in IBD move toward mucosa healing, bowel ultrasound offers the promise of a cost-effective, noninvasive, point-of care test that can be performed during an office consultation. The operator dependent nature of this modality may limit its uptake and utilization. The International Bowel Ultrasound Group (IBUS) has collaborated with the European Crohn’s and Colitis organization as well as the Canadian Association of Gastroenterology to establish training and research in bowel ultrasound. Soon, patients can expect a bowel ultrasound to become part of their routine assessment during an office consultation. 

Manreet Kaur, MD, is medical director of the Inflammatory Bowel Disease Center and an associate professor in the division of gastroenterology and hepatology at Baylor College of Medicine, Houston. She has no relevant conflicts of interest.
 

Publications
Topics
Sections
Body

Patients with Crohn’s disease (CD) undergo multiple colonoscopies during their lifetime. Endoscopic assessment is often necessary to determine extent and severity of inflammation to guide choice of therapy, assess mucosal healing on current therapy, and for surveillance examination for colorectal dysplasia. Multiple colonoscopies over a lifetime present a significant financial burden for patients. The invasive nature of the procedure, along with the small but potential risk of perforation and patient discomfort make for an undesirable experience. Cross-sectional imaging offers the advantage of noninvasive modality to assess bowel wall and extraluminal complications related to CD. Bowel ultrasound, performed as point of care imaging by gastroenterologists, is an emerging imaging alternative to visualize the bowel.

In the study by Allocca et al., the authors developed a bowel ultrasound–based score incorporating bowel wall thickness, pattern, flow, and presence of extraluminal complications. The score was developed by comparing ultrasound parameters with colonoscopy findings for each segment of the colon and terminal ileum. In a cohort of 225 patients, a bowel ultrasound score of >3.52 along with at least one extraluminal complication, baseline fecal calprotectin of >250 mcg/g, and male gender were linked with adverse outcomes within 12 months (defined as need for steroids, change of therapy, hospitalization, or surgery).

Dr. Manreet Kaur

While these observations need to be validated externally, this study further consolidates the role for bowel ultrasound as a viable imaging modality to monitor disease and response to therapy in CD. Prior studies have shown bowel ultrasound is a valid alternative to MR enterography – without the expense, limited availability, and need for gadolinium contrast. As the therapeutic targets in IBD move toward mucosa healing, bowel ultrasound offers the promise of a cost-effective, noninvasive, point-of care test that can be performed during an office consultation. The operator dependent nature of this modality may limit its uptake and utilization. The International Bowel Ultrasound Group (IBUS) has collaborated with the European Crohn’s and Colitis organization as well as the Canadian Association of Gastroenterology to establish training and research in bowel ultrasound. Soon, patients can expect a bowel ultrasound to become part of their routine assessment during an office consultation. 

Manreet Kaur, MD, is medical director of the Inflammatory Bowel Disease Center and an associate professor in the division of gastroenterology and hepatology at Baylor College of Medicine, Houston. She has no relevant conflicts of interest.
 

Body

Patients with Crohn’s disease (CD) undergo multiple colonoscopies during their lifetime. Endoscopic assessment is often necessary to determine extent and severity of inflammation to guide choice of therapy, assess mucosal healing on current therapy, and for surveillance examination for colorectal dysplasia. Multiple colonoscopies over a lifetime present a significant financial burden for patients. The invasive nature of the procedure, along with the small but potential risk of perforation and patient discomfort make for an undesirable experience. Cross-sectional imaging offers the advantage of noninvasive modality to assess bowel wall and extraluminal complications related to CD. Bowel ultrasound, performed as point of care imaging by gastroenterologists, is an emerging imaging alternative to visualize the bowel.

In the study by Allocca et al., the authors developed a bowel ultrasound–based score incorporating bowel wall thickness, pattern, flow, and presence of extraluminal complications. The score was developed by comparing ultrasound parameters with colonoscopy findings for each segment of the colon and terminal ileum. In a cohort of 225 patients, a bowel ultrasound score of >3.52 along with at least one extraluminal complication, baseline fecal calprotectin of >250 mcg/g, and male gender were linked with adverse outcomes within 12 months (defined as need for steroids, change of therapy, hospitalization, or surgery).

Dr. Manreet Kaur

While these observations need to be validated externally, this study further consolidates the role for bowel ultrasound as a viable imaging modality to monitor disease and response to therapy in CD. Prior studies have shown bowel ultrasound is a valid alternative to MR enterography – without the expense, limited availability, and need for gadolinium contrast. As the therapeutic targets in IBD move toward mucosa healing, bowel ultrasound offers the promise of a cost-effective, noninvasive, point-of care test that can be performed during an office consultation. The operator dependent nature of this modality may limit its uptake and utilization. The International Bowel Ultrasound Group (IBUS) has collaborated with the European Crohn’s and Colitis organization as well as the Canadian Association of Gastroenterology to establish training and research in bowel ultrasound. Soon, patients can expect a bowel ultrasound to become part of their routine assessment during an office consultation. 

Manreet Kaur, MD, is medical director of the Inflammatory Bowel Disease Center and an associate professor in the division of gastroenterology and hepatology at Baylor College of Medicine, Houston. She has no relevant conflicts of interest.
 

Title
A 'significant financial burden' avoided
A 'significant financial burden' avoided

Bowel ultrasound predicts the clinical course of Crohn’s disease for up to 1 year, according to results of a prospective trial involving 225 patients.

After additional confirmation in larger studies, ultrasound could serve as a noninvasive alternative to colonoscopy for monitoring and predicting disease course, reported lead author Mariangela Allocca, MD, PhD, of Humanitas University, Milan, and colleagues.

“Frequent colonoscopies are expensive, invasive, and not well tolerated by patients, thus noninvasive tools for assessment and monitoring are strongly needed,” the investigators wrote in Clinical Gastroenterology and Hepatology. “Bowel ultrasound accurately detects inflammatory bowel disease activity, extent, and complications, particularly in Crohn’s disease. Considering its low cost, minimal invasiveness ... and easy repeatability, bowel ultrasound may be a simple, readily available tool for assessing and monitoring Crohn’s disease.”

To test this hypothesis, Dr. Allocca and colleagues enrolled 225 consecutive patients with ileal and/or colonic Crohn’s disease diagnosed for at least 6 months and managed at a tertiary hospital in Italy. All patients underwent both colonoscopy and bowel ultrasound with no more than 3 months between each procedure.

Colonoscopy results were characterized by the Simplified Endoscopic Score for Crohn’s disease (SES-CD), whereas ultrasound was scored using a several parameters, including bowel wall pattern, bowel thickness, bowel wall flow, presence of complications (abscess, fistula, stricture), and characteristics of mesenteric lymph nodes and tissue. Ultrasound scores were considered high if they exceeded a cut-off of 3.52, which was determined by a receiver operating characteristic curve analysis.

Participants were followed for 12 months after baseline ultrasound. The primary objective was to determine the relationship between baseline ultrasound findings and negative disease course, defined by steroid usage, need for surgery, need for hospitalization, and/or change in therapy. The secondary objective was to understand the relationship between ultrasound findings and endoscopy activity.

Multivariable analysis revealed that ultrasound scores greater than 3.52 predicted a negative clinical disease course for up to one year (odds ratio, 6.97; 95% confidence interval, 2.87-16.93; P < .001), as did the presence of at least one disease complication at baseline (OR, 3.90; 95% CI, 1.21-12.53; P = 0.21). A worse clinical course at one-year was also predicted by a baseline fecal calprotectin value of at least 250 mcg/g (OR, 5.43; 95% CI, 2.25-13.11; P < .001) and male sex (OR, 2.60; 95% CI, 1.12-6.02; P = .025).

Investigators then assessed individual disease outcomes at 12 months and baseline results. For example, high ultrasound score and calprotectin at baseline each predicted the need for treatment escalation. In comparison, disease behavior (inflammatory, stricturing, penetrating) and C reactive protein predicted need for corticosteroids. The only significant predictor of hospitalization a year later was CRP.

“[B]owel ultrasound is able to predict disease course in Crohn’s disease patients,” they wrote. “It may identify patients at high risk of a negative course to adopt effective strategies to prevent any disease progression. Our data need to be confirmed and validated in further large studies.”

The investigators disclosed relationships with Janssen, AbbVie, Mundipharma, and others.

Bowel ultrasound predicts the clinical course of Crohn’s disease for up to 1 year, according to results of a prospective trial involving 225 patients.

After additional confirmation in larger studies, ultrasound could serve as a noninvasive alternative to colonoscopy for monitoring and predicting disease course, reported lead author Mariangela Allocca, MD, PhD, of Humanitas University, Milan, and colleagues.

“Frequent colonoscopies are expensive, invasive, and not well tolerated by patients, thus noninvasive tools for assessment and monitoring are strongly needed,” the investigators wrote in Clinical Gastroenterology and Hepatology. “Bowel ultrasound accurately detects inflammatory bowel disease activity, extent, and complications, particularly in Crohn’s disease. Considering its low cost, minimal invasiveness ... and easy repeatability, bowel ultrasound may be a simple, readily available tool for assessing and monitoring Crohn’s disease.”

To test this hypothesis, Dr. Allocca and colleagues enrolled 225 consecutive patients with ileal and/or colonic Crohn’s disease diagnosed for at least 6 months and managed at a tertiary hospital in Italy. All patients underwent both colonoscopy and bowel ultrasound with no more than 3 months between each procedure.

Colonoscopy results were characterized by the Simplified Endoscopic Score for Crohn’s disease (SES-CD), whereas ultrasound was scored using a several parameters, including bowel wall pattern, bowel thickness, bowel wall flow, presence of complications (abscess, fistula, stricture), and characteristics of mesenteric lymph nodes and tissue. Ultrasound scores were considered high if they exceeded a cut-off of 3.52, which was determined by a receiver operating characteristic curve analysis.

Participants were followed for 12 months after baseline ultrasound. The primary objective was to determine the relationship between baseline ultrasound findings and negative disease course, defined by steroid usage, need for surgery, need for hospitalization, and/or change in therapy. The secondary objective was to understand the relationship between ultrasound findings and endoscopy activity.

Multivariable analysis revealed that ultrasound scores greater than 3.52 predicted a negative clinical disease course for up to one year (odds ratio, 6.97; 95% confidence interval, 2.87-16.93; P < .001), as did the presence of at least one disease complication at baseline (OR, 3.90; 95% CI, 1.21-12.53; P = 0.21). A worse clinical course at one-year was also predicted by a baseline fecal calprotectin value of at least 250 mcg/g (OR, 5.43; 95% CI, 2.25-13.11; P < .001) and male sex (OR, 2.60; 95% CI, 1.12-6.02; P = .025).

Investigators then assessed individual disease outcomes at 12 months and baseline results. For example, high ultrasound score and calprotectin at baseline each predicted the need for treatment escalation. In comparison, disease behavior (inflammatory, stricturing, penetrating) and C reactive protein predicted need for corticosteroids. The only significant predictor of hospitalization a year later was CRP.

“[B]owel ultrasound is able to predict disease course in Crohn’s disease patients,” they wrote. “It may identify patients at high risk of a negative course to adopt effective strategies to prevent any disease progression. Our data need to be confirmed and validated in further large studies.”

The investigators disclosed relationships with Janssen, AbbVie, Mundipharma, and others.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CLINICAL GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

In-hospital detox or not, anti-CGRPs show efficacy for medication overuse headache

Article Type
Changed
Thu, 12/15/2022 - 15:39

Anti–calcitonin gene-related peptide (anti-CGRP) monoclonal antibodies are effective for patients with chronic migraine and medication overuse headache regardless of detoxification strategy, according to investigators.

Abruptly discontinuing overused analgesics with health care provider oversight – a frequently resource-intensive and challenging process – is no more effective for controlling medication overuse headache than simply advising patients to stop, reported lead author Umberto Pensato, MD, of the University of Bologna, Italy, and colleagues.

“[C]urrently, the abrupt discontinuation of the overused painkiller(s), accompanied by the start of a pharmacological preventive therapy, is the most recommended strategy [for medication overuse headache],” the investigators wrote in Cephalalgia. “While painkiller(s) withdrawal could be accomplished on an outpatient basis in most cases, an in-hospital setting may be required to achieve successful discontinuation in a subgroup of patients with medication overuse headache, further weighing on individual and hospital costs. Additionally hampering this approach, the abrupt discontinuation of the overused painkiller(s) invariably results in disabling withdrawal symptoms for up to 2 weeks, including a transitory worsening of headache, the so-called ‘rebound headache.’ ”
 

Inpatient or outpatient: Does it matter?

According to Dr. Pensato and colleagues, early evidence suggests that previous painkiller withdrawal does not impact the efficacy of anti-CGRPs for medication overuse headache, yet relevant data remain scarce. To address this knowledge gap, they conducted a prospective, real-world study exploring the relationship between detoxification and outcomes after starting anti-CGRP therapy.

Out of 401 patients enrolled based on initiation of erenumab or galcanezumab, 111 satisfied inclusion criteria, including diagnosis of chronic migraine and medication overuse headache, at least 28 days of analgesic usage and headache days per month in the preceding 3 months, and other factors. Of these 111 patients, 83 underwent in-hospital detox, while the remaining 28 patients, who declined detox based on personal reasons or COVID-19–related bed shortage, were advised to discontinue overused medication on an outpatient basis (without oversight).

The primary endpoint was medication overuse headache responder rate after 3 months, as defined by ICHD-3 diagnostic criteria. Secondary endpoints included 6-item headache impact test (HIT-6), monthly headache days (MHD), migraine disability assessment score (MIDAS), mean pain intensity (MPI), monthly pain medication intake (MPMI), baseline predictors of response/refractoriness, and safety.

Three months after starting anti-CGRP therapy, 59% of patients had resolution of medication overuse headache, including 57% in the inpatient detox group and 64% in the outpatient group, a difference that was not statistically significant (P = .4788). Approximately half of the patients (51%) had at least 50% reduction in monthly headache days; although the rate was numerically lower in the inpatient group compared with the outpatient group, the difference was again not significant (51% vs. 54%; P = .8393).

“Our results support the emerging evidence that anti-CGRP drugs may be effective in these patients irrespective of the detoxification program,” the investigators concluded. “Further studies are needed to definitively confirm these results, potentially leading to a paradigm shift in the management of medication overuse headache.”
 

Abrupt or gradual detox?

According to Alan M. Rapoport, MD, clinical professor of neurology at the University of California, Los Angeles, and editor-in-chief of Neurology Reviews, the study was hampered by two major design limitations.

“The biggest problem I see is that the two groups were treated very differently for their detoxification,” Dr. Rapoport said. “One group was detoxified abruptly in the hospital, so the authors were sure that the patients were off acute-care medication before they started their preventives. The other group was advised to stop their medication on an outpatient basis. The issue is that we have no follow-up as to whether the outpatients did or did not abruptly detoxify. A bigger issue was that the two groups were not randomized so there are many other variables that may have come into consideration.”

Still, Dr. Rapoport, a past president of the International Headache Society (IHS), noted that the findings strengthen a growing body of evidence supporting the efficacy of monoclonal antibodies for medication overuse headache regardless of detoxification strategy. He cited a 2020 study by Carlsen and colleagues conducted at the Danish Headache Center in Copenhagen, which reported similar medication overuse headache outcomes across three randomized cohorts whether they received preventive therapy with detoxification, preventive therapy without detoxification, or detoxification followed 2 months later by preventive therapy.

“What I have noticed since we have had monoclonal antibodies in our armamentarium is that these drugs work very well even when the patient has not fully detoxified,” Dr. Rapoport said. “What I do with my patients is not teach them how to detoxify now, but simply educate them to take fewer acute care medications as their headaches get better from the monoclonal antibodies; they should try to take fewer acute care medications for milder, shorter headaches, and just let them go away on their own. Previous research suggests that even when a patient is not educated at all about medication overuse headache and the reason for detoxification, monoclonal antibodies still work in the presence of medication overuse headache, and improve it.”

The investigators disclosed relationships with Allergan, Novartis, Teva, and others. Dr. Rapoport is on the speakers bureau for AbbVie.

Issue
Neurology Reviews - 30(4)
Publications
Topics
Sections

Anti–calcitonin gene-related peptide (anti-CGRP) monoclonal antibodies are effective for patients with chronic migraine and medication overuse headache regardless of detoxification strategy, according to investigators.

Abruptly discontinuing overused analgesics with health care provider oversight – a frequently resource-intensive and challenging process – is no more effective for controlling medication overuse headache than simply advising patients to stop, reported lead author Umberto Pensato, MD, of the University of Bologna, Italy, and colleagues.

“[C]urrently, the abrupt discontinuation of the overused painkiller(s), accompanied by the start of a pharmacological preventive therapy, is the most recommended strategy [for medication overuse headache],” the investigators wrote in Cephalalgia. “While painkiller(s) withdrawal could be accomplished on an outpatient basis in most cases, an in-hospital setting may be required to achieve successful discontinuation in a subgroup of patients with medication overuse headache, further weighing on individual and hospital costs. Additionally hampering this approach, the abrupt discontinuation of the overused painkiller(s) invariably results in disabling withdrawal symptoms for up to 2 weeks, including a transitory worsening of headache, the so-called ‘rebound headache.’ ”
 

Inpatient or outpatient: Does it matter?

According to Dr. Pensato and colleagues, early evidence suggests that previous painkiller withdrawal does not impact the efficacy of anti-CGRPs for medication overuse headache, yet relevant data remain scarce. To address this knowledge gap, they conducted a prospective, real-world study exploring the relationship between detoxification and outcomes after starting anti-CGRP therapy.

Out of 401 patients enrolled based on initiation of erenumab or galcanezumab, 111 satisfied inclusion criteria, including diagnosis of chronic migraine and medication overuse headache, at least 28 days of analgesic usage and headache days per month in the preceding 3 months, and other factors. Of these 111 patients, 83 underwent in-hospital detox, while the remaining 28 patients, who declined detox based on personal reasons or COVID-19–related bed shortage, were advised to discontinue overused medication on an outpatient basis (without oversight).

The primary endpoint was medication overuse headache responder rate after 3 months, as defined by ICHD-3 diagnostic criteria. Secondary endpoints included 6-item headache impact test (HIT-6), monthly headache days (MHD), migraine disability assessment score (MIDAS), mean pain intensity (MPI), monthly pain medication intake (MPMI), baseline predictors of response/refractoriness, and safety.

Three months after starting anti-CGRP therapy, 59% of patients had resolution of medication overuse headache, including 57% in the inpatient detox group and 64% in the outpatient group, a difference that was not statistically significant (P = .4788). Approximately half of the patients (51%) had at least 50% reduction in monthly headache days; although the rate was numerically lower in the inpatient group compared with the outpatient group, the difference was again not significant (51% vs. 54%; P = .8393).

“Our results support the emerging evidence that anti-CGRP drugs may be effective in these patients irrespective of the detoxification program,” the investigators concluded. “Further studies are needed to definitively confirm these results, potentially leading to a paradigm shift in the management of medication overuse headache.”
 

Abrupt or gradual detox?

According to Alan M. Rapoport, MD, clinical professor of neurology at the University of California, Los Angeles, and editor-in-chief of Neurology Reviews, the study was hampered by two major design limitations.

“The biggest problem I see is that the two groups were treated very differently for their detoxification,” Dr. Rapoport said. “One group was detoxified abruptly in the hospital, so the authors were sure that the patients were off acute-care medication before they started their preventives. The other group was advised to stop their medication on an outpatient basis. The issue is that we have no follow-up as to whether the outpatients did or did not abruptly detoxify. A bigger issue was that the two groups were not randomized so there are many other variables that may have come into consideration.”

Still, Dr. Rapoport, a past president of the International Headache Society (IHS), noted that the findings strengthen a growing body of evidence supporting the efficacy of monoclonal antibodies for medication overuse headache regardless of detoxification strategy. He cited a 2020 study by Carlsen and colleagues conducted at the Danish Headache Center in Copenhagen, which reported similar medication overuse headache outcomes across three randomized cohorts whether they received preventive therapy with detoxification, preventive therapy without detoxification, or detoxification followed 2 months later by preventive therapy.

“What I have noticed since we have had monoclonal antibodies in our armamentarium is that these drugs work very well even when the patient has not fully detoxified,” Dr. Rapoport said. “What I do with my patients is not teach them how to detoxify now, but simply educate them to take fewer acute care medications as their headaches get better from the monoclonal antibodies; they should try to take fewer acute care medications for milder, shorter headaches, and just let them go away on their own. Previous research suggests that even when a patient is not educated at all about medication overuse headache and the reason for detoxification, monoclonal antibodies still work in the presence of medication overuse headache, and improve it.”

The investigators disclosed relationships with Allergan, Novartis, Teva, and others. Dr. Rapoport is on the speakers bureau for AbbVie.

Anti–calcitonin gene-related peptide (anti-CGRP) monoclonal antibodies are effective for patients with chronic migraine and medication overuse headache regardless of detoxification strategy, according to investigators.

Abruptly discontinuing overused analgesics with health care provider oversight – a frequently resource-intensive and challenging process – is no more effective for controlling medication overuse headache than simply advising patients to stop, reported lead author Umberto Pensato, MD, of the University of Bologna, Italy, and colleagues.

“[C]urrently, the abrupt discontinuation of the overused painkiller(s), accompanied by the start of a pharmacological preventive therapy, is the most recommended strategy [for medication overuse headache],” the investigators wrote in Cephalalgia. “While painkiller(s) withdrawal could be accomplished on an outpatient basis in most cases, an in-hospital setting may be required to achieve successful discontinuation in a subgroup of patients with medication overuse headache, further weighing on individual and hospital costs. Additionally hampering this approach, the abrupt discontinuation of the overused painkiller(s) invariably results in disabling withdrawal symptoms for up to 2 weeks, including a transitory worsening of headache, the so-called ‘rebound headache.’ ”
 

Inpatient or outpatient: Does it matter?

According to Dr. Pensato and colleagues, early evidence suggests that previous painkiller withdrawal does not impact the efficacy of anti-CGRPs for medication overuse headache, yet relevant data remain scarce. To address this knowledge gap, they conducted a prospective, real-world study exploring the relationship between detoxification and outcomes after starting anti-CGRP therapy.

Out of 401 patients enrolled based on initiation of erenumab or galcanezumab, 111 satisfied inclusion criteria, including diagnosis of chronic migraine and medication overuse headache, at least 28 days of analgesic usage and headache days per month in the preceding 3 months, and other factors. Of these 111 patients, 83 underwent in-hospital detox, while the remaining 28 patients, who declined detox based on personal reasons or COVID-19–related bed shortage, were advised to discontinue overused medication on an outpatient basis (without oversight).

The primary endpoint was medication overuse headache responder rate after 3 months, as defined by ICHD-3 diagnostic criteria. Secondary endpoints included 6-item headache impact test (HIT-6), monthly headache days (MHD), migraine disability assessment score (MIDAS), mean pain intensity (MPI), monthly pain medication intake (MPMI), baseline predictors of response/refractoriness, and safety.

Three months after starting anti-CGRP therapy, 59% of patients had resolution of medication overuse headache, including 57% in the inpatient detox group and 64% in the outpatient group, a difference that was not statistically significant (P = .4788). Approximately half of the patients (51%) had at least 50% reduction in monthly headache days; although the rate was numerically lower in the inpatient group compared with the outpatient group, the difference was again not significant (51% vs. 54%; P = .8393).

“Our results support the emerging evidence that anti-CGRP drugs may be effective in these patients irrespective of the detoxification program,” the investigators concluded. “Further studies are needed to definitively confirm these results, potentially leading to a paradigm shift in the management of medication overuse headache.”
 

Abrupt or gradual detox?

According to Alan M. Rapoport, MD, clinical professor of neurology at the University of California, Los Angeles, and editor-in-chief of Neurology Reviews, the study was hampered by two major design limitations.

“The biggest problem I see is that the two groups were treated very differently for their detoxification,” Dr. Rapoport said. “One group was detoxified abruptly in the hospital, so the authors were sure that the patients were off acute-care medication before they started their preventives. The other group was advised to stop their medication on an outpatient basis. The issue is that we have no follow-up as to whether the outpatients did or did not abruptly detoxify. A bigger issue was that the two groups were not randomized so there are many other variables that may have come into consideration.”

Still, Dr. Rapoport, a past president of the International Headache Society (IHS), noted that the findings strengthen a growing body of evidence supporting the efficacy of monoclonal antibodies for medication overuse headache regardless of detoxification strategy. He cited a 2020 study by Carlsen and colleagues conducted at the Danish Headache Center in Copenhagen, which reported similar medication overuse headache outcomes across three randomized cohorts whether they received preventive therapy with detoxification, preventive therapy without detoxification, or detoxification followed 2 months later by preventive therapy.

“What I have noticed since we have had monoclonal antibodies in our armamentarium is that these drugs work very well even when the patient has not fully detoxified,” Dr. Rapoport said. “What I do with my patients is not teach them how to detoxify now, but simply educate them to take fewer acute care medications as their headaches get better from the monoclonal antibodies; they should try to take fewer acute care medications for milder, shorter headaches, and just let them go away on their own. Previous research suggests that even when a patient is not educated at all about medication overuse headache and the reason for detoxification, monoclonal antibodies still work in the presence of medication overuse headache, and improve it.”

The investigators disclosed relationships with Allergan, Novartis, Teva, and others. Dr. Rapoport is on the speakers bureau for AbbVie.

Issue
Neurology Reviews - 30(4)
Issue
Neurology Reviews - 30(4)
Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CEPHALALGIA

Citation Override
Publish date: February 25, 2022
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

ILAE offers first guide to treating depression in epilepsy

Article Type
Changed
Thu, 12/15/2022 - 15:39

The International League Against Epilepsy (ILAE) has issued recommendations for treating depression in patients with epilepsy.

The new guidance highlights the high prevalence of depression among patients with epilepsy while offering the first systematic approach to treatment, reported lead author Marco Mula, MD, PhD, of Atkinson Morley Regional Neuroscience Centre at St George’s University Hospital, London, and colleagues.

“Despite evidence that depression represents a frequently encountered comorbidity [among patients with epilepsy], data on the treatment of depression in epilepsy [are] still limited and recommendations rely mostly on individual clinical experience and expertise,” the investigators wrote in Epilepsia.

Recommendations cover first-line treatment of unipolar depression in epilepsy without other psychiatric disorders.

For patients with mild depression, the guidance supports psychological intervention without pharmacologic therapy; however, if the patient wishes to use medication, has had a positive response to medication in the past, or nonpharmacologic treatments have previously failed or are unavailable, then SSRIs should be considered first-choice therapy. For moderate to severe depression, SSRIs are the first choice, according to Dr. Mula and colleagues.

“It has to be acknowledged that there is considerable debate in the psychiatric literature about the treatment of mild depression in adults,” the investigators noted. “A patient-level meta-analysis pointed out that the magnitude of benefit of antidepressant medications compared with placebo increases with severity of depression symptoms and it may be minimal or nonexistent, on average, in patients with mild or moderate symptoms.”

If a patient does not respond to first-line therapy, then venlafaxine should be considered, according to the guidance. When a patient does respond to therapy, treatment should be continued for at least 6 months, and when residual symptoms persist, treatment should be continued until resolution.

“In people with depression it is established that around two-thirds of patients do not achieve full remission with first-line treatment,” Dr. Mula and colleagues wrote. “In people with epilepsy, current data show that up to 50% of patients do not achieve full remission from depression. For this reason, augmentation strategies are often needed. They should be adopted by psychiatrists, neuropsychiatrists, or mental health professionals familiar with such therapeutic strategies.”

Beyond these key recommendations, the guidance covers a range of additional topics, including other pharmacologic options, medication discontinuation strategies, electroconvulsive therapy, light therapy, exercise training, vagus nerve stimulation, and repetitive transcranial magnetic stimulation.
 

Useful advice that counters common misconceptions

According to Jacqueline A. French, MD, a professor at NYU Langone Medical Center, Dr. Mula and colleagues are “top notch,” and their recommendations “hit every nail on the head.”

Dr. Jacqueline A. French

Dr. French, chief medical officer of The Epilepsy Foundation, emphasized the importance of the publication, which addresses two common misconceptions within the medical community: First, that standard antidepressants are insufficient to treat depression in patients with epilepsy, and second, that antidepressants may trigger seizures.

“The first purpose [of the publication] is to say, yes, these antidepressants do work,” Dr. French said, “and no, they don’t worsen seizures, and you can use them safely, and they are appropriate to use.”

Dr. French explained that managing depression remains a practice gap among epileptologists and neurologists because it is a diagnosis that doesn’t traditionally fall into their purview, yet many patients with epilepsy forgo visiting their primary care providers, who more frequently diagnose and manage depression. Dr. French agreed with the guidance that epilepsy specialists should fill this gap.

“We need to at least be able to take people through their first antidepressant, even though we were not trained to be psychiatrists,” Dr. French said. “That’s part of the best care of our patients.”

Imad Najm, MD, director of the Charles Shor Epilepsy Center, Cleveland Clinic, said the recommendations are a step forward in the field, as they are supported by clinical data, instead of just clinical experience and expertise.

Dr. Imad Najm

Still, Dr. Najm noted that more work is needed to stratify risk of depression in epilepsy and evaluate a possible causal relationship between epilepsy therapies and depression.

He went on to emphasizes the scale of issue at hand, and the stakes involved.

“Depression, anxiety, and psychosis affect a large number of patients with epilepsy,” Dr. Najm said. “Clinical screening and recognition of these comorbidities leads to the institution of treatment options and significant improvement in quality of life. Mental health professionals should be an integral part of any comprehensive epilepsy center.”

The investigators disclosed relationships with Esai, UCB, Elsevier, and others. Dr. French is indirectly involved with multiple pharmaceutical companies developing epilepsy drugs through her role as director of The Epilepsy Study Consortium, a nonprofit organization. Dr. Najm reported no conflicts of interest.

Issue
Neurology Reviews - 30(3)
Publications
Topics
Sections

The International League Against Epilepsy (ILAE) has issued recommendations for treating depression in patients with epilepsy.

The new guidance highlights the high prevalence of depression among patients with epilepsy while offering the first systematic approach to treatment, reported lead author Marco Mula, MD, PhD, of Atkinson Morley Regional Neuroscience Centre at St George’s University Hospital, London, and colleagues.

“Despite evidence that depression represents a frequently encountered comorbidity [among patients with epilepsy], data on the treatment of depression in epilepsy [are] still limited and recommendations rely mostly on individual clinical experience and expertise,” the investigators wrote in Epilepsia.

Recommendations cover first-line treatment of unipolar depression in epilepsy without other psychiatric disorders.

For patients with mild depression, the guidance supports psychological intervention without pharmacologic therapy; however, if the patient wishes to use medication, has had a positive response to medication in the past, or nonpharmacologic treatments have previously failed or are unavailable, then SSRIs should be considered first-choice therapy. For moderate to severe depression, SSRIs are the first choice, according to Dr. Mula and colleagues.

“It has to be acknowledged that there is considerable debate in the psychiatric literature about the treatment of mild depression in adults,” the investigators noted. “A patient-level meta-analysis pointed out that the magnitude of benefit of antidepressant medications compared with placebo increases with severity of depression symptoms and it may be minimal or nonexistent, on average, in patients with mild or moderate symptoms.”

If a patient does not respond to first-line therapy, then venlafaxine should be considered, according to the guidance. When a patient does respond to therapy, treatment should be continued for at least 6 months, and when residual symptoms persist, treatment should be continued until resolution.

“In people with depression it is established that around two-thirds of patients do not achieve full remission with first-line treatment,” Dr. Mula and colleagues wrote. “In people with epilepsy, current data show that up to 50% of patients do not achieve full remission from depression. For this reason, augmentation strategies are often needed. They should be adopted by psychiatrists, neuropsychiatrists, or mental health professionals familiar with such therapeutic strategies.”

Beyond these key recommendations, the guidance covers a range of additional topics, including other pharmacologic options, medication discontinuation strategies, electroconvulsive therapy, light therapy, exercise training, vagus nerve stimulation, and repetitive transcranial magnetic stimulation.
 

Useful advice that counters common misconceptions

According to Jacqueline A. French, MD, a professor at NYU Langone Medical Center, Dr. Mula and colleagues are “top notch,” and their recommendations “hit every nail on the head.”

Dr. Jacqueline A. French

Dr. French, chief medical officer of The Epilepsy Foundation, emphasized the importance of the publication, which addresses two common misconceptions within the medical community: First, that standard antidepressants are insufficient to treat depression in patients with epilepsy, and second, that antidepressants may trigger seizures.

“The first purpose [of the publication] is to say, yes, these antidepressants do work,” Dr. French said, “and no, they don’t worsen seizures, and you can use them safely, and they are appropriate to use.”

Dr. French explained that managing depression remains a practice gap among epileptologists and neurologists because it is a diagnosis that doesn’t traditionally fall into their purview, yet many patients with epilepsy forgo visiting their primary care providers, who more frequently diagnose and manage depression. Dr. French agreed with the guidance that epilepsy specialists should fill this gap.

“We need to at least be able to take people through their first antidepressant, even though we were not trained to be psychiatrists,” Dr. French said. “That’s part of the best care of our patients.”

Imad Najm, MD, director of the Charles Shor Epilepsy Center, Cleveland Clinic, said the recommendations are a step forward in the field, as they are supported by clinical data, instead of just clinical experience and expertise.

Dr. Imad Najm

Still, Dr. Najm noted that more work is needed to stratify risk of depression in epilepsy and evaluate a possible causal relationship between epilepsy therapies and depression.

He went on to emphasizes the scale of issue at hand, and the stakes involved.

“Depression, anxiety, and psychosis affect a large number of patients with epilepsy,” Dr. Najm said. “Clinical screening and recognition of these comorbidities leads to the institution of treatment options and significant improvement in quality of life. Mental health professionals should be an integral part of any comprehensive epilepsy center.”

The investigators disclosed relationships with Esai, UCB, Elsevier, and others. Dr. French is indirectly involved with multiple pharmaceutical companies developing epilepsy drugs through her role as director of The Epilepsy Study Consortium, a nonprofit organization. Dr. Najm reported no conflicts of interest.

The International League Against Epilepsy (ILAE) has issued recommendations for treating depression in patients with epilepsy.

The new guidance highlights the high prevalence of depression among patients with epilepsy while offering the first systematic approach to treatment, reported lead author Marco Mula, MD, PhD, of Atkinson Morley Regional Neuroscience Centre at St George’s University Hospital, London, and colleagues.

“Despite evidence that depression represents a frequently encountered comorbidity [among patients with epilepsy], data on the treatment of depression in epilepsy [are] still limited and recommendations rely mostly on individual clinical experience and expertise,” the investigators wrote in Epilepsia.

Recommendations cover first-line treatment of unipolar depression in epilepsy without other psychiatric disorders.

For patients with mild depression, the guidance supports psychological intervention without pharmacologic therapy; however, if the patient wishes to use medication, has had a positive response to medication in the past, or nonpharmacologic treatments have previously failed or are unavailable, then SSRIs should be considered first-choice therapy. For moderate to severe depression, SSRIs are the first choice, according to Dr. Mula and colleagues.

“It has to be acknowledged that there is considerable debate in the psychiatric literature about the treatment of mild depression in adults,” the investigators noted. “A patient-level meta-analysis pointed out that the magnitude of benefit of antidepressant medications compared with placebo increases with severity of depression symptoms and it may be minimal or nonexistent, on average, in patients with mild or moderate symptoms.”

If a patient does not respond to first-line therapy, then venlafaxine should be considered, according to the guidance. When a patient does respond to therapy, treatment should be continued for at least 6 months, and when residual symptoms persist, treatment should be continued until resolution.

“In people with depression it is established that around two-thirds of patients do not achieve full remission with first-line treatment,” Dr. Mula and colleagues wrote. “In people with epilepsy, current data show that up to 50% of patients do not achieve full remission from depression. For this reason, augmentation strategies are often needed. They should be adopted by psychiatrists, neuropsychiatrists, or mental health professionals familiar with such therapeutic strategies.”

Beyond these key recommendations, the guidance covers a range of additional topics, including other pharmacologic options, medication discontinuation strategies, electroconvulsive therapy, light therapy, exercise training, vagus nerve stimulation, and repetitive transcranial magnetic stimulation.
 

Useful advice that counters common misconceptions

According to Jacqueline A. French, MD, a professor at NYU Langone Medical Center, Dr. Mula and colleagues are “top notch,” and their recommendations “hit every nail on the head.”

Dr. Jacqueline A. French

Dr. French, chief medical officer of The Epilepsy Foundation, emphasized the importance of the publication, which addresses two common misconceptions within the medical community: First, that standard antidepressants are insufficient to treat depression in patients with epilepsy, and second, that antidepressants may trigger seizures.

“The first purpose [of the publication] is to say, yes, these antidepressants do work,” Dr. French said, “and no, they don’t worsen seizures, and you can use them safely, and they are appropriate to use.”

Dr. French explained that managing depression remains a practice gap among epileptologists and neurologists because it is a diagnosis that doesn’t traditionally fall into their purview, yet many patients with epilepsy forgo visiting their primary care providers, who more frequently diagnose and manage depression. Dr. French agreed with the guidance that epilepsy specialists should fill this gap.

“We need to at least be able to take people through their first antidepressant, even though we were not trained to be psychiatrists,” Dr. French said. “That’s part of the best care of our patients.”

Imad Najm, MD, director of the Charles Shor Epilepsy Center, Cleveland Clinic, said the recommendations are a step forward in the field, as they are supported by clinical data, instead of just clinical experience and expertise.

Dr. Imad Najm

Still, Dr. Najm noted that more work is needed to stratify risk of depression in epilepsy and evaluate a possible causal relationship between epilepsy therapies and depression.

He went on to emphasizes the scale of issue at hand, and the stakes involved.

“Depression, anxiety, and psychosis affect a large number of patients with epilepsy,” Dr. Najm said. “Clinical screening and recognition of these comorbidities leads to the institution of treatment options and significant improvement in quality of life. Mental health professionals should be an integral part of any comprehensive epilepsy center.”

The investigators disclosed relationships with Esai, UCB, Elsevier, and others. Dr. French is indirectly involved with multiple pharmaceutical companies developing epilepsy drugs through her role as director of The Epilepsy Study Consortium, a nonprofit organization. Dr. Najm reported no conflicts of interest.

Issue
Neurology Reviews - 30(3)
Issue
Neurology Reviews - 30(3)
Publications
Publications
Topics
Article Type
Sections
Article Source

FROM EPILEPSIA

Citation Override
Publish date: February 14, 2022
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Moderate-vigorous stepping seen to lower diabetes risk in older women

Article Type
Changed
Tue, 05/03/2022 - 15:02

More steps per day, particularly at a higher intensity, may reduce the risk of type 2 diabetes in older women, based on a prospective cohort study.

The link between daily stepping and diabetes was not significantly modified by body mass index (BMI) or other common diabetes risk factors, suggesting that the relationship is highly generalizable, lead author Alexis C. Garduno, MPH, a PhD student at the University of California, San Diego, and colleagues reported.

“Physical activity is a key modifiable behavior for diabetes prevention and management,” the investigators wrote in Diabetes Care. “Many prevention studies have demonstrated that regular physical activity, along with improved diet, reduces the risk of diabetes in adults. ... To the best of our knowledge, there are few studies examining the association between objectively measured steps per day and incident diabetes in a community-based setting.”

To this end, the investigators analyzed data from 4,838 older, community-living women in the Objective Physical Activity and Cardiovascular Health Study. Upon enrollment, women were without physician-diagnosed diabetes and had a mean age of 78.9 years. For 1 week, participants wore ActiGraph GT3X+ accelerometers to measure steps per day, as well as step intensity, graded as light or moderate to vigorous.

The relationship between daily activity and diabetes was analyzed using three multivariate models: The first included race/ethnicity and age; the second also included family history of diabetes, education, physical functioning, self-rated health, smoking status, and alcohol consumption; and the third added BMI, “a potential mediator in the causal pathway between steps per day and diabetes,” the investigators wrote.

Participants took an average of 3,729 steps per day, divided roughly evenly between light and moderate to vigorous intensity.

After a median follow-up of 5.7 years, 8.1% of women developed diabetes. The least-adjusted model showed a 14% reduction in diabetes risk per 2,000 steps (hazard ratio, 0.86; 95% confidence interval, 0.80-0.92; P = .007), whereas the second model, adjusting for more confounding variables, showed a 12% reduction in diabetes risk per 2,000 steps (HR, 0.88; 95% CI, 0.78-1.00; P = .045).

The final model, which added BMI, showed a 10% reduction in risk, although it didn’t reach statistical significance (HR, 0.90; 95% CI, 0.80-1.02; P = .11). Furthermore, accelerated failure time models suggested that BMI did not significantly impact the link between steps and diabetes (proportion mediated, 17.7%;95% CI, –55.0 to 142.0; P = .09). Further analyses also found no significant interactions between BMI or other possible confounders.

“The steps per day–diabetes association was not modified by age, race/ethnicity, BMI, physical functioning, or family history of diabetes, which supports the generalizability of these findings to community-living older women,” the investigators wrote.

Increased stepping intensity also appeared to lower risk of diabetes. After adjusting for confounding variables, light stepping was not linked to reduced risk (HR, 0.97; 95% CI, 0.73-1.29; P = .83), whereas moderate to vigorous stepping reduced risk by 14% per 2,000 steps (HR, 0.86; 95% CI, 0.74-1.00; P = .04).

“This study provides evidence supporting an association between steps per day and lower incident diabetes,” the investigators concluded. “While further work is needed to identify whether there is a minimum number of steps per day that results in a clinically significant reduction of diabetes and to evaluate the role that step intensity plays in diabetes etiology for older adults, findings from this study suggest that moderate-vigorous–intensity steps may be more important than lower-intensity steps with respect to incident diabetes. Steps per day–based interventions are needed to advance diabetes prevention science in older adults.”

The study was supported by the National Institute on Aging, the National Institute of Diabetes and Digestive and Kidney Diseases, the Tobacco-Related Disease Research Program, and others. The investigators had no potential conflicts of interest.

Publications
Topics
Sections

More steps per day, particularly at a higher intensity, may reduce the risk of type 2 diabetes in older women, based on a prospective cohort study.

The link between daily stepping and diabetes was not significantly modified by body mass index (BMI) or other common diabetes risk factors, suggesting that the relationship is highly generalizable, lead author Alexis C. Garduno, MPH, a PhD student at the University of California, San Diego, and colleagues reported.

“Physical activity is a key modifiable behavior for diabetes prevention and management,” the investigators wrote in Diabetes Care. “Many prevention studies have demonstrated that regular physical activity, along with improved diet, reduces the risk of diabetes in adults. ... To the best of our knowledge, there are few studies examining the association between objectively measured steps per day and incident diabetes in a community-based setting.”

To this end, the investigators analyzed data from 4,838 older, community-living women in the Objective Physical Activity and Cardiovascular Health Study. Upon enrollment, women were without physician-diagnosed diabetes and had a mean age of 78.9 years. For 1 week, participants wore ActiGraph GT3X+ accelerometers to measure steps per day, as well as step intensity, graded as light or moderate to vigorous.

The relationship between daily activity and diabetes was analyzed using three multivariate models: The first included race/ethnicity and age; the second also included family history of diabetes, education, physical functioning, self-rated health, smoking status, and alcohol consumption; and the third added BMI, “a potential mediator in the causal pathway between steps per day and diabetes,” the investigators wrote.

Participants took an average of 3,729 steps per day, divided roughly evenly between light and moderate to vigorous intensity.

After a median follow-up of 5.7 years, 8.1% of women developed diabetes. The least-adjusted model showed a 14% reduction in diabetes risk per 2,000 steps (hazard ratio, 0.86; 95% confidence interval, 0.80-0.92; P = .007), whereas the second model, adjusting for more confounding variables, showed a 12% reduction in diabetes risk per 2,000 steps (HR, 0.88; 95% CI, 0.78-1.00; P = .045).

The final model, which added BMI, showed a 10% reduction in risk, although it didn’t reach statistical significance (HR, 0.90; 95% CI, 0.80-1.02; P = .11). Furthermore, accelerated failure time models suggested that BMI did not significantly impact the link between steps and diabetes (proportion mediated, 17.7%;95% CI, –55.0 to 142.0; P = .09). Further analyses also found no significant interactions between BMI or other possible confounders.

“The steps per day–diabetes association was not modified by age, race/ethnicity, BMI, physical functioning, or family history of diabetes, which supports the generalizability of these findings to community-living older women,” the investigators wrote.

Increased stepping intensity also appeared to lower risk of diabetes. After adjusting for confounding variables, light stepping was not linked to reduced risk (HR, 0.97; 95% CI, 0.73-1.29; P = .83), whereas moderate to vigorous stepping reduced risk by 14% per 2,000 steps (HR, 0.86; 95% CI, 0.74-1.00; P = .04).

“This study provides evidence supporting an association between steps per day and lower incident diabetes,” the investigators concluded. “While further work is needed to identify whether there is a minimum number of steps per day that results in a clinically significant reduction of diabetes and to evaluate the role that step intensity plays in diabetes etiology for older adults, findings from this study suggest that moderate-vigorous–intensity steps may be more important than lower-intensity steps with respect to incident diabetes. Steps per day–based interventions are needed to advance diabetes prevention science in older adults.”

The study was supported by the National Institute on Aging, the National Institute of Diabetes and Digestive and Kidney Diseases, the Tobacco-Related Disease Research Program, and others. The investigators had no potential conflicts of interest.

More steps per day, particularly at a higher intensity, may reduce the risk of type 2 diabetes in older women, based on a prospective cohort study.

The link between daily stepping and diabetes was not significantly modified by body mass index (BMI) or other common diabetes risk factors, suggesting that the relationship is highly generalizable, lead author Alexis C. Garduno, MPH, a PhD student at the University of California, San Diego, and colleagues reported.

“Physical activity is a key modifiable behavior for diabetes prevention and management,” the investigators wrote in Diabetes Care. “Many prevention studies have demonstrated that regular physical activity, along with improved diet, reduces the risk of diabetes in adults. ... To the best of our knowledge, there are few studies examining the association between objectively measured steps per day and incident diabetes in a community-based setting.”

To this end, the investigators analyzed data from 4,838 older, community-living women in the Objective Physical Activity and Cardiovascular Health Study. Upon enrollment, women were without physician-diagnosed diabetes and had a mean age of 78.9 years. For 1 week, participants wore ActiGraph GT3X+ accelerometers to measure steps per day, as well as step intensity, graded as light or moderate to vigorous.

The relationship between daily activity and diabetes was analyzed using three multivariate models: The first included race/ethnicity and age; the second also included family history of diabetes, education, physical functioning, self-rated health, smoking status, and alcohol consumption; and the third added BMI, “a potential mediator in the causal pathway between steps per day and diabetes,” the investigators wrote.

Participants took an average of 3,729 steps per day, divided roughly evenly between light and moderate to vigorous intensity.

After a median follow-up of 5.7 years, 8.1% of women developed diabetes. The least-adjusted model showed a 14% reduction in diabetes risk per 2,000 steps (hazard ratio, 0.86; 95% confidence interval, 0.80-0.92; P = .007), whereas the second model, adjusting for more confounding variables, showed a 12% reduction in diabetes risk per 2,000 steps (HR, 0.88; 95% CI, 0.78-1.00; P = .045).

The final model, which added BMI, showed a 10% reduction in risk, although it didn’t reach statistical significance (HR, 0.90; 95% CI, 0.80-1.02; P = .11). Furthermore, accelerated failure time models suggested that BMI did not significantly impact the link between steps and diabetes (proportion mediated, 17.7%;95% CI, –55.0 to 142.0; P = .09). Further analyses also found no significant interactions between BMI or other possible confounders.

“The steps per day–diabetes association was not modified by age, race/ethnicity, BMI, physical functioning, or family history of diabetes, which supports the generalizability of these findings to community-living older women,” the investigators wrote.

Increased stepping intensity also appeared to lower risk of diabetes. After adjusting for confounding variables, light stepping was not linked to reduced risk (HR, 0.97; 95% CI, 0.73-1.29; P = .83), whereas moderate to vigorous stepping reduced risk by 14% per 2,000 steps (HR, 0.86; 95% CI, 0.74-1.00; P = .04).

“This study provides evidence supporting an association between steps per day and lower incident diabetes,” the investigators concluded. “While further work is needed to identify whether there is a minimum number of steps per day that results in a clinically significant reduction of diabetes and to evaluate the role that step intensity plays in diabetes etiology for older adults, findings from this study suggest that moderate-vigorous–intensity steps may be more important than lower-intensity steps with respect to incident diabetes. Steps per day–based interventions are needed to advance diabetes prevention science in older adults.”

The study was supported by the National Institute on Aging, the National Institute of Diabetes and Digestive and Kidney Diseases, the Tobacco-Related Disease Research Program, and others. The investigators had no potential conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM DIABETES CARE

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Dairy intake may increase risk of Parkinson’s disease in men

Article Type
Changed
Thu, 12/15/2022 - 15:39

Dairy intake may increase risk of Parkinson’s disease in men, according to investigators. Men of European ancestry with a genetic marker predicting dairy consumption had significantly greater risk of Parkinson’s disease than individuals without the marker, suggesting a causal relationship between dairy intake and Parkinson’s disease, lead author Cloé Domenighetti, MSc, a PhD student at UVSQ, Université Paris Sud, and colleagues reported.

“Previous studies highlighted dairy intake as a risk factor of Parkinson’s disease,” the investigators wrote in Movement Disorders. “A meta-analysis of prospective studies reported a 40% increased Parkinson’s disease risk in participants with the highest intake. It is unclear whether the association is causal or explained by confounding or reverse causation, given the long prodromal phase of Parkinson’s disease.”
 

A Mendelian randomization study

The investigators evaluated this link by comparing 9,823 cases of Parkinson’s disease with 8,376 controls, all individuals of European ancestry from the Courage-Parkinson’s disease consortium, comprising 23 studies. Data were analyzed by two-sample Mendelian randomization, a technique that uses genotype to predict behavior, thereby replacing conventional methods of capturing behavior, such as questionnaires. In this case, the investigators screened all participants for rs4988235, a single-nucleotide polymorphism (SNP) upstream of the lactase gene that is well documented to predict dairy intake among individuals of European ancestry.

“Mendelian randomization uses genetic variants associated with exposures as instrumental variables to estimate causal relationships between exposures and outcomes,” the investigators wrote. “Mendelian randomization analyses are less likely to be biased by confounding or reverse causation than observational studies if a set of assumptions are met.”

The approach uncovered a significant association between rs4988235 and Parkinson’s disease, with a 70% increase in disease risk per one serving of dairy per day (odds ratio, 1.70; 95% confidence interval, 1.12-2.60; P = .013). Further analysis revealed that this finding was driven by men, who had a 2.5-fold increased risk of Parkinson’s disease per one serving per day (OR, 2.50; 95% CI, 1.37-4.56; P = .003) versus women, among whom there was no significant association (OR, 1.04; 95% CI, 0.56-1.92; P = .91). No significant associations were observed among individuals grouped by age or Parkinson’s disease duration.

“Our findings suggest that dairy intake increases Parkinson’s disease risk,” the investigators concluded. “Therefore, diets with limited milk intake (e.g., Mediterranean diet) may be beneficial with respect to Parkinson’s disease.”
 

Further evidence supporting a link between diet and Parkinson’s disease

According to Silke Appel-Cresswell, MD, Marg Meikle Professor for Parkinson’s Research at the University of British Columbia, Vancouver, the findings align with previous prospective cohort studies demonstrating an increased risk of Parkinson’s disease with greater consumption of dairy.

Dr. Silke Appel-Cresswell

“What the current study adds,” Dr. Appel-Cresswell said, “is a complementary approach to assess the association where the risk of reverse causation and of confounding are minimized. Like in some of the previous studies, the authors find sex differences with an increased risk for men but not women.”

Dr. Appel-Cresswell noted that an increasing body of evidence supports a link between diet and Parkinson’s disease, including a study of her own published last year, which showed later onset of Parkinson’s disease among individuals with a Mediterranean-style diet.

“We are accumulating evidence for a role of diet (or more broadly, the food exposome) for the risk to develop Parkinson’s disease,” Dr. Appel-Cresswell said, noting that “key pieces are still missing, including mechanisms underlying associations, clinical trials in individuals with established Parkinson’s disease and – eventually – preventive interventions. This research is urgently needed and analyses will need to take sex differences and a large range of potential other factors into account.”
 

 

 

A ‘modest’ contributing factor?

Vikas Kotagal, MD, associate professor of neurology at the University of Michigan, Ann Arbor, offered a perspective on the study methodology, and suggested that a causal link between dairy intake and Parkinson’s disease, if present, is likely minimal.

Dr. Vikas Kotagal

“Limitations to the study include the fact that participants weren’t actually asked or tested for how much dairy they truly consumed,” Dr. Kotagal said*. “Their dairy intake was estimated based on their genetic background – there are certainly many assumptions baked into this analytic approach which may or may not be true. It is also worth noting the fact that this causal association was seen in men and not women, suggesting that even if dairy intake was truly causal, it is likely to be a modest contributing factor and not a significant cause of Parkinson’s disease in the broader population in general.”

Still, Dr. Kotagal agreed with Dr. Appel-Cresswell that underlying mechanisms need further investigation.

“The biggest takeaway here is to heighten the urgency for researchers and funders to explore whether factors that might cluster with dairy intake – including pesticide exposure in milk or even the make-up of bacterial populations in different peoples’ intestines – might deserve closer scrutiny as a missing link connecting dairy consumption to increased Parkinson’s disease risk,” Dr. Kotagal said.
 

Dietary advice

Considering all available evidence, Dr. Appel-Cresswell offered some dietary advice with benefits that may extend beyond prevention of Parkinson’s disease.

“From a clinical point of view, I suggest to limit dairy intake to a moderate amount,” she said. “Mediterranean diets so far have the best supporting evidence for a lower Parkinson’s disease risk, although data is lacking for benefits in established Parkinson’s disease. Given the low risk of the Mediterranean diet and the established benefits for a host of other medical conditions, this is generally a safe and delicious recommendation whether one is living with Parkinson’s or not.”

The study was supported by the European Union Joint Program for Neurodegenerative Disease Research, the National Centre of Excellence in Research on Parkinson’s Disease, the National Institutes of Health, and others. The investigators disclosed additional relationships with Astellas Pharma, Sanofi, Pfizer, and others. Dr. Kotagal and Dr. Appel-Cresswell reported no relevant conflicts of interest.

*Correction, 2/10/22: An earlier version of this article misstated Dr. Kotagal's name in certain instances, including a photo caption.

Publications
Topics
Sections

Dairy intake may increase risk of Parkinson’s disease in men, according to investigators. Men of European ancestry with a genetic marker predicting dairy consumption had significantly greater risk of Parkinson’s disease than individuals without the marker, suggesting a causal relationship between dairy intake and Parkinson’s disease, lead author Cloé Domenighetti, MSc, a PhD student at UVSQ, Université Paris Sud, and colleagues reported.

“Previous studies highlighted dairy intake as a risk factor of Parkinson’s disease,” the investigators wrote in Movement Disorders. “A meta-analysis of prospective studies reported a 40% increased Parkinson’s disease risk in participants with the highest intake. It is unclear whether the association is causal or explained by confounding or reverse causation, given the long prodromal phase of Parkinson’s disease.”
 

A Mendelian randomization study

The investigators evaluated this link by comparing 9,823 cases of Parkinson’s disease with 8,376 controls, all individuals of European ancestry from the Courage-Parkinson’s disease consortium, comprising 23 studies. Data were analyzed by two-sample Mendelian randomization, a technique that uses genotype to predict behavior, thereby replacing conventional methods of capturing behavior, such as questionnaires. In this case, the investigators screened all participants for rs4988235, a single-nucleotide polymorphism (SNP) upstream of the lactase gene that is well documented to predict dairy intake among individuals of European ancestry.

“Mendelian randomization uses genetic variants associated with exposures as instrumental variables to estimate causal relationships between exposures and outcomes,” the investigators wrote. “Mendelian randomization analyses are less likely to be biased by confounding or reverse causation than observational studies if a set of assumptions are met.”

The approach uncovered a significant association between rs4988235 and Parkinson’s disease, with a 70% increase in disease risk per one serving of dairy per day (odds ratio, 1.70; 95% confidence interval, 1.12-2.60; P = .013). Further analysis revealed that this finding was driven by men, who had a 2.5-fold increased risk of Parkinson’s disease per one serving per day (OR, 2.50; 95% CI, 1.37-4.56; P = .003) versus women, among whom there was no significant association (OR, 1.04; 95% CI, 0.56-1.92; P = .91). No significant associations were observed among individuals grouped by age or Parkinson’s disease duration.

“Our findings suggest that dairy intake increases Parkinson’s disease risk,” the investigators concluded. “Therefore, diets with limited milk intake (e.g., Mediterranean diet) may be beneficial with respect to Parkinson’s disease.”
 

Further evidence supporting a link between diet and Parkinson’s disease

According to Silke Appel-Cresswell, MD, Marg Meikle Professor for Parkinson’s Research at the University of British Columbia, Vancouver, the findings align with previous prospective cohort studies demonstrating an increased risk of Parkinson’s disease with greater consumption of dairy.

Dr. Silke Appel-Cresswell

“What the current study adds,” Dr. Appel-Cresswell said, “is a complementary approach to assess the association where the risk of reverse causation and of confounding are minimized. Like in some of the previous studies, the authors find sex differences with an increased risk for men but not women.”

Dr. Appel-Cresswell noted that an increasing body of evidence supports a link between diet and Parkinson’s disease, including a study of her own published last year, which showed later onset of Parkinson’s disease among individuals with a Mediterranean-style diet.

“We are accumulating evidence for a role of diet (or more broadly, the food exposome) for the risk to develop Parkinson’s disease,” Dr. Appel-Cresswell said, noting that “key pieces are still missing, including mechanisms underlying associations, clinical trials in individuals with established Parkinson’s disease and – eventually – preventive interventions. This research is urgently needed and analyses will need to take sex differences and a large range of potential other factors into account.”
 

 

 

A ‘modest’ contributing factor?

Vikas Kotagal, MD, associate professor of neurology at the University of Michigan, Ann Arbor, offered a perspective on the study methodology, and suggested that a causal link between dairy intake and Parkinson’s disease, if present, is likely minimal.

Dr. Vikas Kotagal

“Limitations to the study include the fact that participants weren’t actually asked or tested for how much dairy they truly consumed,” Dr. Kotagal said*. “Their dairy intake was estimated based on their genetic background – there are certainly many assumptions baked into this analytic approach which may or may not be true. It is also worth noting the fact that this causal association was seen in men and not women, suggesting that even if dairy intake was truly causal, it is likely to be a modest contributing factor and not a significant cause of Parkinson’s disease in the broader population in general.”

Still, Dr. Kotagal agreed with Dr. Appel-Cresswell that underlying mechanisms need further investigation.

“The biggest takeaway here is to heighten the urgency for researchers and funders to explore whether factors that might cluster with dairy intake – including pesticide exposure in milk or even the make-up of bacterial populations in different peoples’ intestines – might deserve closer scrutiny as a missing link connecting dairy consumption to increased Parkinson’s disease risk,” Dr. Kotagal said.
 

Dietary advice

Considering all available evidence, Dr. Appel-Cresswell offered some dietary advice with benefits that may extend beyond prevention of Parkinson’s disease.

“From a clinical point of view, I suggest to limit dairy intake to a moderate amount,” she said. “Mediterranean diets so far have the best supporting evidence for a lower Parkinson’s disease risk, although data is lacking for benefits in established Parkinson’s disease. Given the low risk of the Mediterranean diet and the established benefits for a host of other medical conditions, this is generally a safe and delicious recommendation whether one is living with Parkinson’s or not.”

The study was supported by the European Union Joint Program for Neurodegenerative Disease Research, the National Centre of Excellence in Research on Parkinson’s Disease, the National Institutes of Health, and others. The investigators disclosed additional relationships with Astellas Pharma, Sanofi, Pfizer, and others. Dr. Kotagal and Dr. Appel-Cresswell reported no relevant conflicts of interest.

*Correction, 2/10/22: An earlier version of this article misstated Dr. Kotagal's name in certain instances, including a photo caption.

Dairy intake may increase risk of Parkinson’s disease in men, according to investigators. Men of European ancestry with a genetic marker predicting dairy consumption had significantly greater risk of Parkinson’s disease than individuals without the marker, suggesting a causal relationship between dairy intake and Parkinson’s disease, lead author Cloé Domenighetti, MSc, a PhD student at UVSQ, Université Paris Sud, and colleagues reported.

“Previous studies highlighted dairy intake as a risk factor of Parkinson’s disease,” the investigators wrote in Movement Disorders. “A meta-analysis of prospective studies reported a 40% increased Parkinson’s disease risk in participants with the highest intake. It is unclear whether the association is causal or explained by confounding or reverse causation, given the long prodromal phase of Parkinson’s disease.”
 

A Mendelian randomization study

The investigators evaluated this link by comparing 9,823 cases of Parkinson’s disease with 8,376 controls, all individuals of European ancestry from the Courage-Parkinson’s disease consortium, comprising 23 studies. Data were analyzed by two-sample Mendelian randomization, a technique that uses genotype to predict behavior, thereby replacing conventional methods of capturing behavior, such as questionnaires. In this case, the investigators screened all participants for rs4988235, a single-nucleotide polymorphism (SNP) upstream of the lactase gene that is well documented to predict dairy intake among individuals of European ancestry.

“Mendelian randomization uses genetic variants associated with exposures as instrumental variables to estimate causal relationships between exposures and outcomes,” the investigators wrote. “Mendelian randomization analyses are less likely to be biased by confounding or reverse causation than observational studies if a set of assumptions are met.”

The approach uncovered a significant association between rs4988235 and Parkinson’s disease, with a 70% increase in disease risk per one serving of dairy per day (odds ratio, 1.70; 95% confidence interval, 1.12-2.60; P = .013). Further analysis revealed that this finding was driven by men, who had a 2.5-fold increased risk of Parkinson’s disease per one serving per day (OR, 2.50; 95% CI, 1.37-4.56; P = .003) versus women, among whom there was no significant association (OR, 1.04; 95% CI, 0.56-1.92; P = .91). No significant associations were observed among individuals grouped by age or Parkinson’s disease duration.

“Our findings suggest that dairy intake increases Parkinson’s disease risk,” the investigators concluded. “Therefore, diets with limited milk intake (e.g., Mediterranean diet) may be beneficial with respect to Parkinson’s disease.”
 

Further evidence supporting a link between diet and Parkinson’s disease

According to Silke Appel-Cresswell, MD, Marg Meikle Professor for Parkinson’s Research at the University of British Columbia, Vancouver, the findings align with previous prospective cohort studies demonstrating an increased risk of Parkinson’s disease with greater consumption of dairy.

Dr. Silke Appel-Cresswell

“What the current study adds,” Dr. Appel-Cresswell said, “is a complementary approach to assess the association where the risk of reverse causation and of confounding are minimized. Like in some of the previous studies, the authors find sex differences with an increased risk for men but not women.”

Dr. Appel-Cresswell noted that an increasing body of evidence supports a link between diet and Parkinson’s disease, including a study of her own published last year, which showed later onset of Parkinson’s disease among individuals with a Mediterranean-style diet.

“We are accumulating evidence for a role of diet (or more broadly, the food exposome) for the risk to develop Parkinson’s disease,” Dr. Appel-Cresswell said, noting that “key pieces are still missing, including mechanisms underlying associations, clinical trials in individuals with established Parkinson’s disease and – eventually – preventive interventions. This research is urgently needed and analyses will need to take sex differences and a large range of potential other factors into account.”
 

 

 

A ‘modest’ contributing factor?

Vikas Kotagal, MD, associate professor of neurology at the University of Michigan, Ann Arbor, offered a perspective on the study methodology, and suggested that a causal link between dairy intake and Parkinson’s disease, if present, is likely minimal.

Dr. Vikas Kotagal

“Limitations to the study include the fact that participants weren’t actually asked or tested for how much dairy they truly consumed,” Dr. Kotagal said*. “Their dairy intake was estimated based on their genetic background – there are certainly many assumptions baked into this analytic approach which may or may not be true. It is also worth noting the fact that this causal association was seen in men and not women, suggesting that even if dairy intake was truly causal, it is likely to be a modest contributing factor and not a significant cause of Parkinson’s disease in the broader population in general.”

Still, Dr. Kotagal agreed with Dr. Appel-Cresswell that underlying mechanisms need further investigation.

“The biggest takeaway here is to heighten the urgency for researchers and funders to explore whether factors that might cluster with dairy intake – including pesticide exposure in milk or even the make-up of bacterial populations in different peoples’ intestines – might deserve closer scrutiny as a missing link connecting dairy consumption to increased Parkinson’s disease risk,” Dr. Kotagal said.
 

Dietary advice

Considering all available evidence, Dr. Appel-Cresswell offered some dietary advice with benefits that may extend beyond prevention of Parkinson’s disease.

“From a clinical point of view, I suggest to limit dairy intake to a moderate amount,” she said. “Mediterranean diets so far have the best supporting evidence for a lower Parkinson’s disease risk, although data is lacking for benefits in established Parkinson’s disease. Given the low risk of the Mediterranean diet and the established benefits for a host of other medical conditions, this is generally a safe and delicious recommendation whether one is living with Parkinson’s or not.”

The study was supported by the European Union Joint Program for Neurodegenerative Disease Research, the National Centre of Excellence in Research on Parkinson’s Disease, the National Institutes of Health, and others. The investigators disclosed additional relationships with Astellas Pharma, Sanofi, Pfizer, and others. Dr. Kotagal and Dr. Appel-Cresswell reported no relevant conflicts of interest.

*Correction, 2/10/22: An earlier version of this article misstated Dr. Kotagal's name in certain instances, including a photo caption.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM MOVEMENT DISORDERS

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

High GI spending reveals research, public health need

Article Type
Changed
Wed, 01/05/2022 - 11:04

 

GI, liver, and pancreatic diseases cost the U.S. health care system about $120B per year and account for approximately 250,000 annual deaths, according to a “conservative” estimate from a recent analysis.

These figures emphasize the need for more research funding in the area, along with additional clinical and public health initiatives, reported lead author Anne F. Peery, MD, of the University of North Carolina School of Medicine, Chapel Hill, and colleagues.

“Reports detailing the burden of GI diseases are necessary for clinical research, decision making, and priority setting,” the investigators wrote in Gastroenterology. “Our aim was to describe health care use, expenditures, and research funding across GI, liver, and pancreatic diseases in the United States.”

Dr. Peery and colleagues analyzed data from 14 sources, including the National Institutes of Health; the Centers for Disease Control and Prevention; the National Ambulatory Medical Care Survey; and others. GI-specific outcomes included mortality, readmissions, hospitalizations, office-based visits, and emergency department visits. The investigators also characterized trends in cancers, organ transplants, and GI endoscopy, as well as GI-specific health care costs and NIH research funding. Annual findings were presented for various periods.

Total GI health care spending was $119.6 billion in 2018, down from $135.9 billion in 2015. The top five most costly conditions were biliary tract diseases ($16.9 billion), esophageal disorders ($12.1 billion), abdominal pain ($9.5 billion), abdominal hernias ($9.0 billion), and diverticular disease ($9.0 billion). The investigators noted that medication costs were particularly high for two categories: inflammatory bowel diseases and esophageal disorders, which had prescription drug costs relative to total expenditures of 71% and 53%, respectively.

“This conservative estimate [of $119.6 billion] did not include most GI cancers and likely underestimated the costs associated with some GI conditions,” the investigators noted. “For example, the Medical Expenditure Panel Survey estimate associated with GI bleeding was $300 million. In comparison, the aggregate cost of GI bleeding was more realistically $3.7 billion, as estimated using inpatient data from the National Inpatient Sample.”

In 2016, the most common GI-related diagnosis in the U.S. was abdominal pain (15.7 million annual visits), followed by nausea and vomiting (5.0 million visits), gastroesophageal reflux disorder and reflux esophagitis (4.7 million visits), constipation (3.1 million visits), and abdominal wall/inguinal hernia (2.8 million visits).

The top three most common GI-related hospital admissions in 2018 were GI bleeding (1.3 million admissions), followed by cholelithiasis and cholecystitis (741,060 admissions), then pancreatitis (685,880 admissions). GI bleeding was also the leading cause of 30-day readmission in 2018 (84,533 readmissions).

“We found substantial numbers of GI conditions and symptoms listed in secondary positions on the discharge record,” the investigators wrote. “For example, liver disease accounted for 280,645 discharges with a primary diagnosis; however, there were 13-fold as many discharges (3.6 million in 2018) with liver disease as a secondary diagnosis. Including all diagnoses captures a burden of GI disease not previously reported.”

In 2018 and 2019, GI diseases and cancers caused 255,407 annual deaths. The most common noncancer deaths were caused by alcohol-associated liver disease (24,110 deaths), hepatic fibrosis/cirrhosis (20,184 deaths), and GI bleeding (9,548 deaths). Among GI-cancer related deaths, colorectal cancer (CRC) caused the most mortalities (52,163 deaths), followed by pancreatic cancer (44,914 deaths), and hepatic/biliary cancer (44,914 deaths). The investigators noted that CRC was disproportionately common among non-Hispanic Black individuals, whereas gastric cancer was relatively high among Hispanic individuals.

“GI cancers account for a large number of diagnoses and deaths annually, with persistent disparities in incidence and mortality rates by race/ethnicity,” the investigators wrote. “Racial, ethnic, and regional disparities in access to most GI endoscopy procedures exist, which suggests an unmet need for GI procedures across the United States.”

A total of 22.2 million endoscopies were performed in 2019, most commonly colonoscopy (13.8 million procedures), followed by upper endoscopy (7.5 million procedures), and flexible sigmoidoscopy (379,883 procedures).

In 2020, the NIH spent $3.1 billion, or approximately 7.5% of its budget, on GI disease research. Digestive diseases captured the bulk of this spending, with $2.3 billion. In the same year, the NIH spent 10.5% of its cancer research budget on GI cancers, with the greatest proportion ($325 million) awarded to CRC research.

“Carefully examining the data in this report can help generate areas for future investigation, prioritize research funding, identify areas of unmet need or disparities, and provide an important overview of the impact of digestive and liver conditions,” the investigators concluded. “We hope that others will use this report as motivation to take a deeper dive into individual diseases. There is much to learn from carefully studying existing data sources.”

The study was supported by the National Center for Advancing Translational Sciences, National Institutes of Health. The investigators disclosed no conflicts of interest.

Publications
Topics
Sections

 

GI, liver, and pancreatic diseases cost the U.S. health care system about $120B per year and account for approximately 250,000 annual deaths, according to a “conservative” estimate from a recent analysis.

These figures emphasize the need for more research funding in the area, along with additional clinical and public health initiatives, reported lead author Anne F. Peery, MD, of the University of North Carolina School of Medicine, Chapel Hill, and colleagues.

“Reports detailing the burden of GI diseases are necessary for clinical research, decision making, and priority setting,” the investigators wrote in Gastroenterology. “Our aim was to describe health care use, expenditures, and research funding across GI, liver, and pancreatic diseases in the United States.”

Dr. Peery and colleagues analyzed data from 14 sources, including the National Institutes of Health; the Centers for Disease Control and Prevention; the National Ambulatory Medical Care Survey; and others. GI-specific outcomes included mortality, readmissions, hospitalizations, office-based visits, and emergency department visits. The investigators also characterized trends in cancers, organ transplants, and GI endoscopy, as well as GI-specific health care costs and NIH research funding. Annual findings were presented for various periods.

Total GI health care spending was $119.6 billion in 2018, down from $135.9 billion in 2015. The top five most costly conditions were biliary tract diseases ($16.9 billion), esophageal disorders ($12.1 billion), abdominal pain ($9.5 billion), abdominal hernias ($9.0 billion), and diverticular disease ($9.0 billion). The investigators noted that medication costs were particularly high for two categories: inflammatory bowel diseases and esophageal disorders, which had prescription drug costs relative to total expenditures of 71% and 53%, respectively.

“This conservative estimate [of $119.6 billion] did not include most GI cancers and likely underestimated the costs associated with some GI conditions,” the investigators noted. “For example, the Medical Expenditure Panel Survey estimate associated with GI bleeding was $300 million. In comparison, the aggregate cost of GI bleeding was more realistically $3.7 billion, as estimated using inpatient data from the National Inpatient Sample.”

In 2016, the most common GI-related diagnosis in the U.S. was abdominal pain (15.7 million annual visits), followed by nausea and vomiting (5.0 million visits), gastroesophageal reflux disorder and reflux esophagitis (4.7 million visits), constipation (3.1 million visits), and abdominal wall/inguinal hernia (2.8 million visits).

The top three most common GI-related hospital admissions in 2018 were GI bleeding (1.3 million admissions), followed by cholelithiasis and cholecystitis (741,060 admissions), then pancreatitis (685,880 admissions). GI bleeding was also the leading cause of 30-day readmission in 2018 (84,533 readmissions).

“We found substantial numbers of GI conditions and symptoms listed in secondary positions on the discharge record,” the investigators wrote. “For example, liver disease accounted for 280,645 discharges with a primary diagnosis; however, there were 13-fold as many discharges (3.6 million in 2018) with liver disease as a secondary diagnosis. Including all diagnoses captures a burden of GI disease not previously reported.”

In 2018 and 2019, GI diseases and cancers caused 255,407 annual deaths. The most common noncancer deaths were caused by alcohol-associated liver disease (24,110 deaths), hepatic fibrosis/cirrhosis (20,184 deaths), and GI bleeding (9,548 deaths). Among GI-cancer related deaths, colorectal cancer (CRC) caused the most mortalities (52,163 deaths), followed by pancreatic cancer (44,914 deaths), and hepatic/biliary cancer (44,914 deaths). The investigators noted that CRC was disproportionately common among non-Hispanic Black individuals, whereas gastric cancer was relatively high among Hispanic individuals.

“GI cancers account for a large number of diagnoses and deaths annually, with persistent disparities in incidence and mortality rates by race/ethnicity,” the investigators wrote. “Racial, ethnic, and regional disparities in access to most GI endoscopy procedures exist, which suggests an unmet need for GI procedures across the United States.”

A total of 22.2 million endoscopies were performed in 2019, most commonly colonoscopy (13.8 million procedures), followed by upper endoscopy (7.5 million procedures), and flexible sigmoidoscopy (379,883 procedures).

In 2020, the NIH spent $3.1 billion, or approximately 7.5% of its budget, on GI disease research. Digestive diseases captured the bulk of this spending, with $2.3 billion. In the same year, the NIH spent 10.5% of its cancer research budget on GI cancers, with the greatest proportion ($325 million) awarded to CRC research.

“Carefully examining the data in this report can help generate areas for future investigation, prioritize research funding, identify areas of unmet need or disparities, and provide an important overview of the impact of digestive and liver conditions,” the investigators concluded. “We hope that others will use this report as motivation to take a deeper dive into individual diseases. There is much to learn from carefully studying existing data sources.”

The study was supported by the National Center for Advancing Translational Sciences, National Institutes of Health. The investigators disclosed no conflicts of interest.

 

GI, liver, and pancreatic diseases cost the U.S. health care system about $120B per year and account for approximately 250,000 annual deaths, according to a “conservative” estimate from a recent analysis.

These figures emphasize the need for more research funding in the area, along with additional clinical and public health initiatives, reported lead author Anne F. Peery, MD, of the University of North Carolina School of Medicine, Chapel Hill, and colleagues.

“Reports detailing the burden of GI diseases are necessary for clinical research, decision making, and priority setting,” the investigators wrote in Gastroenterology. “Our aim was to describe health care use, expenditures, and research funding across GI, liver, and pancreatic diseases in the United States.”

Dr. Peery and colleagues analyzed data from 14 sources, including the National Institutes of Health; the Centers for Disease Control and Prevention; the National Ambulatory Medical Care Survey; and others. GI-specific outcomes included mortality, readmissions, hospitalizations, office-based visits, and emergency department visits. The investigators also characterized trends in cancers, organ transplants, and GI endoscopy, as well as GI-specific health care costs and NIH research funding. Annual findings were presented for various periods.

Total GI health care spending was $119.6 billion in 2018, down from $135.9 billion in 2015. The top five most costly conditions were biliary tract diseases ($16.9 billion), esophageal disorders ($12.1 billion), abdominal pain ($9.5 billion), abdominal hernias ($9.0 billion), and diverticular disease ($9.0 billion). The investigators noted that medication costs were particularly high for two categories: inflammatory bowel diseases and esophageal disorders, which had prescription drug costs relative to total expenditures of 71% and 53%, respectively.

“This conservative estimate [of $119.6 billion] did not include most GI cancers and likely underestimated the costs associated with some GI conditions,” the investigators noted. “For example, the Medical Expenditure Panel Survey estimate associated with GI bleeding was $300 million. In comparison, the aggregate cost of GI bleeding was more realistically $3.7 billion, as estimated using inpatient data from the National Inpatient Sample.”

In 2016, the most common GI-related diagnosis in the U.S. was abdominal pain (15.7 million annual visits), followed by nausea and vomiting (5.0 million visits), gastroesophageal reflux disorder and reflux esophagitis (4.7 million visits), constipation (3.1 million visits), and abdominal wall/inguinal hernia (2.8 million visits).

The top three most common GI-related hospital admissions in 2018 were GI bleeding (1.3 million admissions), followed by cholelithiasis and cholecystitis (741,060 admissions), then pancreatitis (685,880 admissions). GI bleeding was also the leading cause of 30-day readmission in 2018 (84,533 readmissions).

“We found substantial numbers of GI conditions and symptoms listed in secondary positions on the discharge record,” the investigators wrote. “For example, liver disease accounted for 280,645 discharges with a primary diagnosis; however, there were 13-fold as many discharges (3.6 million in 2018) with liver disease as a secondary diagnosis. Including all diagnoses captures a burden of GI disease not previously reported.”

In 2018 and 2019, GI diseases and cancers caused 255,407 annual deaths. The most common noncancer deaths were caused by alcohol-associated liver disease (24,110 deaths), hepatic fibrosis/cirrhosis (20,184 deaths), and GI bleeding (9,548 deaths). Among GI-cancer related deaths, colorectal cancer (CRC) caused the most mortalities (52,163 deaths), followed by pancreatic cancer (44,914 deaths), and hepatic/biliary cancer (44,914 deaths). The investigators noted that CRC was disproportionately common among non-Hispanic Black individuals, whereas gastric cancer was relatively high among Hispanic individuals.

“GI cancers account for a large number of diagnoses and deaths annually, with persistent disparities in incidence and mortality rates by race/ethnicity,” the investigators wrote. “Racial, ethnic, and regional disparities in access to most GI endoscopy procedures exist, which suggests an unmet need for GI procedures across the United States.”

A total of 22.2 million endoscopies were performed in 2019, most commonly colonoscopy (13.8 million procedures), followed by upper endoscopy (7.5 million procedures), and flexible sigmoidoscopy (379,883 procedures).

In 2020, the NIH spent $3.1 billion, or approximately 7.5% of its budget, on GI disease research. Digestive diseases captured the bulk of this spending, with $2.3 billion. In the same year, the NIH spent 10.5% of its cancer research budget on GI cancers, with the greatest proportion ($325 million) awarded to CRC research.

“Carefully examining the data in this report can help generate areas for future investigation, prioritize research funding, identify areas of unmet need or disparities, and provide an important overview of the impact of digestive and liver conditions,” the investigators concluded. “We hope that others will use this report as motivation to take a deeper dive into individual diseases. There is much to learn from carefully studying existing data sources.”

The study was supported by the National Center for Advancing Translational Sciences, National Institutes of Health. The investigators disclosed no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

US Multi-Society Task Force lowers recommended CRC screening age

Article Type
Changed
Mon, 01/03/2022 - 10:35

The U.S. Multi-Society Task Force on Colorectal Cancer (CRC) has lowered the recommended age to start CRC screening from 50 to 45 years of age for all average-risk individuals.

Although no studies have directly demonstrated the result of lowering the age of screening, lead author Swati G. Patel, MD, of University of Colorado Anschutz Medical Center, Aurora, and colleagues suggested that the increasing incidence of advanced CRC among younger individuals, coupled with the net benefit of screening, warrant a lower age threshold.

“Recent data ... show that CRC incidence rates in individuals ages 50 to 64 have increased by 1% annually between 2011 and 2016,” the authors wrote in Gastroenterology. “Similarly, CRC incidence and mortality rates in persons under age 50, termed early-age onset CRC (EAO-CRC), are also increasing.”

The task force of nine experts, representing the American Gastroenterological Association, the American College of Gastroenterology, and the American Society for Gastrointestinal Endoscopy, conducted a literature review and generated recommendations using the Grading of Recommendations Assessment, Development, and Evaluation (GRADE) criteria. In addition to recommending a lower age for initial screening, Dr. Patel and colleagues provided guidance for cessation of screening among older individuals.
 

Guidance for screening initiation

According to the authors, the present risk of CRC among younger individuals mirrors the historical risk for older individuals before screening was prevalent.

“The current CRC incidence rates in individuals ages 45 to 49 are similar to the incidence rates observed in 50-year-olds in 1992, before widespread CRC screening was performed,” they wrote.

Elevated rates among younger people have been disproportionately driven by rectal cancer, according to the authors. From 2006 to 2015, incidence of rectal cancer among Americans under 50 increased 1.7% per year, compared with 0.7% per year for colon cancer, based on data from the North American Association of Central Cancer Registries.

Associated mortality rates also increased, the authors noted. From 1999-2019, mortality from colon cancer among people 45-49 years increased from 6.4 to 6.6 deaths per 100,000 individuals, while deaths from rectal cancer increased from 1.3 to 1.7 per 100,000, according to the CDC. Concurrently, CRC-associated mortality rates among older individuals generally declined.

While these findings suggest a growing disease burden among the under-50-year age group, controlled data demonstrating the effects of earlier screening are lacking, Dr. Patel and colleagues noted. Still, they predicted that expanded screening would generate a net benefit.

“Although there are no CRC screening safety data for average-risk individuals [younger than] 50, there are ample data that colonoscopy for other indications (screening based on family history, symptom evaluation, etc.) is safer when comparing younger versus older individuals,” they wrote.

Supporting this claim, the authors cited three independently generated microsimulation models from the Agency for Healthcare Research and Quality that “showed a favorable balance of life-years gained compared with adverse events,” given 100% compliance.
 

Guidance for screening cessation

Like the situation with younger individuals, minimal data are available to determine the best time for screening cessation, according to the task force.

“There are no randomized or observational studies after 2017 that enrolled individuals over age 75 to inform the appropriate time to stop CRC screening,” the authors wrote. “In our search of 37 relevant articles, only one presented primary data for when to stop screening.”

This one available study showed that some individuals older than 74 do in fact gain benefit from screening,

“For example,” Dr. Patel and colleagues wrote, “women without a history of screening and no comorbidities benefitted from annual fecal immunochemical test (FIT) screening until age 90, whereas unscreened men with or without comorbidities benefited from annual FIT screening until age 88. Conversely, screening was not beneficial beyond age 66 in men or women with severe comorbidities.”

The task force therefore recommended personalized screening for individuals 76-85 years of age “based on the balance of benefits and harms and individual patient clinical factors and preferences.”

Screening for individuals 86 years and older, according to the task force, is unnecessary.

The authors disclosed relationships with Olympus America, Bayer Pharmaceuticals, Janssen Pharmaceuticals, and others.

This article was updated on Jan. 3, 2022.

Publications
Topics
Sections

The U.S. Multi-Society Task Force on Colorectal Cancer (CRC) has lowered the recommended age to start CRC screening from 50 to 45 years of age for all average-risk individuals.

Although no studies have directly demonstrated the result of lowering the age of screening, lead author Swati G. Patel, MD, of University of Colorado Anschutz Medical Center, Aurora, and colleagues suggested that the increasing incidence of advanced CRC among younger individuals, coupled with the net benefit of screening, warrant a lower age threshold.

“Recent data ... show that CRC incidence rates in individuals ages 50 to 64 have increased by 1% annually between 2011 and 2016,” the authors wrote in Gastroenterology. “Similarly, CRC incidence and mortality rates in persons under age 50, termed early-age onset CRC (EAO-CRC), are also increasing.”

The task force of nine experts, representing the American Gastroenterological Association, the American College of Gastroenterology, and the American Society for Gastrointestinal Endoscopy, conducted a literature review and generated recommendations using the Grading of Recommendations Assessment, Development, and Evaluation (GRADE) criteria. In addition to recommending a lower age for initial screening, Dr. Patel and colleagues provided guidance for cessation of screening among older individuals.
 

Guidance for screening initiation

According to the authors, the present risk of CRC among younger individuals mirrors the historical risk for older individuals before screening was prevalent.

“The current CRC incidence rates in individuals ages 45 to 49 are similar to the incidence rates observed in 50-year-olds in 1992, before widespread CRC screening was performed,” they wrote.

Elevated rates among younger people have been disproportionately driven by rectal cancer, according to the authors. From 2006 to 2015, incidence of rectal cancer among Americans under 50 increased 1.7% per year, compared with 0.7% per year for colon cancer, based on data from the North American Association of Central Cancer Registries.

Associated mortality rates also increased, the authors noted. From 1999-2019, mortality from colon cancer among people 45-49 years increased from 6.4 to 6.6 deaths per 100,000 individuals, while deaths from rectal cancer increased from 1.3 to 1.7 per 100,000, according to the CDC. Concurrently, CRC-associated mortality rates among older individuals generally declined.

While these findings suggest a growing disease burden among the under-50-year age group, controlled data demonstrating the effects of earlier screening are lacking, Dr. Patel and colleagues noted. Still, they predicted that expanded screening would generate a net benefit.

“Although there are no CRC screening safety data for average-risk individuals [younger than] 50, there are ample data that colonoscopy for other indications (screening based on family history, symptom evaluation, etc.) is safer when comparing younger versus older individuals,” they wrote.

Supporting this claim, the authors cited three independently generated microsimulation models from the Agency for Healthcare Research and Quality that “showed a favorable balance of life-years gained compared with adverse events,” given 100% compliance.
 

Guidance for screening cessation

Like the situation with younger individuals, minimal data are available to determine the best time for screening cessation, according to the task force.

“There are no randomized or observational studies after 2017 that enrolled individuals over age 75 to inform the appropriate time to stop CRC screening,” the authors wrote. “In our search of 37 relevant articles, only one presented primary data for when to stop screening.”

This one available study showed that some individuals older than 74 do in fact gain benefit from screening,

“For example,” Dr. Patel and colleagues wrote, “women without a history of screening and no comorbidities benefitted from annual fecal immunochemical test (FIT) screening until age 90, whereas unscreened men with or without comorbidities benefited from annual FIT screening until age 88. Conversely, screening was not beneficial beyond age 66 in men or women with severe comorbidities.”

The task force therefore recommended personalized screening for individuals 76-85 years of age “based on the balance of benefits and harms and individual patient clinical factors and preferences.”

Screening for individuals 86 years and older, according to the task force, is unnecessary.

The authors disclosed relationships with Olympus America, Bayer Pharmaceuticals, Janssen Pharmaceuticals, and others.

This article was updated on Jan. 3, 2022.

The U.S. Multi-Society Task Force on Colorectal Cancer (CRC) has lowered the recommended age to start CRC screening from 50 to 45 years of age for all average-risk individuals.

Although no studies have directly demonstrated the result of lowering the age of screening, lead author Swati G. Patel, MD, of University of Colorado Anschutz Medical Center, Aurora, and colleagues suggested that the increasing incidence of advanced CRC among younger individuals, coupled with the net benefit of screening, warrant a lower age threshold.

“Recent data ... show that CRC incidence rates in individuals ages 50 to 64 have increased by 1% annually between 2011 and 2016,” the authors wrote in Gastroenterology. “Similarly, CRC incidence and mortality rates in persons under age 50, termed early-age onset CRC (EAO-CRC), are also increasing.”

The task force of nine experts, representing the American Gastroenterological Association, the American College of Gastroenterology, and the American Society for Gastrointestinal Endoscopy, conducted a literature review and generated recommendations using the Grading of Recommendations Assessment, Development, and Evaluation (GRADE) criteria. In addition to recommending a lower age for initial screening, Dr. Patel and colleagues provided guidance for cessation of screening among older individuals.
 

Guidance for screening initiation

According to the authors, the present risk of CRC among younger individuals mirrors the historical risk for older individuals before screening was prevalent.

“The current CRC incidence rates in individuals ages 45 to 49 are similar to the incidence rates observed in 50-year-olds in 1992, before widespread CRC screening was performed,” they wrote.

Elevated rates among younger people have been disproportionately driven by rectal cancer, according to the authors. From 2006 to 2015, incidence of rectal cancer among Americans under 50 increased 1.7% per year, compared with 0.7% per year for colon cancer, based on data from the North American Association of Central Cancer Registries.

Associated mortality rates also increased, the authors noted. From 1999-2019, mortality from colon cancer among people 45-49 years increased from 6.4 to 6.6 deaths per 100,000 individuals, while deaths from rectal cancer increased from 1.3 to 1.7 per 100,000, according to the CDC. Concurrently, CRC-associated mortality rates among older individuals generally declined.

While these findings suggest a growing disease burden among the under-50-year age group, controlled data demonstrating the effects of earlier screening are lacking, Dr. Patel and colleagues noted. Still, they predicted that expanded screening would generate a net benefit.

“Although there are no CRC screening safety data for average-risk individuals [younger than] 50, there are ample data that colonoscopy for other indications (screening based on family history, symptom evaluation, etc.) is safer when comparing younger versus older individuals,” they wrote.

Supporting this claim, the authors cited three independently generated microsimulation models from the Agency for Healthcare Research and Quality that “showed a favorable balance of life-years gained compared with adverse events,” given 100% compliance.
 

Guidance for screening cessation

Like the situation with younger individuals, minimal data are available to determine the best time for screening cessation, according to the task force.

“There are no randomized or observational studies after 2017 that enrolled individuals over age 75 to inform the appropriate time to stop CRC screening,” the authors wrote. “In our search of 37 relevant articles, only one presented primary data for when to stop screening.”

This one available study showed that some individuals older than 74 do in fact gain benefit from screening,

“For example,” Dr. Patel and colleagues wrote, “women without a history of screening and no comorbidities benefitted from annual fecal immunochemical test (FIT) screening until age 90, whereas unscreened men with or without comorbidities benefited from annual FIT screening until age 88. Conversely, screening was not beneficial beyond age 66 in men or women with severe comorbidities.”

The task force therefore recommended personalized screening for individuals 76-85 years of age “based on the balance of benefits and harms and individual patient clinical factors and preferences.”

Screening for individuals 86 years and older, according to the task force, is unnecessary.

The authors disclosed relationships with Olympus America, Bayer Pharmaceuticals, Janssen Pharmaceuticals, and others.

This article was updated on Jan. 3, 2022.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article