Article Text

Download PDFPDF

A diagnostic model for chronic hypersensitivity pneumonitis
Free
  1. Kerri A Johannson1,2,
  2. Brett M Elicker3,
  3. Eric Vittinghoff4,
  4. Deborah Assayag5,
  5. Kaïssa de Boer1,
  6. Jeffrey A Golden1,
  7. Kirk D Jones6,
  8. Talmadge E King Jr1,
  9. Laura L Koth1,
  10. Joyce S Lee7,
  11. Brett Ley1,
  12. Paul J Wolters1,
  13. Harold R Collard1
  1. 1 Department of Medicine, University of California, San Francisco, San Francisco, California, USA
  2. 2 Department of Medicine, University of Calgary, Calgary, Alberta, Canada
  3. 3 Department of Radiology, University of California, San Francisco, San Francisco, California, USA
  4. 4 Department of Epidemiology and Biostatistics, University of California, San Francisco, San Francisco, California, USA
  5. 5 Montreal Jewish General Hospital, Montreal, Quebec, Canada
  6. 6 Department of Pathology, University of California, San Francisco, San Francisco, California, USA
  7. 7 University of Colorado Denver, Aurora, Colorado, USA
  1. Correspondence to Dr Kerri A Johannson, Department of Medicine, University of Calgary, 6th Floor, 4448 Front Street S.E., Calgary, Alberta, Canada T3M-1M4; kerri.johannson{at}ahs.ca

Abstract

The objective of this study was to develop a diagnostic model that allows for a highly specific diagnosis of chronic hypersensitivity pneumonitis using clinical and radiological variables alone. Chronic hypersensitivity pneumonitis and other interstitial lung disease cases were retrospectively identified from a longitudinal database. High-resolution CT scans were blindly scored for radiographic features (eg, ground-glass opacity, mosaic perfusion) as well as the radiologist's diagnostic impression. Candidate models were developed then evaluated using clinical and radiographic variables and assessed by the cross-validated C-statistic. Forty-four chronic hypersensitivity pneumonitis and eighty other interstitial lung disease cases were identified. Two models were selected based on their statistical performance, clinical applicability and face validity. Key model variables included age, down feather and/or bird exposure, radiographic presence of ground-glass opacity and mosaic perfusion and moderate or high confidence in the radiographic impression of chronic hypersensitivity pneumonitis. Models were internally validated with good performance, and cut-off values were established that resulted in high specificity for a diagnosis of chronic hypersensitivity pneumonitis.

  • Interstitial Fibrosis
  • Hypersensitivity pneumonitis
  • Clinical Epidemiology

Statistics from Altmetric.com

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.

Introduction

Chronic hypersensitivity pneumonitis (HP) is an interstitial lung disease (ILD) caused by long-standing exposure to an environmental antigen1 that can be challenging to distinguish from other forms of chronic ILD.2 ,3 Making an accurate diagnosis of chronic HP is essential as chronic HP requires a distinct management plan and informs prognosis.3 ,4 Previously proposed multidimensional diagnostic models for chronic HP have failed to gain widespread acceptance, perhaps due to their complexity and lack of specificity for chronic disease.5 ,6 The purpose of this study was to construct a clinically feasible, multidimensional diagnostic model for chronic HP that would allow for a highly specific diagnosis in the absence of surgical lung biopsy.

Methods

Study population

Patients with chronic HP and other ILDs seen between 2006 and 2012 were identified from a longitudinal cohort at the University of California, San Francisco. All patients were required to have a high-resolution CT (HRCT) scan of the chest performed within 1 year of their baseline clinic visit. Chronic HP cases were required to have a surgical lung biopsy and the criteria to establish a diagnosis of HP were consistent with those previously published from our institution.4 ,7 Non-HP ILD cases were identified from the database by selecting the non-HP patients immediately before and after each HP case included, resulting in a 1:2 case:control ratio. All non-HP ILD diagnoses were allowed; only ‘unclassifiable ILD’ cases were excluded.

Study variables

Baseline demographic and clinical data included in the analyses were age, sex, smoking history and pulmonary function test values. Self-reported, standardised and prospectively collected exposure histories to birds, down feathers, moulds, water damage, indoor hot tubs, steam room/saunas, humidifiers or farming were also included. All HRCT scans included thin section (<1.25 mm collimation) inspiratory images. Each scan was scored for specific features (see online supplementary table S1) and overall diagnostic impression by an expert thoracic radiologist (BME) blinded to the patient's clinical data and diagnosis.

Supplemental material

Prediction models were derived in the subset of patients with expiratory images,8 and validated in the remainder. Predictor variables associated (p<0.10) with a diagnosis of chronic HP were considered for inclusion in multivariate models. Candidate models were compared using the C-statistic, estimated using 10-fold cross-validation. For the models with the highest cross-validated C-statistic (CV-C), goodness of fit (GOF) was assessed using the p value for the Hosmer-Lemeshow (H-L) statistic.9 Two diagnostic models were developed that varied by HRCT variables included. Model 1 included specific radiographic findings, while model 2 included the radiologist's diagnostic impression (dichotomised into low vs moderate or high likelihood of chronic HP). Final models were selected for CV-C at or near the maximum observed value, H-L p value >0.10, clinical feasibility (ie, the availability and ease of use for included variables) and face validity (ie, the clinical appropriateness of included variables). Receiver operating characteristic curves were constructed for each model. Model performance was tested in the validation cohort, as well as among the subset of patients with either chronic HP or idiopathic pulmonary fibrosis (IPF). All analyses were done using Stata V.13.1 (Stata, College Station, Texas, USA).

Results

Two hundred and sixty-one patients were identified from the database (87 chronic HP, 174 non-HP ILD) and 190 met eligibility criteria (66 HP, 124 non-HP ILD controls) (see online supplementary figure S2). The derivation cohort consisted of 124 patients (44 HP, 80 non-HP ILD controls). The validation cohort consisted of 66 patients (22 HP, 44 non-HP ILD controls). Baseline characteristics of the study population are presented in table 1. Diagnoses included in the control populations are presented in online supplementary table S2.

Table 1

Baseline patient characteristics

The top three candidates for model 1 and model 2 are presented in table 2. For model 1, the best overall model as defined by its performance, clinical feasibility and face validity included the following predictor variables: age, a history of down feather and/or bird exposure, the presence of diffuse craniocaudal ground-glass opacity on HRCT, and the presence of mosaic perfusion on HRCT (CV-C=77.8, H-L GOF p value=0.60) (figure 1A). This model demonstrated consistent performance in the validation cohort (CV-C=79.8, H-L GOF p value=0.89). Using model 1 to derive an ‘HP score’ point value (range 0–100), HP scores ≥63 demonstrated a specificity of 91% and sensitivity of 48% for the diagnosis of chronic HP. For model 2, the best overall model included the following: age, history of down feather and/or bird exposure and a moderate–high confidence in the radiographic diagnosis of HP (CV-C=75.8, H-L GOF p value=0.72) (figure 1B). This model also demonstrated consistent performance in the validation cohort (CV-C=86.9, H-L GOF p value=0.35). Using model 2, HP scores ≥57 demonstrated a specificity of 91% and sensitivity of 50% for the diagnosis of chronic HP. Details of both models, their performance in the subgroup of patients with chronic HP or IPF, and determining the performance characteristics of various cut-off values are provided in the online supplementary materials and supplementary tables S3a and S3b.

Table 2

Clinical prediction model performance for diagnosis of HP

Figure 1

Receiver operating characteristic curve for (A) the derivation and validation prediction models incorporating clinical, exposure and individual radiological variables and (B) the derivation and validation prediction models incorporating clinical, exposure and radiologist's confidence variables. The solid and dashed lines represent the derivation and validation cohorts, respectively. The curve areas to the left of the vertical line indicate diagnostic specificity ≥90%.

Discussion

In this study, we developed multidimensional diagnostic models to identify patients with chronic HP with a high degree of specificity. In such patients, surgical lung biopsy can likely be avoided. While a clinical diagnosis of HP is already made in many ILD practices, these data are the first we are aware of that provide evidence to support this approach in selected patients.

There are several strengths to our models. First, they use readily available clinical parameters such as patient demographics, self-reported exposures and radiographic features. Second, they were derived in biopsy-proven patients with chronic HP, likely representative of those patients with increased diagnostic uncertainty (ie, those in whom a surgical biopsy would be considered), the most challenging patients encountered in clinical practice. Lastly, our models were internally validated.

mportantly, our models do not incorporate diagnostic tests that may have value in the diagnostic evaluation of chronic HP, as they were not routinely performed in our cohort. These include bronchoalveolar lavage differential cell counts, antigen-specific circulating antibodies and inhalational challenge.1 ,10 We excluded unclassifiable ILD and therefore cannot comment on model performance in this subset of patients. This represents a limitation to its application in clinical practice and requires further investigation. In addition, inciting antigens likely vary geographically, and specific exposures may be more or less predictive in other parts of the world. While this would be expected to primarily affect the sensitivity and not the specificity of the models, validation of these models in geographically diverse cohorts is essential.

Acknowledgments

We would like to thank the patients, providers and staff of the UCSF Interstitial Lung Disease Program for their time and commitment to clinical research, and Jane Berkeley for data management and retrieval. We would also like to thank Dr Cyrus Shariat, Dr W Richard Webb, Dr Theodore Lee and Dr Antonio Gomez for their thoughtful contributions to the development of this project.

References

View Abstract

Footnotes

  • Contributors KAJ and HRC conceived of the project; KAJ, BME, DA, KdB, JAG, KDJ, LLK, JSL, BL, PJW and HRC generated data for the project; KAJ, EV and HRC analysed data for the project and wrote the manuscript; KAJ, BME, EV, DA, KdB, JAG, KDJ, TEK, LLK, JSL, BL, PJW and HRC revised the manuscript.

  • Funding GSK/University of Calgary Advanced Fellowship in Respirology Grant (KAJ); Nina Ireland Program for Lung Health (HRC).

  • Competing interests KAJ reports travel support from InterMune and personal fees from Boehringer Ingelheim and Hoffman-La Roche; KdB reports travel support from InterMune and personal fees from AstraZeneca and Boehringer Ingelheim; TEK reports personal fees from InterMune, ImmuneWorks, GlaxoSmithKline, Daiichi Sankyo and Boehringer Ingelheim; HRC reports personal fees from AstraZeneca, Bayer, Biogen, FibroGen, Genentech/InterMune, Genoa, Gilead, GlaxoSmithKline, Mesoblast, Moerae Matrix, Pfizer, Promedior, Prometic and PatientsLikeMe, grants from Boehringer Ingelheim, consultancy fees from the Pulmonary Fibrosis Foundation and grant funding from the NIH/NHLBI.

  • Ethics approval University of California, San Francisco's Human Research Protection Program Committee on Human Research.

  • Provenance and peer review Not commissioned; externally peer reviewed.