Skip to main content

Artificial intelligence in COPD CT images: identification, staging, and quantitation

Abstract

Chronic obstructive pulmonary disease (COPD) stands as a significant global health challenge, with its intricate pathophysiological manifestations often demanding advanced diagnostic strategies. The recent applications of artificial intelligence (AI) within the realm of medical imaging, especially in computed tomography, present a promising avenue for transformative changes in COPD diagnosis and management. This review delves deep into the capabilities and advancements of AI, particularly focusing on machine learning and deep learning, and their applications in COPD identification, staging, and imaging phenotypes. Emphasis is laid on the AI-powered insights into emphysema, airway dynamics, and vascular structures. The challenges linked with data intricacies and the integration of AI in the clinical landscape are discussed. Lastly, the review casts a forward-looking perspective, highlighting emerging innovations in AI for COPD imaging and the potential of interdisciplinary collaborations, hinting at a future where AI doesn’t just support but pioneers breakthroughs in COPD care. Through this review, we aim to provide a comprehensive understanding of the current state and future potential of AI in shaping the landscape of COPD diagnosis and management.

Background

The prevalence and impact of COPD

Chronic obstructive pulmonary disease (COPD) remains one of the predominant public health challenges of the 21st century. With a global footprint spanning diverse demographic and geographic settings, COPD presents a multifaceted clinical picture marked by persistent respiratory symptoms and airflow limitations due to airway and alveolar abnormalities [1]. Primarily driven by prolonged exposure to noxious particles or gases, especially those originating from tobacco smoking, the pathological underpinnings of COPD are complex and varied, involving chronic inflammation, structural changes, and repair processes that affect both the larger airways and the peripheral lung [2].

As of the latest global estimates, over 250 million people suffer from COPD worldwide, making it the third leading cause of death by 2030 [3]. Economically, the disease also places a heavy toll on healthcare systems, with direct medical costs and productivity losses amounting to billions annually [4]. These alarming statistics underline the pressing need for early identification, accurate staging, and effective management strategies.

This review aims to journey through the convergence of COPD imaging from computed tomography (CT) with the capabilities of AI, underscoring the current achievements, challenges, and the road ahead.

The power of CT imaging in COPD

Medical imaging has proven indispensable in the landscape of respiratory medicine [5]. CT imaging, with its high resolution and ability to visualize lung structures in detail, has provided a platform for in-depth investigations into the complex manifestations of COPD [6, 7].

Unlike spirometry, which gives a global measure of lung function, CT imaging can localize and characterize the pathological abnormalities of COPD [8]. It illuminates the heterogeneity inherent in the disease-whether it be the bullous formations of emphysema, thickening of the bronchial walls, or alterations in the pulmonary vessels [2, 6, 9, 10]. CT imaging is crucial not just for diagnosis but also for tailoring patient-specific interventions, monitoring disease progression, and evaluating therapeutic efficacy [11].

Further, staging of COPD, which is vital for prognostication and management decisions, has traditionally leaned heavily on physiological parameters. However, the staging paradigm is experiencing a shift. With the advent of quantitative CT techniques, objective measurements related to airway thickness, lung volume, and parenchymal attenuation are becoming part of the COPD assessment lexicon [12, 13]. These metrics provide a more nuanced understanding of disease severity and its spatial distribution within the lungs.

As COPD comprises heterogeneous imaging phenotypes, including emphysema, airway changes, and vessel modifications, CT imaging offers a non-invasive window into the structural abnormalities that define these phenotypes [6, 14, 15]. With the surge in technological advancements, particularly the integration of artificial intelligence (AI) in medical imaging, the landscape of COPD diagnosis and management is set for a transformative shift [16, 17].

Yet, the vast potential of CT imaging is not without its challenges. The sheer volume of data from high-resolution scans demands intensive manual scrutiny, making the interpretation time-consuming and prone to variability [18]. Herein lies the promise of AI. By harnessing algorithms trained on vast datasets, AI can automate, augment, and refine the image analysis process, bringing precision and consistency to the forefront [19,20,21].

AI in medical imaging: an overview

Machine learning vs. deep learning: differentiating the two foundational approaches

AI encompasses a spectrum of techniques aimed at simulating human-like intelligence. Two standout sub-domains are machine learning (ML) and deep learning (DL), each having distinct attributes and applications in medical imaging.

ML employs algorithms that learn from data to make predictions. In medical imaging, these algorithms often use labeled datasets to discern patterns and offer diagnostics [22, 23]. Various types include: supervised learning: algorithms learn from labeled data, aiding in tasks like disease subtype classification [24] and treatment response prediction [25]. Unsupervised learning: patterns are identified without labeled data, useful for segmenting similar regions in CT images [26].

DL, an advanced subset of ML, operates on multi-layered neural networks. The star of DL in medical imaging is the convolutional neural network (CNN), optimized for image processing by detecting features ranging from basic edges to complex patterns [27]. Its prowess in COPD imaging stems from its ability to discern intricate image features, offering detailed classifications [28]. A trinity of factors supports DL’s ascendancy in imaging: extensive labeled datasets, sophisticated network designs, and burgeoning computational power [29].

Figure 1 illustrates two fundamental concepts in deep learning: the multi-layer perceptron (MLP) and convolutional operations. Figure 1a depicts a multi-layer perceptron, which is a type of feedforward artificial neural network. The MLP consists of an input layer, one or more hidden layers, and an output layer. In this specific example, the input layer comprises three variables (or features), denoted as x1, x2, and x3. These input variables are fed into the hidden layers, where each neuron applies a nonlinear activation function to the weighted sum of its inputs. The activation function introduces non-linearity into the network, enabling it to learn complex mappings between the input and output spaces. The output layer of the MLP produces two values, representing the probabilities of the input belonging to two different classes, denoted as p1 and p2. The MLP learns to classify the input data by adjusting the weights of the connections between neurons during the training process, typically using optimization algorithms such as stochastic gradient descent.

Fig. 1
figure 1

Fundamental concepts of deep learning. a MLP Prediction with 2 outputs. b Convolutional operation

Figure 1b illustrates the process of convolutional operations, which are fundamental building blocks in CNN. CNN is particularly effective in processing grid-like data, such as images or time series. In this example, the input is a three-channel tensor, which can be thought of as an RGB image. The convolutional operation involves sliding a set of learnable filters (or kernels) over the input tensor, performing element-wise multiplications, and summing up the results to produce a feature map. The figure shows two convolutional kernels being applied to the input tensor, resulting in two output channels. Each kernel has a specific set of weights that are learned during the training process to extract relevant features from the input. The convolutional operation exploits spatial locality and parameter sharing, enabling the network to learn translation-invariant features and capture local patterns in the data. By stacking multiple convolutional layers, CNNs can learn hierarchical representations, with lower layers capturing low-level features (e.g., edges and textures) and higher layers capturing more abstract and semantic information.

The combination of MLPs and convolutional operations forms the backbone of many deep learning architectures. MLPs are commonly used in the final stages of CNNs for classification or regression tasks, while convolutional layers are employed to extract spatial features from the input data. By leveraging the power of deep neural networks and convolutional operations, deep learning has achieved remarkable success in various domains, including computer vision [30], natural language processing [31], and speech recognition. The ability to automatically learn hierarchical representations from raw data has revolutionized the field of artificial intelligence and opened up new possibilities for solving complex problems.

Figure 2 presents an overview of several influential CNN architectures. The figures are referred to the previous study [32]: (a) AlexNet: Developed by Krizhevsky et al. in 2012 [33], AlexNet is a pioneering CNN architecture that achieved remarkable performance on the ImageNet classification task. It consists of five convolutional layers followed by three fully connected layers. AlexNet introduced the use of rectified linear unit (ReLU) activation functions and employed techniques such as data augmentation and dropout regularization to improve generalization. (b) ResNet18: ResNet, short for Residual Network, is a family of CNN architectures introduced by He et al. in 2015 [30]. ResNet18 is a specific instance of the ResNet architecture with 18 layers. The key innovation of ResNet is the introduction of residual connections, which allow the network to learn residual functions with reference to the input layer, enabling the training of much deeper networks without the vanishing gradient problem. (c) MobileNet-v2: MobileNet is a family of efficient CNN architectures designed for mobile and embedded vision applications. MobileNet-v2, proposed by Sandler et al. in 2018 [34], builds upon the ideas of depthwise separable convolutions and introduces inverted residual connections. This architecture achieves a good balance between accuracy and computational efficiency, making it suitable for resource-constrained devices. (d) ResNet26: ResNet26 is another variant of the ResNet architecture, similar to ResNet18 but with 26 layers [35]. It follows the same principles of residual learning, allowing for the training of deeper networks while mitigating the vanishing gradient problem. ResNet26 offers a trade-off between network depth and computational complexity. (e) VGG16: VGG16, introduced by Simonyan and Zisserman in 2014 [36], is a CNN architecture known for its simplicity and effectiveness. It consists of 16 layers, including 13 convolutional layers and 3 fully connected layers. VGG16 uses small convolutional filters (3x3) and employs a uniform architecture, making it easy to understand and implement. Despite its depth, VGG16 has been widely adopted and has served as a foundation for many subsequent CNN architectures.

Fig. 2
figure 2

Some main CNN architectures referring to [32]. a AlexNet, b ResNet18, c Mobilenet-v2, d ResNet26, and e VGG16

These CNN architectures have been instrumental in advancing the field of deep learning and have been applied to various computer vision tasks, including image classification, object detection, and semantic segmentation. Each architecture has its own unique characteristics and trade-offs in terms of accuracy, computational complexity, and memory requirements. Understanding the design principles and performance characteristics of these architectures is crucial for selecting the appropriate model for a given task and for developing new CNN architectures tailored to specific domain requirements, such as medical image analysis for the characterization and recognition of COPD using CT scans [37, 38].

In sum, while ML and DL both offer value in medical imaging, the selection between them hinges on the specific demands of the task at hand. Their differentiated roles will be further proved as we delve deeper into AI’s integration with COPD imaging.

Applications of AI in medical imaging

Radiology, an indispensable branch of medical diagnostics, has always been at the forefront of technological integration. The recent embrace of AI within radiology signifies a paradigm shift, augmenting the accuracy, efficiency, and capabilities of imaging modalities. (1) Image analysis and interpretation: AI-powered tools have streamlined image analysis, enhancing detection sensitivity and reducing manual errors. For instance, algorithms can assist radiologists in identifying early signs of pathologies like tumors, vascular anomalies, or pulmonary conditions that might be subtle or ambiguous in initial scans [39,40,41,42]. (2) Workflow optimization: Beyond image interpretation, AI facilitates streamlined workflows in radiological settings. Tools can prioritize reading lists based on urgency, predict no-show appointments, or automate documentation processes, thus improving clinical efficiency [43, 44]. (3) Radiation dose reduction: One of the concerns in radiological procedures is the radiation dose. AI algorithms can reconstruct high-quality images from lower-dose scans, striking a balance between image clarity and patient safety [45, 46]. (4) Predictive analysis: AI’s ability to integrate imaging data with electronic health records allows for predictive modeling. This offers insights into potential disease trajectories, response to treatments, or even risks of complications, enabling personalized patient care [47, 48]. (5) Advanced imaging techniques: AI augments traditional imaging modalities like MRI or CT with advanced techniques. For instance, AI-powered synthetic MRI can generate multiple image sequences from a single acquisition, reducing scan times [49,50,51]. As radiology continues its evolutionary journey, AI stands as a beacon, promising transformative changes. Its integration not only amplifies diagnostic precision but also heralds a more patient-centric approach, where tailored interventions and enhanced safety become the norm.

According to our previous studies [32, 52, 53], the following will provide three examples of the application of AI, especially deep learning methods, to lung CT images.

Figure 3 introduces the Vision Transformer (ViT) and its application in emphysema subtype classification. The ViT model, introduced by Dosovitskiy et al. in 2020 [54], adapts the Transformer architecture, originally designed for natural language processing, to the task of image classification. In this specific application, the ViT model is employed to classify emphysema subtypes, which are important indicators of COPD severity and progression.

Fig. 3
figure 3

The architecture of Vision Transformer and its application in emphysema subytpe classification [52]

The input to the ViT model is a large patch of slice in CT scan, which is preprocessed and divided into a sequence of fixed-size patches. These patches are linearly embedded and augmented with positional embeddings to preserve spatial information. The embedded patches are then fed into the Transformer encoder, which consists of multiple layers of multi-head self-attention and feed-forward networks. The self-attention mechanism allows the model to capture long-range dependencies and learn global context, enabling it to effectively capture the spatial patterns and textural characteristics of emphysema subtypes.

The output of the Transformer encoder is a sequence of feature vectors, which are then aggregated using a classification token and passed through a MLP to obtain the final emphysema subtype predictions. The ViT model is trained on a labeled dataset of CT scans with annotated emphysema subtypes, using techniques such as data augmentation and transfer learning to improve generalization and performance.

The application of ViT for emphysema subtype classification offers several advantages. First, the self-attention mechanism enables the model to capture long-range dependencies and global context, which is crucial for accurately identifying the spatial patterns and textural characteristics of different emphysema subtypes. Second, the ViT architecture is highly scalable and can be trained on large datasets, for example, ImageNet, allowing for the learning of rich and expressive feature representations. Finally, the ViT model has shown promising results in various medical image analysis tasks, demonstrating its potential for improving the accuracy and efficiency of emphysema subtype classification in COPD assessment.

Moreover, as depicted in Fig. 4, The generative adversarial network (GAN) related method was proposed for synthesizing contrasted-enhanced or non-contrasted CT. The proposed synthesizer’s network architecture comprises a generator and a discriminator. The generator’s purpose is to create synthetic images, while the discriminator’s role is to differentiate between authentic and generated images, enabling the generator to learn to produce realistic contrasted enhanced (CE) CT or Non-contrasted (NC) CT images.

Fig. 4
figure 4

The generative adversarial network (GAN) for synthesizing contrasted-enhanced or non-contrasted CT [53]. The proposed Synthesizer based on GAN and consists the generator and the discriminator

The NC CT is input into generator (G1) to generate SynCECT. The NC CT and SynCECT are then concatenated along the channel dimension and fed into the discriminator (D1). The discriminator generates a probability map indicating whether the input image is a SynCECT or a real CE CT. Additionally, the NC CT and real CE CT are combined and input into the discriminator to produce a probability map. The generator and discriminator continue to compete until an equilibrium is achieved. The backbone of the generator (Fig. 4) is based on the 3D U-Net architecture, which consists of an encoder and a decoder section.

Figure 5 depicts the multiple instance earning (MIL) for COPD identification using CT scans. MIL is a weakly supervised learning paradigm where the training data consists of labeled bags, each containing multiple instances [55]. In the context of COPD recognition, a bag corresponds to a CT scan, and instances within the bag represent different regions or patches of the scan. The MIL framework is particularly suitable for COPD recognition because it can handle the heterogeneous nature of the disease, where the presence of COPD may be indicated by local patterns or abnormalities in specific regions of the CT scan. Instead of requiring detailed pixel-level annotations, MIL allows for the learning of COPD patterns from weakly labeled data, where only the overall COPD status of each CT scan is provided.

Fig. 5
figure 5

The attention-guided MIL for COPD identification using CT images [32]

The MIL-based COPD identification pipeline consists of the following steps: (1) Patch extraction: The CT scans are divided into multiple slices, which serve as instances within the MIL framework. These slices can be extracted using techniques such as sliding windows or unsupervised segmentation methods.(2) Feature extraction: Each patch is transformed into a feature representation using a deep learning model, such as a CNN. The CNN is trained to extract discriminative features that capture the local patterns and abnormalities associated with COPD. (3) Attention-guided instance-level prediction: An attention mechanism is introduced to assign different weights to the instances based on their relevance to the COPD prediction. The attention mechanism learns to focus on the most informative slices while suppressing the less relevant ones. This is achieved by computing attention scores for each instance, which are then used to weigh the instance-level predictions. The attention scores can be learned through a separate attention network or by incorporating attention layers within the MIL framework. (4) Bag-level aggregation: The instance-level predictions are aggregated to obtain a bag-level prediction, representing the overall COPD status of the CT scan. Common aggregation strategies include max pooling, average pooling, or attention-based mechanisms that assign different weights to the instances based on their relevance to the COPD prediction.

The application of MIL for COPD identification offers several advantages. First, it allows for the learning of COPD patterns from weakly labeled data, reducing the need for detailed pixel-level annotations. Second, MIL can handle the heterogeneous nature of COPD, capturing local patterns and abnormalities that may be indicative of the disease. Third, the MIL framework is flexible and can be combined with various deep learning architectures and aggregation strategies to improve the accuracy and robustness of COPD recognition.

Literature search and review

To compile the necessary literature for this study, two separate reviewers independently scoured Google Scholar, Web of Science, and PubMed for articles applying AI/ML methodologies in COPD-related research, from their inception through November 2023.

A targeted search was conducted on Google Scholar, a well-regarded bibliographic retrieval database, using [“Artificial Intelligence” AND COPD], [“Machine Learning” AND COPD], [“Deep Learning” AND COPD], [“Convolutional Neural Networks” AND COPD], [“Detection” AND COPD], [“Classification” AND COPD], [“Airway” AND COPD], [“Vessel” AND COPD], [“Classification” AND “emphysema”], [“Segmentation” AND “emphysema”], [“Classification” AND “Airway”], [“Segmentation” AND “Airway”], [“Classification” AND “Vessel”], [“Segmentation” AND “Vessel”], [“Segmentation” AND “artery-vein”]. as the main search terms. The same search strings are employed to search on Web of Science and PubMed.

The articles gathered from the three databases were aggregated, and duplicate entries were eliminated. We retained only original research articles written in English. The remaining articles underwent a review process to ensure that only those pertinent to the study were kept.

Upon finalization, we had a collection of 126 articles. These were classified into four key categories based on their content and objectives: (1) COPD identification and staging (2) emphysema subtype classification and segmentation, (3) airway segmentation and quantification in COPD, and (4) vessel segmentation and quantification in COPD. The step-by-step process is illustrated in Fig. 6.

Fig. 6
figure 6

Literature search and analysis. The PRISMA flowchart for this review

AI techniques for COPD identification, staging, and imaging phenotype

Some articles have reviewed the progress of AI techniques in COPD [56,57,58]. Exarchos et al. reviewed the general adoption of AI in COPD research, categorizing the studies into ‘COPD diagnosis’, ‘COPD prognosis’, ‘Patient classification’, and ‘COPD management’. It identified an acceleration of AI use in COPD research and calls for broader adoption due to the large and complex data involved [57]. The article published by Nikolaou et al. focused on the use of machine learning algorithms, specifically cluster analysis, to better characterize COPD through integration of patient characteristics like symptoms, comorbidities, biomarkers, and genomic information. It reviewed the progress of research in the past decade using cluster analysis for COPD phenotypes [56]. Estépar’s article provided an introduction to AI and deep learning, discussing their role in understanding the evolution and divergent trajectories of COPD. It highlighted the successes of AI in clinical decision making, radiological interpretation, prognostication, and presents opportunities, challenges, and limitations of AI in COPD [58].

Our review article, compared to the others, provides a more in-depth focus on machine learning and deep learning techniques for COPD identification, staging, and imaging phenotypes, emphasizing the roles of emphysema, airway dynamics, and vascular structures. While all articles discuss AI’s role in COPD, we have delved deeper into the specifics of AI-powered insights and the complexities of integrating AI into the clinical landscape. Unlike the other articles, we also provide a comprehensive understanding of the current state and future potential of AI in shaping COPD diagnosis and management.

COPD remains a prevalent respiratory condition, with imaging playing a pivotal role in its diagnosis and management. Figure 7 shows emphysematous destruction, airway, and vascular structure variability in two COPD subjects with different stages obtained with our previous studies [59, 60]. In the following section, we will review the AI in COPD Imaging from four aspects: COPD Identification and Staging, emphysema subtyping, airway analysis, and vascular changes.

Fig. 7
figure 7

Coronal computed tomography (CT) view, emphysema region(in red), airway and vascular morphology in two COPD subjects with different stages. (Top) Male, Age = 51, BMI = 28.52, GOLD 1, FEV1 = 3.29, FVC = 4.75. (Bottom) Male, Age = 81, BMI = 25.49, GOLD 2, FEV1 = 1.71, FVC = 3.77. Emphysema index was calculated using a threshold of -950 HU in inspiratory CT scan [61]. The lung airway was extracted using the method in [59]. And lung vessel was segmented by [60]. BMI: body mass index; COPD: chronic obstructive pulmonary disease; FEV1: forced expiratory volume in one second; FVC: Forced Vital Capacity; GOLD: Global Initiative for Chronic Obstructive Lung Disease

COPD identification and staging with AI

Understanding COPD through radiomics: advanced features extracted from CT images

All of the papers related to COPD identification and staging with AI are summarized in Table 1. Radiomics, an emerging field in medical imaging, involves the extraction of a large number of quantitative features from radiographic images [62]. In the context of COPD, radiomics can provide numerous features from CT scans, capturing detailed information about lung morphology and texture that can indicate disease presence and severity [63]. For example, first order features describe the distribution of voxel intensities within the image through basic metrics such as mean, variance, skewness, and kurtosis. Shape features provide information about the 3D shape of the lung and its irregularities. Matrix-based features like the gray level co-occurrence matrix (GLCM), gray level size zone matrix (GLSZM), gray level run length matrix (GLRLM), neighboring gray tone difference matrix (NGTDM), and gray level dependence matrix (GLDM) describe more complex characteristics. The GLCM quantifies the texture by examining the spatial distribution of voxels. GLSZM captures the size zones of different gray levels, while GLRLM gives information about the lengths of consecutive voxels with the same gray level. NGTDM measures the difference between a voxel and its neighbor, and GLDM quantifies the dependencies of gray level values in the image.

Table 1 Summary of COPD identification and staging by different teams

Cheplygina and his colleagues demonstrated that machine learning methods could effectively utilize radiomics features to classify and detect COPD in chest CT images [67]. The proposed method had validated on multi-dataset and achieved a moderate area under the curve (AUC) (DLCST:0.684, COPDGene: 0.962, and Frederikshavn: 0.969). Li et al. showed that lung radiomics features, when combined with a support vector machine (SVM) model, could accurately classify different stages of COPD, outperforming conventional methods, achieving 0.970 of AUC in identification and 0.799 AUC in staging [76]. According to the study, machine learning methods, coupled with radiomics features, accurately classified the stages of COPD and outperformed classical CNN [77]. Yang et al. explored how a multi-layer perceptron classifier, coupled with selected lung radiomics features, can effectively characterize and classify COPD stages [80]. Moreover, he proposed a novel lung radiomics combination vector and an auto-metric graph neural network with a meta-learning strategy for effective COPD stage classification [81]. Amudala’s study suggested that a combination of radiomics features, representing parenchymal texture and lung and airway shape of CT images, could be used to accurately detect COPD [86]. The proposed method gain tremendous classification potential in inspiratory low-dose (AUC=0.88) and standard-dose CT scans (AUC=0.90) .

Deep learning and COPD: unveiling disease markers from CT images

Deep learning techniques, especially CNNs, have shown remarkable success in image analysis tasks. CNNs have the capability of automatically learning hierarchical features from raw CT images, which can capture both primitive (edges, textures) and more abstract features (shapes, structures) in the data. In the context of COPD, CNNs can be trained to identify structural changes in the lung indicative of the disease, including emphysema, airway thickening, and the presence of mucus plugs. They can also be used to quantify the extent of these changes, providing a measure of disease severity. The advantage of deep learning lies in its ability to learn from large-scale data, capturing subtle changes and complex patterns that might be overlooked by human experts or traditional image processing techniques.

In COPD identification, Sathiya et al. presented a computer-aided diagnosis system for COPD that employed CNN to classify CT images, with a process encompassing preprocessing, feature extraction, segmentation, and classification, aiming to enhance the accuracy of COPD diagnosis under noisy conditions [68]. Tang and his colleagues demonstrated the effectiveness of deep residual networks for the automated detection of COPD using low-dose CT scans, achieving an AUC of 0.889 in PanCAN and 0.886 in ECLIPSE [72]. Li et al. presented a novel method for early COPD detection using a graph convolution network model applied on small and weakly labeled chest CT images [79]. It gained robust classification performance (AUC:0.81). Moreover, deep CNN transferred Multiple Instance Learning (DCT-MIL) approach [32, 70, 84] and self-supervised learning [82, 87] were employed to identify COPD from CT images. In COPD staging, Ying et al. utilized a deep belief network to develop a highly accurate automatic classifier for COPD severity, demonstrating its effectiveness as a tool for exacerbation risk assessment in COPD patients (accuracy (ACC)=97.2%) [65]. While another study developed a deep learning-based algorithm to stage COPD severity through quantification of emphysema and air trapping from CT images, showing that automated CT algorithms can facilitate COPD severity staging and provided prognostic value [74]. The proposed method achieved remarkable discriminative capacity in COPD staging with a sensitivity (SEN) of 88.25%, a specificity (SPE) of 74.5%, and an AUC of 0.905.

Some works propose deep learning methods for both identification and staging. Gonzalez et al. demonstrated that the several layers’ CNN could effectively detect and stage COPD, predict acute respiratory disease events and mortality in smokers using only CT imaging data, suggesting that CNN analysis could be a powerful tool for risk assessment at a population level [66]. Singla and his colleagues developed a deep learning model to extract informative regional image features from HRCT of COPD patients, demonstrating strong predictive capabilities for spirometric obstruction, emphysema severity, exacerbation risk, and mortality, and potentially improving both research and clinical practice [69]. Moreover, the CNN models which were trained using images of lung parenchyma and bronchial wall [78] and weakly supervised deep learning [38] to diagnose and grade COPD.

In the deep learning methods for COPD identification and staging, researchers have employed various neural network architectures, numbers of layers, and activation functions. Based on the information provided in Table 2, the following three points can be summarized: CNNs are the most commonly used neural network architecture in COPD identification and staging. Researchers have utilized various classic CNN architectures, such as AlexNet, ResNet, DenseNet, and VGG, as well as some self-designed CNN architectures. These architectures typically consist of multiple convolutional layers and pooling layers to extract features from the images. The number of layers in the neural networks varies considerably, ranging from as few as 3 layers to as many as 201 layers. Some studies have used shallower networks, such as González et al. [66] and Ho et al. [73], who employed CNNs with 3 convolutional and pooling layers. In contrast, other studies have utilized deeper networks, such as Zhang et al. [78] with DenseNet-201 (201 layers) and Tang et al. [72] with ResNet-152 (152 layers). The Rectified Linear Unit (ReLU) is the most frequently used activation function in these deep learning models for COPD identification and staging. Most of the studies listed in the table have employed ReLU as their activation function, with the exception of Du et al. [71] who used Leaky ReLU and Yu et al. [87] who used Sigmoid. The choice of activation function can impact the model’s ability to learn complex patterns and generate meaningful representations of the input data.

Table 2 More details of deep learning method in COPD identification and stage, including neural network architectures, number of layers, and activation functions

In the preprocessing, feature extraction, and feature selection methods employed in deep learning approaches for COPD identification and staging, researchers have utilized various techniques to enhance the performance of their models. Based on the information provided in Table 3, the following three points can be summarized: (1) Preprocessing techniques, such as segmentation of the lung and airways, are commonly used to focus the analysis on the relevant regions of interest in the CT images. Several studies, including Mets et al. [64], Xu et al. [70], Hasenstab et al. [74], and Puchakayala et al. [86], have employed lung and airway segmentation as a preprocessing step. Other preprocessing methods include joining multiple views into a single montage (González et al. [66]), extracting 3D regions of interest (Cheplygina et al. [67]), and applying grayscale conversion (Sathiya et al. [68]).(2) Feature extraction methods can be broadly categorized into two main approaches: (a) using pre-trained or self-designed CNN architectures to automatically learn relevant features from the preprocessed images, and (b) calculating handcrafted features, such as radiomics features or specific quantitative CT biomarkers. Studies by González et al. [66], Xu et al. [70], and Tang et al. [72] have utilized CNN-based feature extraction, while others, such as Mets et al. [64], Hasenstab et al. [74], Li et al. [76], Yang et al. [80, 81], and Puchakayala et al. [86], have employed handcrafted features. (3) Feature selection methods are occasionally used to reduce the dimensionality of the extracted features and select the most informative ones for COPD identification and staging. Some studies, such as Xu et al. [70], have used principal component analysis (PCA) for feature selection, while others, like Li et al. [76] and Yang et al. [80, 81], have employed techniques such as variance threshold, Select K Best method, least absolute shrinkage and selection operator (LASSO), and generalized linear models. However, not all studies have explicitly mentioned the use of feature selection methods, suggesting that the choice of feature selection techniques may depend on the specific requirements and characteristics of the dataset and the model being used.

Table 3 The preprocessing, feature extraction, and feature selection method of deep learning method in COPD identification and stage

Beyond traditional methods: unique feature extraction from CT images for COPD identification and staging

Other innovative features can be extracted from CT images for COPD identification . For instance, the extraction of the bronchial tree structure provides a unique, detailed representation of the airways. Analyzing the branching patterns, diameters, and wall thicknesses of the bronchial tree can reveal valuable insights into the disease [78]. Additionally, A novel method diagnosed COPD using deep CNN to assess snapshots of 3D airway trees extracted from CT images [71] with an accuracy of 88.6%. Moreover, Wu and his colleagues further utilized snapshots of 3D airway trees and lung fields for COPD identification [37], achieving a higher accuracy of 94.7%. By combining these diverse features and original CT image features, They also created a comprehensive representation of the lung that can potentially improve the performance of machine learning models in COPD identification [32] with the noteworthy classification performance (ACC:95.8%, SEN:95.3%, and SPE:96.5%). Other study developed a new classification method for COPD based on CNN and a parametric-response mapping with an accuracy of 89.3% [73]. Future research should focus on developing effective feature integration strategies and exploring the potential of novel imaging features.

Emphysema imaging analysis through AI lenses

Emphysema is characterized by the irreversible destruction of alveoli, the lung’s air sacs. CT imaging is instrumental in its detection and quantification. However, with the integration of AI, the landscape of emphysema identification, subtyping, and phenotypic visualization is undergoing transformative changes [88].

Characteristics and subtypes

In emphysema, alveoli are damaged and enlarged causing breathlessness. CT scans, particularly high-resolution CT (HRCT), play a crucial role in diagnosing and characterizing emphysema. As defined in [10], emphysema can be categorized into several subtypes based on the distribution and appearance of the disease on CT scans (as shown in Fig. 8): centrilobular emphysema (CLE) is the most common type of emphysema seen in smokers. It typically begins in the center of the secondary pulmonary lobule, primarily affecting the bronchioles while sparing the peripheral portion. On CT images, it appears as areas of low attenuation without visible walls, often in an upper lobe predominant distribution. Panlobular emphysema (PLE) involves the entire secondary lobule. It is most commonly associated with \(\alpha\)1-antitrypsin deficiency. On CT scans, it presents as diffuse, low attenuation areas affecting all lung zones, but more severe in the lower lobes and anterior lung zones. Paraseptal emphysema (PSE) is characterized by the involvement of the distal airway structures adjacent to the pleura. It is often found in the upper lobes. On CT, it appears as subpleural areas of low attenuation, often with visible walls, and may be associated with bulla formation. In all these types, the degree of emphysema can be quantified on CT through densitometric analysis, which measures the lung’s mean attenuation and the percentage of the lung volume with low attenuation. These quantitative measurements can help assess the severity of the disease and monitor its progression. Moreover, the severity of emphysema is often categorized based on the percentage of the lung volume that falls below a certain Hounsfield Unit (HU) threshold on CT. It’s important to note these are general guidelines and the specific thresholds can vary depending on the source [89].

Fig. 8
figure 8

Examples of different lung tissue patterns are indicated by the red arrow.(a)Normal lung parenchyma (NLP);(b) centrilobular emphysema (CLE);(c) panlobular emphysema (PLE); and (d) paraseptal emphysema (PSE)

AI-based severity identification and subtype classification

AI has enhanced the diagnostic horizon for emphysema in multiple ways (Table 4):

Table 4 Summary of emphysema subtype classification methods by different teams, NT:normal tissue, CLE: centrilobular emphysema, PLE: panlobular emphysema, and PSE: paraseptal emphysema

Enhanced sensitivity with machine learning and traditional features: Machine learning models, trained on texton features, are proficient in identifying subtypes of emphysema. Gangeh et al. introduced a new texton-based classification system, coupled with a SVM model, for classifying emphysema in CT lung images, demonstrating superior accuracy (96.43%) over common techniques and slight improvement over recent methods based on local binary patterns [90]. Other methods including local binary patterns and joint intensity histograms [91], ensemble features based on log-Gabor filters, mean difference technique, and intensity values [104], representation by rotation invariant uniform local ternary pattern with Weber’s law [94], combination of local quinary patterns, multifractal features, and intensity histograms [100], was proposed for emphysema subtyping. Moreover, Zulueta et al. explored the use of a manifold learning technique with embedded probabilistic PCA for classifying different types of emphysema in CT lung images, demonstrating competitive performance with traditional texture-based and intensity distribution methods, as well as good visual agreement with actual emphysema types in full lung analysis [92].

Automatically extract features with deep learning: CNNs can identify and isolate emphysematous regions without the need for manual feature engineering, thus capturing the intricate patterns indicative of the disease. Karabulut et al. explored a CNN model to automatically identify and discriminate between subtypes of emphysema in high-resolution CT lung images, demonstrating promising accuracy levels and reduced processing time [93]. Peng and his colleagues presented a novel multi-scale residual network for automated emphysema tissue classification, achieving 93.74% accuracy, and introduced a new measure of emphysema severity based on the sum of centrilobular and panlobular emphysema, demonstrating strong correlation with pulmonary functions [98]. As illustrated in the work by Bermejo’s group, the Slice-Recovery network (SR-Net) was introduced, which was a novel convolutional network architecture that utilizes 3D contextual information for 2D segmentation of PSE lesions in CT images with the dice of 0.764 [97]. In the study led by Wu, they proposed a vision Transformer (ViT) [54] model for the classification of emphysema subtypes based on CT images, using large patches cropped from the images for embedding and classification. Using pre-training on ImageNet to overcome data limitations, the ViT model achieved an average accuracy of 95.95% on a proprietary dataset, outperforming AlexNet [33], Inception-V3 [105], MobileNet-V2 [106], ResNet34 [30], and ResNet50, as well as a non-pretrained ViT model. These results suggested that the proposed ViT model can accurately classify emphysema subtypes and had potential for other medical applications [52].

Automated evaluation the severity of emphysema: AI algorithms can categorize the severity of emphysema by automatically quantifying the affected lung volume, providing a more objective and consistent staging compared to manual densitometry. Bortsova et al. introduced an end-to-end deep learning method to estimate the extent of emphysema based on the proportion of diseased tissue. It outperforms traditional lung densitometry and other recent methods by a significant margin with an AUC of 0.89 [96].

Segmentation using deep learning: Deep learning, especially CNNs, also has demonstrated a marked aptitude for emphysema segmentation in CT scans. As reported by Peng and his fellow researchers introduced a new end-to-end semi-supervised framework for the semantic segmentation of emphysema in CT images using both annotated and unannotated areas. It was designed to reduce the workload for radiologists and annotation workload. The authors also propose a new loss function, the Fisher loss, to improve the model’s discriminative power. Experimental results demonstrate that this approach outperforms both the baseline supervised approach (which uses only annotated areas) and other state-of-the-art methods for emphysema segmentation ACC=82.6% [99].

Derived subtypes: Yang and his team investigated the possibility of using texture learning to identify novel emphysema specific lung texture patterns (sLTPs), which might correspond to previously unrecognized emphysema subtypes with distinct clinical traits. They employed advanced clustering techniques on emphysematous region textons within the MESA COPD cohort 49, pinpointing 12 unique sLTPs. A notable feature of this method was its incorporation of spatial data, since the regional distribution of emphysema is considered a significant phenotype 50. While most sLTPs displayed a strong correlation with dyspnea and exercise capacity, further research is essential to fully grasp their pathological significance [107, 108].

In conclusion, the incorporation of AI, particularly deep learning, has redefined the paradigms of emphysema detection and analysis. These advancements underscore the immense potential of AI in tailoring therapeutic interventions and prognostic assessments for patients with emphysema.

Airway analysis with AI

The airways, comprising bronchi and bronchioles, undergo significant structural changes in the face of COPD. HRCT imaging captures these alterations. AI’s integration with radiology accentuates the precision, scale, and depth of airway analysis, amplifying our understanding and management of COPD.

From visual cues to machine insights

Traditionally, radiologists discern airway changes by identifying bronchial wall thickening, luminal narrowing, and mucous plugging in scans [109]. While these visual cues remain foundational, they often necessitate expert scrutiny and can be subjective. AI bridges this gap by two aspects: Enhancing detection: AI models, trained on diverse datasets, unearth subtle airway alterations, magnifying early detection prospects [110, 111]. Standardizing evaluations: Algorithms ensure consistent airway analysis, minimizing inter-observer variability that can arise from manual evaluations [112, 113].

Segmentation and quantification: using AI to segment and measure airway changes

Segmenting the intricate airway structure and quantifying its changes is a complex endeavor. AI shines in this domain by precise segmentation. As shown in Table 5, deep learning algorithms, including CNNs and Transformer, can meticulously delineate airway structures from adjacent lung parenchyma, ensuring accuracy [114,115,116,117,118,119,120]. Charbonnier et al. improved the airway segmentation quality by detecting and removing leaks using a convolutional network, and combining multiple segmentations to increase the airway tree length (65.4%) in EXACT’09 while minimizing leaks [121]. A novel 2.5D convolutional neural network-based method was proposed for airway segmentation in volumetric chest CT scans with the tree length of 60.1% and a false positive rate (FPR) of 4.56% [122, 123]. Another research introduced AirwayNet, an innovative voxel-connectivity aware approach for precise airway segmentation in CT scans with a DSC of 90.2%. By transforming the conventional binary segmentation task into 26 connectivity prediction tasks, AirwayNet learns not only the airway structure but also the relationship between neighboring voxels [124]. Nadeem and his colleagues presented a novel multi-parametric freeze-and-grow propagation approach for automated and accurate segmentation of pulmonary airway trees in CT scans for exploring COPD sub-phenotypes. A CT intensity-based FG algorithm and a deep learning-based version are developed [125]. Graph neural network was also employed for extracting airways from chest CT data [126, 127]. A coarse-to-fine framework was proposed for addressing challenges in small airway branch segmentation [59, 128, 129]. Moreover, Wu et al. employed a novel 3D contextual transformer for accurate airway segmentation, extracting significantly more branches and longer lengths of the airway tree [59] with the tree length of 79.6% and a FPR of 8.27% on EXACT’09 dataset . Also, multi-task segmentation including airway and vessel was finished using CNNs-based methods [129,130,131]. Moreover, topology-guided iterative self-learning approach [132] and long-term slice propagation method [133] were proposed for improving tree length and branch detection.

Table 5 Summary of airway segmentation methods by different teams

AI quantifies airway alterations like wall thickness, diameter, or cross-sectional area. Such measurements facilitate objective assessment of airway disease severity and its progression over time. Nardelli et al. introduced a convolutional neural regressor trained with a generative model and Simulated and unsupervised generative adversarial network (SimGAN) to accurately characterize small pulmonary structures from CT images, overcoming the limitations of traditional methods. The validation results, both synthetic and in-vivo, showcase the promise of CNNs in providing accurate measurements of airway lumen, airway wall thickness, and vessel radius on chest CT images, potentially revolutionizing the diagnosis and treatment of pulmonary diseases [134].

AI enables airway changes aid in COPD identification

Beyond segmentation and quantification, AI’s prowess extends to classifying detected airway changes for diagnostic implications. It achieves this by: Pattern recognition: Algorithms detect specific airway patterns linked to COPD subtypes, enhancing diagnostic specificity [135, 136]. Integration with clinical data: AI models that merge imaging data with clinical parameters, such as spirometry, amplify the accuracy of COPD diagnosis and staging [137, 138]. Proactive predictions: Some models predict the risk of exacerbations or disease progression based on airway patterns, enabling clinicians to tailor management strategies proactively [138,139,140,141]. To encapsulate, AI’s role in airway analysis is transformative. By refining detection, segmentation, quantification, and classification processes, AI equips clinicians with invaluable insights, paving the way for personalized and effective COPD management.

Vessel imaging analysis through AI

Vascular changes, both within the pulmonary system and potentially extending extrapulmonary, are intricately linked with COPD’s pathology. The nuanced visualization and interpretation of these vascular dynamics become paramount, and this is where the synergy between radiology and AI offers promising avenues.

Traditional vs. AI-enhanced visualization

Historically, radiological evaluations of pulmonary vessels were reliant on the expertise of radiologists to discern vessel caliber changes, pruning, or other abnormalities from CT scans. While effective, there were some limitations. Interpretations could vary between experts, especially in borderline cases or early disease stages. Some microvascular changes could escape the naked eye, potentially delaying interventions. AI’s integration bridges these gaps. Algorithms ensure a uniform approach, reducing discrepancies in vessel interpretation. AI tools can potentially identify sub-millimeter vascular changes, offering a more comprehensive view of vessel dynamics [111, 142].

Segmentation and quantification of pulmonary vessel

As shown in Table 6, the presented studies focus on the challenging task of vessel segmentation and artery-vein separation in CT images. Various approaches, including CNN methods [143,144,145,146,147], generative adversarial networks [53], and transformer-based networks [60], were proposed to address issues such as the separation of pulmonary arteries and veins, the synthesis of non-contrast and contrast-enhanced CT images, and the segmentation of intricate vessel structures. These methods leverage advanced architectures, such as 3D contextual transformers and channel-enhanced attention modules, to improve accuracy and efficiency. Evaluation on diverse datasets demonstrates the effectiveness of the proposed techniques, showcasing their potential applications in diagnosing and planning treatments for lung diseases. Additionally, the studies highlight the importance of addressing challenges such as limited annotated data and the need for robustness to noise in medical image segmentation tasks. Jimenez et al. presented a graph-cut methodology and a random forest pre-classifier for the segmentation of pulmonary artery-vein (AV) structures in CT images with an F1-score of 79.5% [148]. Cui et al. proposed an efficient 2.5D segmentation network from three orthogonal axes, achieving superior performance in pulmonary vessel segmentation with a Dice score of 0.9272 [149]. Gu et al. introduced two techniques for pulmonary vessel suppression, demonstrating improved nodule detection for early lung cancer diagnosis [145]. Nam et al. developed a deep learning-based pulmonary vessel segmentation algorithm (DLVS), showcasing high accuracy (AUC=0.977) and clinical relevance for assessing vascular remodeling in COPD patients [150]. Wu et al. addressed the limitations of FCN and U-Net in vessel segmentation, proposing the MSI-U-Net with attention mechanisms and achieving state-of-the-art results (DSC:0.7168, SEN:0.7234, and precision:0.7893) [151]. Li et al. presented a novel 3D vessel segmentation network guided by edge profiles, demonstrating superior performance with a DSC of 0.789, especially in scenarios with limited training data [152]. Wang et al. leveraged spatial registration for automatic pulmonary vessel segmentation in NCCT images, achieving high Dice of 0.856 [153]. Pan et al. tackled the challenges of artery-vein separation, introducing the MSIA-Net with multi-scale fusion blocks and achieving remarkable segmentation performance DSC:81.7%, precision:80.5%, and FPR:0.069% [154]. Pang et al. proposed synthesizers for mutual synthesis of NCCT and CECT images, showcasing their effectiveness in pulmonary vessel segmentation [53]. Wu et al. contributed a transformer-based network for vessel segmentation and artery-vein separation, demonstrating high accuracy and applicability in CT images [60]. These methods collectively advance the field, offering innovative solutions for accurate and efficient vessel segmentation in medical imaging.

Table 6 Summary of vessel segmentation and artery-vein separation methods by different teams

Understanding vessel alterations requires not only meticulous segmentation but also quantification. AI quantifies vessel caliber, branching patterns, or other structural changes, offering metrics that are pivotal for COPD assessments and potential therapeutic responses. The quantitative analysis indicators of the vascular tree in patients with COPD provide significant insights about the disease’s progression and the patient’s overall health. These indicators include total blood volume (TBV), surface area, total volume of vessels with a cross-sectional area smaller than 5 mm2 (BV5), and BV5/TBV [155].

Extrapulmonary abnormalities & AI

While COPD predominantly affects the lungs, extrapulmonary manifestations, including cardiovascular implications, are notable [156, 157]. AI plays a role in the following aspects. Algorithms trained on diverse datasets can detect subtle extrapulmonary changes, offering insights into systemic COPD effects [58, 158]. AI not only visualizes but also interprets these manifestations in the context of COPD, potentially predicting risks like cardiac events or other systemic complications [159]. By merging pulmonary and extrapulmonary data, AI offers a holistic view of COPD’s impact, guiding comprehensive patient management [160, 161].

In conclusion, AI’s role in illuminating vessel dynamics, both pulmonary and extrapulmonary, revolutionizes our grasp of COPD’s vascular implications. This fusion of technology and radiology heralds a future where COPD management is not just reactive but proactive, underpinned by deep, data-driven insights.

Clinical applicability and Challenges in clinical implementation of AI for COPD

Clinical applicability of AI methods in COPD management

The clinical applicability of AI methods in COPD management is a critical consideration when assessing the real-world impact of these technologies. While numerous studies have demonstrated the promising performance of AI algorithms in various tasks, such as COPD identification and staging, emphysema region segmentation, and quantitative analysis, it is essential to evaluate their effectiveness and feasibility in actual clinical settings.

Several studies have shown that AI-assisted diagnosis can improve diagnostic accuracy and efficiency compared to traditional methods. For example, González et al. [66] demonstrated that their CNN-based approach achieved a high classification accuracy of 77.3% in identifying COPD patients from CT scans. Similarly, Xu et al. [70] reported an accuracy of 99.29% in identifying COPD using a modified AlexNet architecture. These results suggest that AI algorithms can potentially reduce misdiagnosis and improve the early detection of COPD, which is crucial for timely intervention and management.

Moreover, automatic measurements provided by AI algorithms can significantly enhance clinical workflow efficiency. Hasenstab et al. [74] developed an automated pipeline for quantifying emphysema and air trapping from CT scans, which showed strong correlations with manual measurements. Such automated tools can save substantial time for radiologists and pulmonologists, allowing them to focus on more complex cases and patient care.

However, it is important to acknowledge the limitations of current studies when considering their clinical applicability. Many of the cited works have relatively small sample sizes and lack long-term follow-up data, which may limit the generalizability of their findings. Supervised learning, particularly for segmentation tasks, necessitates accurately annotated images. Inconsistencies in labeling can mislead the model, compromising its efficacy. Deep learning models, especially, hunger for vast data volumes. Without enough examples, these models risk overfitting, limiting their generalizability. Balancing data access for model training and patient privacy is critical. Moreover, ethical considerations regarding data sourcing and utilization can’t be overlooked [162]. Furthermore, the integration of AI tools into existing clinical workflows and the interpretability of AI models remain significant challenges that need to be addressed for successful clinical adoption.

To fully assess the clinical applicability of AI methods in COPD management, future research should focus on conducting large-scale, prospective studies with diverse patient populations and long-term follow-up. Additionally, efforts should be made to develop more interpretable AI models and optimize their integration with clinical workflows. Real-world application cases and ongoing clinical trials, such as the COPDGene study [163], can provide valuable insights into the potential benefits and challenges of implementing AI tools in COPD care.

Challenges in clinical implementation of AI methods

Harnessing the power of AI in the arena of COPD imaging offers boundless opportunities, but it also comes with inherent challenges. These challenges include the interpretability and explainability of AI models, integration into existing clinical workflows, and regulatory issues.

Interpretability and explainability are crucial for building trust and confidence in AI-assisted decision-making. Many AI models, particularly deep learning algorithms, operate as “black boxes”, making it difficult for clinicians to understand how the model arrived at a particular decision [164]. This lack of transparency can hinder the adoption of AI tools in clinical practice. To address this issue, researchers are developing methods to enhance the interpretability of AI models, such as Grad-CAM [165] and attention rollout [166]. These approaches aim to provide insights into the features and reasoning behind AI predictions, enabling clinicians to validate and trust the results.

Integrating AI tools into existing clinical workflows is another significant challenge. AI algorithms should seamlessly fit into the daily routines of healthcare professionals without causing disruptions or increasing workload [167]. This requires close collaboration between AI developers, radiologists, and pulmonologists to design user-friendly interfaces and workflows that align with clinical needs. Furthermore, the outputs of AI tools should be presented in a clear and actionable manner, allowing clinicians to easily incorporate the information into their decision-making process.

Regulatory issues pose another hurdle in the clinical implementation of AI methods. AI algorithms used in healthcare are subject to strict regulatory requirements to ensure patient safety and data privacy [168]. Compliance with these regulations often involves extensive validation, documentation, and monitoring, which can be time-consuming and resource-intensive. Moreover, the rapidly evolving nature of AI technologies presents challenges for regulatory bodies to keep pace with the latest developments and establish appropriate guidelines.

To overcome these challenges, a multi-faceted approach is necessary. Researchers should prioritize the development of interpretable and explainable AI models, engaging clinicians in the process to ensure clinical relevance. Collaborative efforts between AI developers and healthcare professionals are essential for designing intuitive and efficient workflows that integrate AI tools seamlessly. Additionally, regulatory bodies need to adapt and evolve their guidelines to keep up with the advancements in AI technologies while maintaining patient safety and data privacy standards.

Future perspectives

Emerging innovations: potential future directions and innovations in AI for COPD imaging

Enhanced AI algorithms: As AI research progresses, there’s a potential to develop algorithms that can better detect early COPD changes or subtypes not identifiable with current techniques. This implies a move beyond simple identification towards nuanced understanding, such as differentiating various pathophysiological processes or predicting the likelihood of exacerbations based on subtle imaging features. Adaptive learning systems: The concept of AI systems that evolve with each scan, constantly learning and refining their diagnostic abilities, could revolutionize timely and accurate disease detection and staging. 3D imaging reconstructions: With advancements in imaging modalities and AI-driven reconstructions, there’s a possibility of generating dynamic 3D models of the lungs. These models could offer real-time insights into airway dynamics, vascular changes, and tissue alterations, providing a depth of understanding previously unattainable. Beyond imaging - integrated diagnostic platforms: The future might see platforms that merge imaging data with physiological metrics, blood biomarkers, and even genomics. Such an integrated approach would provide a multi-dimensional perspective of COPD, facilitating precision medicine endeavors.

Interdisciplinary collaborations: the fusion of AI with other scientific fields for comprehensive COPD management

AI and molecular biology: By integrating AI with molecular research, we could decipher intricate relationships between imaging phenotypes and molecular signatures. This could aid in identifying potential therapeutic targets or understanding the underpinnings of various COPD subtypes at a molecular level [169].

Neuroimaging and COPD: Emerging research hints at COPD’s neural implications. Collaborations between AI, pulmonology, and neuroimaging could uncover neural patterns associated with COPD, potentially opening avenues for novel interventions [170].

Environmental data and AI: By incorporating environmental data, AI models could predict COPD exacerbation risks based on localized air quality metrics, allergen levels, or other relevant factors. This integrative approach would encompass not just internal but also external factors influencing COPD dynamics [171].

Collaborations: The fusion of AI with other scientific fields for comprehensive COPD care. The potential of AI in enhancing the diagnostic and therapeutic landscape of COPD is undeniable. Yet, its full potential remains to be untapped. As we envisage the future, we recognize a trajectory marked by complex innovations and the melding of diverse disciplines to offer holistic and advanced COPD care.

Patient-centric platforms: The blend of AI with user experience (UX) design could lead to platforms that not only monitor patients but also educate and empower them. Such tools would make patients active participants in their care, promoting adherence and proactive health management [172].

Conclusively, the potential of AI in COPD care transcends mere imaging. Its nexus with diverse disciplines and the ensuing innovations could profoundly reshape COPD diagnostics, therapeutics, and patient engagement in the years to come.

Conclusion

The application of AI, specifically machine learning and deep learning techniques, has shown significant potential in transforming the diagnosis and management of COPD. These technologies are providing unprecedented insights into aspects of the disease such as emphysema, airway dynamics, and vascular structures, which are critical for a holistic understanding and treatment of COPD. Despite the challenges posed by the complex ’black-box’ nature of AI algorithms and the need for robust model training, the future holds promise. The emergence of innovations in AI for COPD imaging and the potential for cross-disciplinary collaborations indicate a future where AI is not just an aid, but a leader in significant advancements in COPD care. However, to fully leverage AI’s potential, it’s imperative to create rich, meticulously annotated datasets that will help develop reliable and generalizable AI models. By doing so, we can ensure that AI contributes more effectively to the refinement of COPD patient care.

Availability of data and materials

No datasets were generated or analysed during the current study.

Abbreviations

COPD:

Chronic obstructive pulmonary disease

AI:

Artificial intelligence

CT:

Computed tomography

ML:

Machine learning

DL:

Deep learning

CNN:

Convolutional neural network

GLCM:

Gray level co-occurrence matrix

GLSZM:

Gray level size zone matrix

GLRLM:

Gray level run length matrix

NGTDM:

Neighboring gray tone difference matrix

GLDM:

Gray level dependence matrix

SVM:

Support vector machine

MIL:

Multiple instance learning

HRCT:

High-resolution CT

CLE:

Centrilobular emphysema

PLE:

Panlobular emphysema

PSE:

Paraseptal emphysema

HU:

Hounsfield Unit

SR-Net:

Slice-recovery network

ViT:

Vision Transformer

BMI:

Body mass index

FEV1:

Forced expiratory volume in one second

FVC:

Forced vital capacity

GOLD:

Global initiative for chronic obstructive lung disease

SimGAN:

Simulated and unsupervised generative adversarial network

AV:

Artery-vein

TBV:

Total blood volume

BV5:

Total volume of vessels with a cross-sectional area smaller than 5 mm2

AUC:

Area under the curve

ACC:

Accuracy

DSC:

Dice similarity coefficient

SEN:

Sensitivity

SPE:

Specificity

FPR:

False positive rate

TPR:

True Positive Rate

CE:

Contrasted enhanced

NC:

Non-contrasted

References

  1. Agustí A, Celli BR, Criner GJ, Halpin D, Anzueto A, Barnes P, et al. Global initiative for chronic obstructive lung disease 2023 report: GOLD executive summary. Am J Respir Crit Care Med. 2023;207(7):819–37.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Washko GR, Coxson HO, O’Donnell DE, Aaron SD. CT imaging of chronic obstructive pulmonary disease: insights, disappointments, and promise. Lancet Respir Med. 2017;5(11):903–8.

    Article  PubMed  Google Scholar 

  3. Doke PP. Chronic respiratory diseases: a rapidly emerging public health menace. Indian J Public Health. 2023;67(2):192–6.

    Article  PubMed  Google Scholar 

  4. Momtazmanesh S, Moghaddam SS, Ghamari SH, Rad EM, Rezaei N, Shobeiri P, et al. Global burden of chronic respiratory diseases and risk factors, 1990–2019: an update from the Global Burden of Disease Study 2019. EClinicalMedicine. 2023;59:101936.

  5. Washko GR, Coxson HO, O"Donnell DE, Aaron SD. CT imaging of chronic obstructive pulmonary disease: insights, disappointments, and promise. Lancet Respir Med. 2017;5(11):903–8.

  6. Bhatt SP, Washko GR, Hoffman EA, Newell JD, Bodduluri S, Diaz AA, et al. Imaging Advances in Chronic Obstructive Pulmonary Disease: Insights from COPDGene. Am J Respir Crit Care Med. 2018;199(3):286–301.

    Article  Google Scholar 

  7. Venkatesan P. GOLD COPD report: 2023 update. Lancet Respir Med. 2023;11(1):18.

    Article  PubMed  Google Scholar 

  8. Bailey KL. The importance of the assessment of pulmonary function in COPD. Med Clin N Am. 2012;96(4):745–52.

    Article  PubMed  Google Scholar 

  9. Bodduluri S, Nakhmani A, Kandhare P, Patel T, Reinhardt J, Wilson C, et al. CT-Based Airway Surface Area to Volume Ratio Is Associated with Lung Function Decline in Chronic Obstructive Pulmonary Disease (COPD). In: A98. PHENOTYPING COPD AND PREDICTING THE DISEASE PROGRESSION. New York: American Thoracic Society; 2019. p. A2411.

  10. Lynch DA, Austin JH, Hogg JC, Grenier PA, Kauczor HU, Bankier AA, et al. CT-definable subtypes of chronic obstructive pulmonary disease: a statement of the Fleischner Society. Radiology. 2015;277(1):192–205.

    Article  PubMed  Google Scholar 

  11. Keogh E, Williams EM. Managing malnutrition in COPD: a review. Respir Med. 2021;176:106248.

    Article  PubMed  Google Scholar 

  12. Gawlitza J, Sturm T, Spohrer K, Henzler T, Akin I, Schönberg S, et al. Predicting pulmonary function testing from quantified computed tomography using machine learning algorithms in patients with COPD. Diagnostics. 2019;9(1):33.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Konietzke P, Wielpütz MO, Wagner WL, Wuennemann F, Kauczor HU, Heussel CP, et al. Quantitative CT detects progression in COPD patients with severe emphysema in a 3-month interval. Eur Radiol. 2020;30:2502–12.

    Article  PubMed  Google Scholar 

  14. Bodduluri S, Reinhardt JM, Hoffman EA, Newell JD, Bhatt SP. Recent Advances in Computed Tomography Imaging in Chronic Obstructive Pulmonary Disease. Ann Am Thorac Soc. 2018;15(3):281–9.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Kauczor HU, Wielpütz MO, Jobst BJ, Weinheimer O, Gompelmann D, Herth FJ, et al. Computed tomography imaging for novel therapies of chronic obstructive pulmonary disease. J Thorac Imaging. 2019;34(3):202–13.

    Article  PubMed  Google Scholar 

  16. Ma J, Song Y, Tian X, Hua Y, Zhang R, Wu J. Survey on deep learning for pulmonary medical imaging. Front Med. 2019;14(4):1–20.

  17. Chassagnon G, Vakalopolou M, Paragios N, Revel MP. Deep learning: definition and perspectives for thoracic imaging. Eur Radiol. 2019;30:1–10.

  18. Williams LH, Drew T. What do we know about volumetric medical image interpretation?: A review of the basic science and medical image perception literatures. Cogn Res Princ Implications. 2019;4:1–24.

    Google Scholar 

  19. Kumar R, Wang W, Kumar J, Yang T, Khan A, Ali W, et al. An integration of blockchain and AI for secure data sharing and detection of CT images for the hospitals. Comput Med Imaging Graph. 2021;87:101812.

    Article  PubMed  Google Scholar 

  20. Castiglioni I, Rundo L, Codari M, Di Leo G, Salvatore C, Interlenghi M, et al. AI applications to medical images: From machine learning to deep learning. Phys Med. 2021;83:9–24.

    Article  PubMed  Google Scholar 

  21. Fourcade A, Khonsari RH. Deep learning in medical image analysis: A third eye for doctors. J Stomatology Oral Maxillofac Surg. 2019;120(4):279–88.

    Article  CAS  Google Scholar 

  22. Myszczynska MA, Ojamies PN, Lacoste AM, Neil D, Saffari A, Mead R, et al. Applications of machine learning to diagnosis and treatment of neurodegenerative diseases. Nat Rev Neurol. 2020;16(8):440–56.

    Article  PubMed  Google Scholar 

  23. Chang R, Qi S, Wu Y, Yue Y, Zhang X, Guan Y, et al. Deep radiomic model based on the sphere-shell partition for predicting treatment response to chemotherapy in lung cancer. Transl Oncol. 2023;35:101719.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  24. Qi S, Zuo Y, Chang R, Huang K, Liu J, Zhang Z. Using CT radiomic features based on machine learning models to subtype adrenal adenoma. BMC Cancer. 2023;23(1):1–12.

    Article  Google Scholar 

  25. Chang R, Qi S, Zuo Y, Yue Y, Zhang X, Guan Y, et al. Predicting chemotherapy response in non-small-cell lung cancer via computed tomography radiomic features: Peritumoral, intratumoral, or combined? Front Oncol. 2022;12:915835.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Xu M, Qi S, Yue Y, Teng Y, Xu L, Yao Y, et al. Segmentation of lung parenchyma in CT images using CNN trained with the clustering algorithm generated dataset. Biomed Eng Online. 2019;18:1–21.

    Article  Google Scholar 

  27. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44.

    Article  CAS  PubMed  Google Scholar 

  28. Srivastava A, Jain S, Miranda R, Patil S, Pandya S, Kotecha K. Deep learning based respiratory sound analysis for detection of chronic obstructive pulmonary disease. PeerJ Comput Sci. 2021;7:e369.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Rusk N. Deep learning. Nat Methods. 2016;13(1):35.

    Article  CAS  Google Scholar 

  30. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway: IEEE; 2016. pp. 770–8.

  31. Wu S, Roberts K, Datta S, Du J, Ji Z, Si Y, et al. Deep learning in clinical natural language processing: a methodical review. J Am Med Inform Assoc. 2020;27(3):457–70.

    Article  PubMed  Google Scholar 

  32. Wu Y, Qi S, Feng J, Chang R, Pang H, Hou J, et al. Attention-guided multiple instance learning for COPD identification: To combine the intensity and morphology. Biocybernetics Biomed Eng. 2023;43(3):568–85.

    Article  Google Scholar 

  33. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst. 2012;25.

  34. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. 2017. [Online] Available: https://arxiv.org/abs/1704.04861.

  35. Lakhani P, Sundaram B. Deep learning at chest radiography: automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology. 2017;284(2):574–82.

    Article  PubMed  Google Scholar 

  36. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. [Online]. Available: https://arxiv.org/abs/1409.1556.

  37. Wu Y, Du R, Feng J, Qi S, Pang H, Xia S, et al. Deep CNN for COPD identification by Multi-View snapshot integration of 3D airway tree and lung field. Biomed Signal Process Control. 2023;79:104162.

    Article  Google Scholar 

  38. Sun J, Liao X, Yan Y, Zhang X, Sun J, Tan W, et al. Detection and staging of chronic obstructive pulmonary disease using a computed tomography-based weakly supervised deep learning approach. Eur Radiol. 2022;32(8):5319–29.

    Article  PubMed  Google Scholar 

  39. Chen S, Duan J, Wang H, Wang R, Li J, Qi M, et al. Automatic detection of stroke lesion from diffusion-weighted imaging via the improved YOLOv5. Comput Biol Med. 2022;150:106120.

    Article  PubMed  Google Scholar 

  40. Chen S, Duan J, Zhang N, Qi M, Li J, Wang H, et al. MSA-YOLOv5: Multi-scale attention-based YOLOv5 for automatic detection of acute ischemic stroke from multi-modality MRI images. Comput Biol Med. 2023;165:107471.

    Article  PubMed  Google Scholar 

  41. Campo MI, Pascau J, Estépar RSJ. Emphysema quantification on simulated X-rays through deep learning techniques. In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018). Piscataway: IEEE; 2018. pp. 273–6.

  42. Resten A, Maitre S, Musset D. CT imaging of peripheral pulmonary vessel disease. Eur Radiol. 2005;15:2045–56.

    Article  PubMed  Google Scholar 

  43. Pang H, Wu Y, Qi S, Li C, Shen J, Yue Y, et al. A fully automatic segmentation pipeline of pulmonary lobes before and after lobectomy from computed tomography images. Comput Biol Med. 2022;147:105792.

    Article  PubMed  Google Scholar 

  44. Qi Q, Qi S, Wu Y, Li C, Tian B, Xia S, et al. Fully automatic pipeline of convolutional neural networks and capsule networks to distinguish COVID-19 from community-acquired pneumonia via CT images. Comput Biol Med. 2022;141:105182.

    Article  CAS  PubMed  Google Scholar 

  45. Greffier J, Hamard A, Pereira F, Barrau C, Pasquier H, Beregi JP, et al. Image quality and dose reduction opportunity of deep learning image reconstruction algorithm for CT: a phantom study. Eur Radiol. 2020;30:3951–9.

    Article  PubMed  Google Scholar 

  46. Koetzier LR, Mastrodicasa D, Szczykutowicz TP, van der Werf NR, Wang AS, Sandfort V, et al. Deep learning image reconstruction for CT: technical principles and clinical prospects. Radiology. 2023;306(3):e221257.

    Article  PubMed  Google Scholar 

  47. Dong Y, Feng Q, Yang W, Lu Z, Deng C, Zhang L, et al. Preoperative prediction of sentinel lymph node metastasis in breast cancer based on radiomics of T2-weighted fat-suppression and diffusion-weighted MRI. Eur Radiol. 2018;28(2):582–91.

    Article  PubMed  Google Scholar 

  48. Liu MZ, Mutasa S, Chang P, Siddique M, Jambawalikar S, Ha R. A novel CNN algorithm for pathological complete response prediction using an I-SPY TRIAL breast MRI database. Magn Reson Imaging. 2020;73:148–51.

    Article  PubMed  PubMed Central  Google Scholar 

  49. Liu Y, Chen A, Shi H, Huang S, Zheng W, Liu Z, et al. CT synthesis from MRI using multi-cycle GAN for head-and-neck radiation therapy. Comput Med Imaging Graph. 2021;91:101953.

    Article  PubMed  Google Scholar 

  50. Lei Y, Harms J, Wang T, Liu Y, Shu HK, Jani AB, et al. MRI-only based synthetic CT generation using dense cycle consistent generative adversarial networks. Med Phys. 2019;46(8):3565–81.

    Article  PubMed  PubMed Central  Google Scholar 

  51. Wu H, Jiang X, Jia F. UC-GAN for MR to CT image synthesis. In: Artificial Intelligence in Radiation Therapy: First International Workshop, AIRT 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 17, 2019, Proceedings 1. New York: Springer; 2019. pp. 146–53.

  52. Wu Y, Qi S, Sun Y, Xia S, Yao Y, Qian W. A vision transformer for emphysema classification using CT images. Phys Med Biol. 2021;66(24):245016.

    Article  Google Scholar 

  53. Pang H, Qi S, Wu Y, Wang M, Li C, Sun Y, et al. NCCT-CECT image synthesizers and their application to pulmonary vessel segmentation. Comput Methods Prog Biomed. 2023;231:107389.

    Article  Google Scholar 

  54. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In: Proceedings of the International Conference on Learning Representations (ICLR); 2021. [Online]. Available: https://openreview.net/forum?id=YicbFdNTTy.

  55. Carbonneau MA, Cheplygina V, Granger E, Gagnon G. Multiple instance learning: A survey of problem characteristics and applications. Pattern Recogn. 2018;77:329–53.

    Article  Google Scholar 

  56. Nikolaou V, Massaro S, Fakhimi M, Stergioulas L, Price D. COPD phenotypes and machine learning cluster analysis: A systematic review and future research agenda. Respir Med. 2020;106093.

  57. Exarchos KP, Aggelopoulou A, Oikonomou A, Biniskou T, Beli V, Antoniadou E, et al. Review of artificial intelligence techniques in chronic obstructive lung disease. IEEE J Biomed Health Inf. 2021;26(5):2331–8.

    Article  Google Scholar 

  58. Estépar RSJ. Artificial intelligence in COPD: New venues to study a complex disease. Barcelona Respir Netw Rev. 2020;6(2):144.

    Google Scholar 

  59. Wu Y, Zhao S, Qi S, Feng J, Pang H, Chang R, et al. Two-stage contextual transformer-based convolutional neural network for airway extraction from CT images. Artif Intell Med. 2023;143:102637.

    Article  PubMed  Google Scholar 

  60. Wu Y, Qi S, Wang M, Zhao S, Pang H, Xu J, et al. Transformer-based 3D U-Net for pulmonary vessel segmentation and artery-vein separation from CT images. Med Biol Eng Comput. 2023;61(10):2649–63.

    Article  PubMed  Google Scholar 

  61. Madani A, De Maertelaer V, Zanen J, Gevenois PA. Pulmonary emphysema: radiation dose and section thickness at multidetector CT quantification-comparison with macroscopic and microscopic morphometry. Radiology. 2007;243(1):250–7.

    Article  PubMed  Google Scholar 

  62. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology. 2016;278(2):563–77.

    Article  PubMed  Google Scholar 

  63. Yang K, Yang Y, Kang Y, Liang Z, Wang F, Li Q, et al. The value of radiomic features in chronic obstructive pulmonary disease assessment: a prospective study. Clin Radiol. 2022;77(6):e466–72.

    Article  CAS  PubMed  Google Scholar 

  64. Mets OM, Schmidt M, Buckens CF, Gondrie MJ, Isgum I, Oudkerk M, et al. Diagnosis of chronic obstructive pulmonary disease in lung cancer screening computed tomography scans: independent contribution of emphysema, air trapping and bronchial wall thickening. Respir Res. 2013;14:1–8.

    Article  Google Scholar 

  65. Ying J, Dutta J, Guo N, Xia L, Sitek A, Li Q. Gold classification of COPDGene cohort based on deep learning. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Piscataway: IEEE; 2016. pp. 2474–8.

  66. González G, Ash SY, Vegas-Sánchez-Ferrero G, Onieva Onieva J, Rahaghi FN, Ross JC, et al. Disease staging and prognosis in smokers using deep learning in chest computed tomography. Am J Respir Crit Care Med. 2018;197(2):193–203.

    Article  PubMed  PubMed Central  Google Scholar 

  67. Cheplygina V, Pena IP, Pedersen JH, Lynch DA, Sørensen L, de Bruijne M. Transfer learning for multicenter classification of chronic obstructive pulmonary disease. IEEE journal of biomedical and health informatics. 2017;22(5):1486–96.

    Article  PubMed  PubMed Central  Google Scholar 

  68. Sathiya S, Priyanka G, Jeyanthi S. Detection of chronic obstructive pulmonary disease in computer aided diagnosis system with CNN classification. Int J Pure Appl Math. 2018;119(12):13815–1382.

    Google Scholar 

  69. Singla S, Gong M, Riley C, Sciurba F, Batmanghelich K. Improving clinical disease subtyping and future events prediction through a chest CT-based deep learning approach. Med Phys. 2021;48(3):1168–81.

    Article  PubMed  Google Scholar 

  70. Xu C, Qi S, Feng J, Xia S, Kang Y, Yao Y, et al. DCT-MIL: deep CNN transferred multiple instance learning for COPD identification using CT images. Phys Med Biol. 2020;65(14):145011.

    Article  PubMed  Google Scholar 

  71. Du R, Qi S, Feng J, Xia S, Kang Y, Qian W, et al. Identification of COPD From Multi-View Snapshots of 3D Lung Airway Tree via Deep CNN. IEEE Access. 2020;8:38907–19.

    Article  Google Scholar 

  72. Tang LY, Coxson HO, Lam S, Leipsic J, Tam RC, Sin DD. Towards large-scale case-finding: training and validation of residual networks for detection of chronic obstructive pulmonary disease using low-dose CT. Lancet Digit Health. 2020;2(5):e259–67.

    Article  PubMed  Google Scholar 

  73. Ho TT, Kim T, Kim WJ, Lee CH, Chae KJ, Bak SH, et al. A 3D-CNN model with CT-based parametric response mapping for classifying COPD subjects. Sci Rep. 2021;11(1):34.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  74. Hasenstab KA, Yuan N, Retson T, Conrad DJ, Kligerman S, Lynch DA, et al. Automated CT staging of chronic obstructive pulmonary disease severity for predicting disease progression and mortality with a deep learning convolutional neural network. Radiol Cardiothorac Imaging. 2021;3(2):e200477.

    Article  PubMed  PubMed Central  Google Scholar 

  75. Chen L, Feng Q, Yin X, Min X, Shi L, Yang D, et al. A Graph Convolutional Multiple Instance Learning on a Hypersphere Manifold Approach for Diagnosing Chronic Obstructive Pulmonary Disease in CT Images. IEEE J Biomed Health Informa. 2022;26(12):6058–69.

    Article  Google Scholar 

  76. Li Z, Liu L, Zhang Z, Yang X, Li X, Gao Y, et al. A novel CT-based radiomics features analysis for identification and severity staging of COPD. Acad Radiol. 2022;29(5):663–73.

    Article  PubMed  Google Scholar 

  77. Sun P, Wang D, Mok VC, Shi L. Comparison of feature selection methods and machine learning classifiers for radiomics analysis in glioma grading. IEEE Access. 2019;7:102010–20.

    Article  Google Scholar 

  78. Zhang L, Jiang B, Wisselink HJ, Vliegenthart R, Xie X. COPD identification and grading based on deep learning of lung parenchyma and bronchial wall in chest CT images. Br J Radiol. 2022;95(1133):20210637.

    Article  PubMed  PubMed Central  Google Scholar 

  79. Li Z, Huang K, Liu L, Zhang Z. Early detection of COPD based on graph convolutional network and small and weakly labeled data. Med Biol Eng Comput. 2022;60(8):2321–33.

    Article  PubMed  PubMed Central  Google Scholar 

  80. Yang Y, Li W, Guo Y, Zeng N, Wang S, Chen Z, et al. Lung radiomics features for characterizing and classifying COPD stage based on feature combination strategy and multi-layer perceptron classifier. Math Biosci Eng. 2022;19(8):7826–55.

    Article  PubMed  Google Scholar 

  81. Yang Y, Wang S, Zeng N, Duan W, Chen Z, Liu Y, et al. Lung radiomics features selection for COPD stage classification based on auto-metric graph neural network. Diagnostics. 2022;12(10):2274.

    Article  PubMed  PubMed Central  Google Scholar 

  82. Almeida SD, Lüth CT, Norajitra T, Wald T, Nolden M, Jäger PF, et al. cOOpD: Reformulating COPD classification on chest CT scans as anomaly detection using contrastive representations. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. New York: Springer; 2023. pp. 33–43.

  83. Karch A, Vogelmeier C, Welte T, Bals R, Kauczor HU, Biederer J, et al. The German COPD cohort COSYCONET: aims, methods and descriptive analysis of the study population at baseline. Respir Med. 2016;114:27–37.

    Article  PubMed  Google Scholar 

  84. Xue M, Jia S, Chen L, Huang H, Yu L, Zhu W. CT-based COPD identification using multiple instance learning with two-stage attention. Comput Methods Prog Biomed. 2023;230:107356.

    Article  Google Scholar 

  85. Zhou HY, Yu Y, Wang C, Zhang S, Gao Y, Pan J, et al. A transformer-based representation-learning model with unified processing of multimodal input for clinical diagnostics. Nat Biomed Eng. 2023;7(6):743–55.

  86. Amudala Puchakayala PR, Sthanam VL, Nakhmani A, Chaudhary MF, Kizhakke Puliyakote A, Reinhardt JM, et al. Radiomics for Improved Detection of Chronic Obstructive Pulmonary Disease in Low-Dose and Standard-Dose Chest CT Scans. Radiology. 2023;307(5):e222998.

    Article  PubMed  Google Scholar 

  87. Yu K, Sun L, Chen J, Reynolds M, Chaudhary T, Batmanghelich K. DrasCLR: A Self-supervised Framework of Learning Disease-related and Anatomy-specific Representation for 3D Medical Images. Med Image Anal. 2024;92:103062.

  88. Martini K, Frauenfelder T. Advances in imaging for lung emphysema. Ann Transl Med. 2020;8(21):1467.

  89. Coxson HO, Rogers RM, Whittall KP, D’YACHKOVA Y, Pare PD, Sciurba FC, et al. A quantification of the lung surface area in emphysema using computed tomography. Am J Respir Crit Care Med. 1999;159(3):851–6.

    Article  CAS  PubMed  Google Scholar 

  90. Gangeh MJ, Sørensen L, Shaker SB, Kamel MS, De Bruijne M, Loog M. A texton-based approach for the classification of lung parenchyma in CT images. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2010: 13th International Conference, Beijing, China, September 20-24, 2010, Proceedings, Part III 13. New York: Springer; 2010. pp. 595–602.

  91. Sorensen L, Shaker SB, De Bruijne M. Quantitative analysis of pulmonary emphysema using local binary patterns. IEEE Trans Med Imaging. 2010;29(2):559–69.

    Article  PubMed  Google Scholar 

  92. Zulueta-Coarasa T, Kurugol S, Ross JC, Washko GG, Estépar RSJ. Emphysema classification based on embedded probabilistic PCA. In: 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). Piscataway: IEEE; 2013. pp. 3969–72.

  93. Karabulut EM, Ibrikci T. Emphysema discrimination from raw HRCT images by convolutional neural networks. In: 2015 9th International Conference on Electrical and Electronics Engineering (ELECO). Piscataway: IEEE; 2015. pp. 705–8.

  94. Peng L, Lin L, Hu H, Ling X, Wang D, Han X, et al. Joint weber-based rotation invariant uniform local ternary pattern for classification of pulmonary emphysema in CT images. In: 2017 IEEE International Conference on Image Processing (ICIP). Piscataway: IEEE; 2017. pp. 2050–4.

  95. Bortsova G, Dubost F, Ørting S, Katramados I, Hogeweg L, Thomsen L, et al. Deep learning from label proportions for emphysema quantification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. New York: Springer; 2018. pp. 768–76.

  96. Bermejo-Peláez D, Estepar RSJ, Ledesma-Carbayo MJ. Emphysema classification using a multi-view convolutional network. In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018). Piscataway: IEEE; 2018. pp. 519–22.

  97. Bermejo-Peláez D, Okajima Y, Washko GR, Ledesma-Carbayo MJ, Estépar RSJ. A SR-NET 3D-to-2D Architecture For Paraseptal Emphysema Segmentation. In: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019). Piscataway: IEEE; 2019. pp. 303–6.

  98. Peng L, Lin L, Hu H, Li H, Chen Q, Ling X, et al. Classification and quantification of emphysema using a multi-scale residual network. IEEE J Biomed Health Inform. 2019;23(6):2526–36.

    Article  PubMed  Google Scholar 

  99. Peng L, Lin L, Hu H, Zhang Y, Li H, Iwamoto Y, et al. Semi-Supervised Learning for Semantic Segmentation of Emphysema With Partial Annotations. IEEE J Biomed Health Inform. 2019;24(8):2327–36.

    Article  PubMed  Google Scholar 

  100. Li H, Mukundan R. Robust Texture Features For Emphysema Classification In CT Images. In: 2020 28th European Signal Processing Conference (EUSIPCO). Piscataway: IEEE; 2020. pp. 1220–4.

  101. Ørting SN, Petersen J, Thomsen LH, Wille MM, De Bruijne M. Detecting emphysema with multiple instance learning. In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018). Piscataway: IEEE; 2018. pp. 510–3.

  102. Humphries SM, Notary AM, Centeno JP, Strand MJ, Crapo JD, Silverman EK, et al. Deep learning enables automatic classification of emphysema pattern at CT. Radiology. 2020;294(2):434–44.

    Article  PubMed  Google Scholar 

  103. Mondal S, Sadhu AK, Dutta PK. Automated diagnosis of pulmonary emphysema using multi-objective binary thresholding and hybrid classification. Biomed Signal Process Control. 2021;69:102886.

    Article  Google Scholar 

  104. Nava R, Olveres J, Kybic J, Escalante B, Cristóbal G. Feature ensemble for quantitative analysis of emphysema in CT imaging. In: 2015 E-Health and Bioengineering Conference (EHB). IEEE. pp. 1–4.

  105. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2016. pp. 2818–26.

  106. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC. Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2018. pp. 4510–20.

  107. Yang J, Angelini ED, Balte PP, Hoffman EA, Austin JH, Smith BM, et al. Novel subtypes of pulmonary emphysema based on spatially-informed lung texture learning: the multi-ethnic study of atherosclerosis (MESA) COPD Study. IEEE Trans Med Imaging. 2021;40(12):3652–62.

    Article  PubMed  PubMed Central  Google Scholar 

  108. Angelini E, Sun Y, Yang J, Balte P, Hoffman E, Allen N, et al. Longitudinal Progression of Machine-Learned Quantitative Emphysema Subtypes. The Mesa COPD Study. In: D28. COPD: FROM PATHOGENESIS TO PROGNOSIS. American Thoracic Society; 2020. p. A6420.

  109. Hackx M, Bankier AA, Gevenois PA. Chronic obstructive pulmonary disease: CT quantification of airways disease. Radiology. 2012;265(1):34–48.

    Article  PubMed  Google Scholar 

  110. Jin D, Harrison AP, Zhang L, Yan K, Wang Y, Cai J, et al. Artificial intelligence in radiology. In: Artificial Intelligence in Medicine. New York: Elsevier; 2021. pp. 265–89.

  111. Saha PK, Nadeem SA, Comellas AP. A survey on artificial intelligence in pulmonary imaging. Wiley Interdiscip Rev Data Min Knowl Disc. 2023;13(6):e1510.

    Article  Google Scholar 

  112. Little S, Sproule M, Cowan M, Macleod K, Robertson M, Love J, et al. High resolution computed tomographic assessment of airway wall thickness in chronic asthma: reproducibility and relationship with lung function and severity. Thorax. 2002;57(3):247–53.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  113. Kocak B, Ates E, Durmaz ES, Ulusan MB, Kilickesmez O. Influence of segmentation margin on machine learning-based high-dimensional quantitative CT texture analysis: a reproducibility study on renal clear cell carcinomas. Eur Radiol. 2019;29:4765–75.

    Article  PubMed  Google Scholar 

  114. Li H, Tang Z, Nan Y, Yang G. Human treelike tubular structure segmentation: A comprehensive review and future perspectives. Comput Biol Med. 2022;151:106241.

  115. Cheng G, Wu X, Xiang W, Guo C, Ji H, He L. Segmentation of the airway tree from chest CT using tiny atrous convolutional network. IEEE Access. 2021;9:33583–94.

    Article  Google Scholar 

  116. Zheng H, Qin Y, Gu Y, Xie F, Yang J, Sun J, et al. Alleviating class-wise gradient imbalance for pulmonary airway segmentation. IEEE Trans Med Imaging. 2021;40(9):2452–62.

    Article  PubMed  Google Scholar 

  117. Wang A, Tam TCC, Poon HM, Yu KC, Lee WN. Naviairway: a bronchiole-sensitive deep learning-based airway segmentation pipeline for planning of navigation bronchoscopy. 2022. [Online]. Available: https://arxiv.org/abs/2203.04294.

  118. Zhao M, Zhao S, Quan Q, Fan L, Qiu X, Liu S, et al. GDDS: Pulmonary Bronchioles Segmentation with Group Deep Dense Supervision. 2023. [Online]. Available: https://arxiv.org/abs/2303.09212.

  119. Nan Y, Del Ser J, Tang Z, Tang P, Xing X, Fang Y, et al. Fuzzy attention neural network to tackle discontinuity in airway segmentation. IEEE Trans Neural Netw Learn Syst. 2023;35(6):7391–404.

  120. Huang W, Gong H, Zhang H, Wang Y, Li H, Li G, et al. BronchusNet: Region and Structure Prior Embedded Representation Learning for Bronchus Segmentation and Classification. 2022. [Online]. Available: https://arxiv.org/abs/2303.09212.

  121. Charbonnier JP, Van Rikxoort EM, Setio AA, Schaefer-Prokop CM, van Ginneken B, Ciompi F. Improving airway segmentation in computed tomography using leak detection with convolutional networks. Med Image Anal. 2017;36:52–60.

    Article  PubMed  Google Scholar 

  122. Yun J, Park J, Yu D, Yi J, Lee M, Park HJ, et al. Improvement of fully automated airway segmentation on volumetric computed tomographic images using a 2.5 dimensional convolutional neural net. Med Image Anal. 2019;51:13–20.

    Article  PubMed  Google Scholar 

  123. Carmo D, Rittner L, Lotufo R. Open-source tool for Airway Segmentation in Computed Tomography using 2.5 D Modified EfficientDet: Contribution to the ATM22 Challenge. 2022. [Online]. Available: https://arxiv.org/abs/2209.15094.

  124. Qin Y, Chen M, Zheng H, Gu Y, Shen M, Yang J, et al. Airwaynet: a voxel-connectivity aware approach for accurate airway segmentation using convolutional neural networks. In: International conference on medical image computing and computer-assisted intervention. New York: Springer; 2019. pp. 212–20.

  125. Nadeem SA, Hoffman EA, Sieren JC, Comellas AP, Bhatt SP, Barjaktarevic IZ, et al. A CT-based automated algorithm for airway segmentation using freeze-and-grow propagation and deep learning. IEEE Trans Med Imaging. 2020;40(1):405–18.

    Article  PubMed  Google Scholar 

  126. Selvan R, Kipf T, Welling M, Juarez AGU, Pedersen JH, Petersen J, et al. Graph refinement based airway extraction using mean-field networks and graph neural networks. Med Image Anal. 2020;64:101751.

    Article  PubMed  Google Scholar 

  127. Xie W, Jacobs C, Charbonnier JP, van Ginneken B. Structure and position-aware graph neural network for airway labeling. 2022. [Online]. Available: https://arxiv.org/abs/2201.04532.

  128. Guo J, Fu R, Pan L, Zheng S, Huang L, Zheng B, et al. Coarse-to-fine airway segmentation using multi information fusion network and CNN-based region growing. Comput Methods Prog Biomed. 2022;215:106610.

    Article  Google Scholar 

  129. Heitz A, Weinzorn J, Noblet V, Naegel B, Charnoz A, Heitz F, et al. Lubrav: a new framework for the segmentation of the lung’s tubular structures. In: 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI). Piscataway: IEEE; 2021. pp. 948–52.

  130. Qin Y, Zheng H, Gu Y, Huang X, Yang J, Wang L, et al. Learning tubule-sensitive CNNs for pulmonary airway and artery-vein segmentation in CT. IEEE Trans Med Imaging. 2021;40(6):1603–17.

    Article  PubMed  Google Scholar 

  131. Chen S, Garcia-Uceda A, Su J, van Tulder G, Wolff L, van Walsum T, et al. Label refinement network from synthetic error augmentation for medical image segmentation. 2022. [Online]. Available: https://arxiv.org/abs/2209.06353.

  132. Wang P, Guo D, Zheng D, Zhang M, Yu H, Sun X, et al. Accurate Airway Tree Segmentation in CT Scans via Anatomy-aware Multi-class Segmentation and Topology-guided Iterative Learning. IEEE Trans Med Imaging. 2024. https://doi.org/10.1109/TMI.2024.3419707.

  133. Wu Y, Zhang M, Yu W, Zheng H, Xu J, Gu Y. LTSP: long-term slice propagation for accurate airway segmentation. Int J CARS. 2022;17(5):857–65.

    Article  Google Scholar 

  134. Nardelli P, Ross JC, Estépar RSJ. Generative-based airway and vessel morphology quantification on chest CT images. Med Image Anal. 2020;63:101691.

    Article  PubMed  PubMed Central  Google Scholar 

  135. Park J, Hobbs BD, Crapo JD, Make BJ, Regan EA, Humphries S, et al. Subtyping COPD by using visual and quantitative CT imaging features. Chest. 2020;157(1):47–60.

    Article  PubMed  Google Scholar 

  136. Castaldi PJ, Boueiz A, Yun J, Estepar RSJ, Ross JC, Washko G, et al. Machine learning characterization of COPD subtypes: insights from the COPDGene study. Chest. 2020;157(5):1147–57.

    Article  CAS  PubMed  Google Scholar 

  137. Tanabe N, Shimizu K, Terada K, Sato S, Suzuki M, Shima H, et al. Central airway and peripheral lung structures in airway disease-dominant COPD. ERJ Open Res. 2021;7(1):00672–2020.

  138. Peng J, Chen C, Zhou M, Xie X, Zhou Y, Luo CH. A machine-learning approach to forecast aggravation risk in patients with acute exacerbation of chronic obstructive pulmonary disease with clinical indicators. Sci Rep. 2020;10(1):3118.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  139. Lanclus M, Clukers J, Van Holsbeke C, Vos W, Leemans G, Holbrechts B, et al. Machine learning algorithms utilizing functional respiratory imaging may predict COPD exacerbations. Acad Radiol. 2019;26(9):1191–9.

    Article  PubMed  Google Scholar 

  140. Zeng S, Arjomandi M, Tong Y, Liao ZC, Luo G. Developing a machine learning model to predict severe chronic obstructive pulmonary disease exacerbations: retrospective cohort study. J Med Internet Res. 2022;24(1):e28953.

    Article  PubMed  PubMed Central  Google Scholar 

  141. Young AL, Bragman FJ, Rangelov B, Han MK, Galbán CJ, Lynch DA, et al. Disease progression modeling in chronic obstructive pulmonary disease. Am J Respir Crit Care Med. 2020;201(3):294–302.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  142. Lin A, Kolossváry M, Motwani M, Išgum I, Maurovich-Horvat P, Slomka PJ, et al. Artificial intelligence in cardiovascular CT: Current status and future implications. J Cardiovasc Comput Tomogr. 2021;15(6):462–9.

    Article  PubMed  PubMed Central  Google Scholar 

  143. Nardelli P, Jimenez-Carretero D, Bermejo-Pelaez D, Washko GR, Rahaghi FN, Ledesma-Carbayo MJ, et al. Pulmonary artery-vein classification in CT images using deep learning. IEEE Trans Med Imaging. 2018;37(11):2428–40.

    Article  PubMed  PubMed Central  Google Scholar 

  144. Xu Y, Mao Z, Liu C, Wang B. Pulmonary vessel segmentation via stage-wise convolutional networks with orientation-based region growing optimization. IEEE Access. 2018;6:71296–305.

    Article  Google Scholar 

  145. Gu X, Wang J, Zhao J, Li Q. Segmentation and suppression of pulmonary vessels in low-dose chest CT scans. Med Phys. 2019;46(8):3603–14.

    Article  PubMed  Google Scholar 

  146. Guo W, Gu X, Fang Q, Li Q. Comparison of performances of conventional and deep learning-based methods in segmentation of lung vessels and registration of chest radiographs. Radiol Phys Technol. 2021;14:6–15.

    Article  PubMed  Google Scholar 

  147. Tan W, Zhou L, Li X, Yang X, Chen Y, Yang J. Automated vessel segmentation in lung CT and CTA images via deep neural networks. J X-ray Sci Technol. 2021;29(6):1123–37.

    Google Scholar 

  148. Jimenez-Carretero D, Bermejo-Peláez D, Nardelli P, Fraga P, Fraile E, Estépar RSJ, et al. A graph-cut approach for pulmonary artery-vein segmentation in noncontrast CT images. Med Image Anal. 2019;52:144–59.

    Article  PubMed  Google Scholar 

  149. Cui H, Liu X, Huang N. Pulmonary vessel segmentation based on orthogonal fused U-Net++ of chest CT images. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part VI 22. New York: Springer; 2019. pp. 293–300.

  150. Nam JG, Witanto JN, Park SJ, Yoo SJ, Goo JM, Yoon SH. Automatic pulmonary vessel segmentation on noncontrast chest CT: deep learning algorithm developed using spatiotemporally matched virtual noncontrast images and low-keV contrast-enhanced vessel maps. Eur Radiol. 2021;31(12):9012–21.

    Article  PubMed  PubMed Central  Google Scholar 

  151. Wu R, Xin Y, Qian J, Dong Y. A multi-scale interactive U-Net for pulmonary vessel segmentation method based on transfer learning. Biomed Signal Process Control. 2023;80:104407.

    Article  Google Scholar 

  152. Li X, Bala R, Monga V. Robust deep 3D blood vessel segmentation using structural priors. IEEE Trans Image Process. 2022;31:1271–84.

    Article  PubMed  Google Scholar 

  153. Wang M, Qi S, Wu Y, Sun Y, Chang R, Pang H, et al. CE-NC-VesselSegNet: Supervised by contrast-enhanced CT images but utilized to segment pulmonary vessels from non-contrast-enhanced CT images. Biomed Signal Process Control. 2023;82:104565.

    Article  Google Scholar 

  154. Pan L, Li Z, Shen Z, Liu Z, Huang L, Yang M, et al. Learning multi-view and centerline topology connectivity information for pulmonary artery-vein separation. Comput Biol Med. 2023;155:106669.

    Article  PubMed  Google Scholar 

  155. Litchman T, de Margerie-Mellon C, Bankier A, Washko G, Rahaghi F, San Jose Estepar R, et al. Pulmonary Vascular Pruning on Computed Tomography and Risk of Progression of Lung Adenocarcinoma. In: D20. NOT YOUR GRANDFATHER’S CT SCAN: RADIOMICS AND DEEP MACHINE LEARNING FOR PULMONARY NODULES AND LUNG CANCER EVALUATION. American Thoracic Society; 2022. p. A5074.

  156. Patel AR, Hurst JR. Extrapulmonary comorbidities in chronic obstructive pulmonary disease: state of the art. Expert Rev Respir Med. 2011;5(5):647–62.

    Article  PubMed  Google Scholar 

  157. Machado A, Marques A, Burtin C. Extra-pulmonary manifestations of COPD and the role of pulmonary rehabilitation: a symptom-centered approach. Expert Rev Respir Med. 2021;15(1):131–42.

    Article  CAS  PubMed  Google Scholar 

  158. Chattopadhyay S, Das R. Statistical Validation of Cardiovascular Digital Biomarkers Towards Monitoring the Cardiac Risk in COPD: A Lyfas Case Study. Artif Intell Evol. 2022;3(1):1–16.

  159. Balbirsingh V, Mohammed AS, Turner AM, Newnham M. Cardiovascular disease in chronic obstructive pulmonary disease: a narrative review. Thorax. 2022;77(9):939–45.

    Article  Google Scholar 

  160. Franssen FM, Alter P, Bar N, Benedikter BJ, Iurato S, Maier D, et al. Personalized medicine for patients with COPD: where are we? Int J Chronic Obstructive Pulm Dis. 2019;14:1465–84.

  161. Wouters EF, Wouters BB, Augustin IM, Houben-Wilke S, Vanfleteren LE, Franssen FM. Personalised pulmonary rehabilitation in COPD. Eur Respir Rev. 2018;27(147).

  162. Bećirović LS, Deumić A, Pokvić LG, Badnjevic A. Aritificial Inteligence Challenges in COPD management: a review. In: 2021 IEEE 21st International Conference on Bioinformatics and Bioengineering (BIBE). Piscataway: IEEE; 2021. pp. 1–7.

  163. Regan EA, Hokanson JE, Murphy JR, Make B, Lynch DA, Beaty TH, et al. Genetic epidemiology of COPD (COPDGene) study design. COPD J Chronic Obstructive Pulm Dis. 2011;7(1):32–43.

    Article  Google Scholar 

  164. Castelvecchi D. Can we open the black box of AI? Nat News. 2016;538(7623):20.

    Article  CAS  Google Scholar 

  165. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE international conference on computer vision. Piscataway: IEEE; 2017. pp. 618–26.

  166. Chefer H, Gur S, Wolf L. Transformer interpretability beyond attention visualization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway: IEEE; 2021. pp. 782–91.

  167. Altan G, Kutlu Y, Allahverdi N. Deep learning on computerized analysis of chronic obstructive pulmonary disease. IEEE J Biomed Health Inform. 2019;24(5):1344–50.

    Article  Google Scholar 

  168. Oh AS, Baraghoshi D, Lynch DA, Ash SY, Crapo JD, Humphries SM, et al. Emphysema Progression at CT by Deep Learning Predicts Functional Impairment and Mortality: Results from the COPDGene Study. Radiology. 2022;304(3):672–9.

    Article  PubMed  Google Scholar 

  169. Chen J, Zhonghui X, Sun L, Yu K, Hersh CP, Boueiz A, et al. Deep Learning Integration of Chest CT Imaging and Gene Expression Identifies Novel Aspects of COPD. Chronic Obstr Pulm Dis. 2023;10(4):355–68.

  170. Zhou SK, Greenspan H, Davatzikos C, Duncan JS, Van Ginneken B, Madabhushi A, et al. A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises. Proc IEEE. 2021;109(5):820–38.

    Article  CAS  Google Scholar 

  171. Gayle AV, Quint JK, Fuertes EI. Understanding the relationships between environmental factors and exacerbations of COPD. Exp Rev Respir Med. 2021;15(1):39–50.

    Article  CAS  Google Scholar 

  172. Bourbeau J, Collet JP, Schwartzman K, Ducruet T, Nault D, Bradley C. Economic benefits of self-management education in COPD. Chest. 2006;130(6):1704–11.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

This work was partly supported by the National Natural Science Foundation of China (Nos. 82072008, 82270044, and 62271131) and the Fundamental Research Funds for the Central Universities (N2424010-19).

Author information

Authors and Affiliations

Authors

Contributions

YW conducted literature search and analysis, writing - original draft preparation. SX, ZL, and RC contributed to manuscript review and supervision. SQ conducted writing, review, and editing. All authors contributed to manuscript read, and approved the final manuscript.

Corresponding author

Correspondence to Shouliang Qi.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wu, Y., Xia, S., Liang, Z. et al. Artificial intelligence in COPD CT images: identification, staging, and quantitation. Respir Res 25, 319 (2024). https://doi.org/10.1186/s12931-024-02913-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12931-024-02913-z

Keywords