Skip to main content

ORIGINAL RESEARCH article

Front. Cell Dev. Biol., 24 February 2023
Sec. Molecular and Cellular Pathology
This article is part of the Research Topic Artificial Intelligence Applications in Chronic Ocular Diseases View all 31 articles

Automatic measurement of exophthalmos based orbital CT images using deep learning

  • 1School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China
  • 2Marshall Laboratory of Biomedical Engineering, Shenzhen, China
  • 3Shenzhen Eye Hospital, Jinan University, Shenzhen, China
  • 4Shenzhen Eye Institute, Shenzhen, China
  • 5Shenzhen Overseas Chinese Town Hospital, Shenzhen, China

Introduction: Objective, accurate, and efficient measurement of exophthalmos is imperative for diagnosing orbital diseases that cause abnormal degrees of exophthalmos (such as thyroid-related eye diseases) and for quantifying treatment effects.

Methods: To address the limitations of existing clinical methods for measuring exophthalmos, such as poor reproducibility, low reliability, and subjectivity, we propose a method that uses deep learning and image processing techniques to measure the exophthalmos. The proposed method calculates two vertical distances; the distance from the apex of the anterior surface of the cornea to the highest protrusion point of the outer edge of the orbit in axial CT images and the distance from the apex of the anterior surface of the cornea to the highest protrusion point of the upper and lower outer edges of the orbit in sagittal CT images.

Results: Based on the dataset used, the results of the present method are in good agreement with those measured manually by clinicians, achieving a concordance correlation coefficient (CCC) of 0.9895 and an intraclass correlation coefficient (ICC) of 0.9698 on axial CT images while achieving a CCC of 0.9902 and an ICC of 0.9773 on sagittal CT images.

Discussion: In summary, our method can provide a fully automated measurement of the exophthalmos based on orbital CT images. The proposed method is reproducible, shows high accuracy and objectivity, aids in the diagnosis of relevant orbital diseases, and can quantify treatment effects.

1 Introduction

The exophthalmos reflects the anterior-posterior position of the eye relative to the orbit (Segni et al., 2002; Ameri and Fenton, 2004) and is associated with various orbital diseases, including Graves’ orbitopathy, orbital tumor, and orbital fracture (Burch and Wartofsky, 1993; Bartalena et al., 2000; Alsuhaibani et al., 2011; Guo et al., 2018; Huh et al., 2020; Huang et al., 2022; Ji et al., 2022). Accurate measurement of exophthalmos can assist in the diagnosis of these related diseases (Segni et al., 2002; Lam et al., 2010) and also quantify the treatment outcome.

Currently, the main clinical methods for measuring exophthalmos can be classified as exophthalmometer and computed tomography (CT) methods. The most widely used method is the Hertel exophthalmometer (Migliori and Gladstone, 1984; Dunsky, 1992), which measures the distance from the lateral orbital rim to the corneal surface in a direction perpendicular to the frontal plane as a quantitative indicator of the degree of exophthalmos (O’Donnell et al., 1999). However, the Hertel exophthalmometer has low inter- and intraobserver reproducibility, which in turn affects the reliability of its results (Frueh et al., 1985; Musch et al., 1985; Dunsky, 1992; Chang et al., 1995; Kim and Choi, 2001; Sleep and Manners, 2002; Ameri and Fenton, 2004). Furthermore, this method is not suitable for subjects with abnormalities, such as severe upper eyelid swelling, ptosis, and hyper-deviated eyes, because it is greatly influenced by facial tissues (Na et al., 2019).

Meanwhile, clinicians using CT scans for diagnosing the degree of exophthalmos measure the relevant distance manually by dragging the mouse after determining physiological structures, such as the outer edge of the orbit and the apex of the anterior surface of the cornea (Nkenke et al., 2003; Bingham et al., 2016; Na et al., 2019). Such a manual method of measuring exophthalmos is not only time-consuming and inefficient but also inevitably subjective to the clinician, resulting in poor reproducibility of the interobserver measurements (Huh et al., 2020). Therefore, an objective, accurate, convenient, and efficient method for measuring exophthalmos is necessary for the timely diagnosis or assessment of treatment outcomes for relevant orbital diseases. The development of image processing and deep learning methods has provided the basis for automatic, objective, efficient, and accurate computer-aided diagnosis, and these methods have been widely applied in a variety of fields—especially in studies related to the diagnosis of ophthalmic diseases (Zhao et al., 2022).

In this paper, we propose an automated method based on image processing and deep learning to measure the vertical distance from the apex of the anterior corneal surface to the lateral orbital rim of both eyes and the longest line of the superior to the inferior orbital rim on the axial and sagittal plane of CT images, respectively. The two distance parameters, related to ocular prominence, can be measured objectively, accurately, and efficiently without relying on the clinician. This method can help clinicians diagnose diseases related to protrusion or depression by measuring the exophthalmos.

2 Materials and methods

2.1 Data

Ocular CT images were collected from 31 subjects in the horizontal position and 43 subjects in the sagittal position at the Shenzhen Eye Hospital and Shenzhen Overseas Chinese Hospital using a Philips Ingenuity core 129—a Dutch computed tomography machine with a CT scan thickness of 0.625 mm using the soft tissue window. For this study, 79 horizontal CT images and 99 sagittal CT images including the thickest lens were selected by clinicians empirically.

To train a deep learning network model for automatic eye region segmentation, we divided 79 axial CT images from 31 CT sequence images and 99 sagittal CT images from 43 CT sequence images into a training, validation set, and test set in the ratio of 48:8:23 and 48:8:43, respectively. The ratio of the number of eyes in the training set, validation set and test set for axial and sagittal images is 96:16:40 and 96:16:43, respectively.

Two ophthalmology clinician measured the vertical distance of the line from the apex of the anterior surface of the eye to the most protruding point of the orbital rim for 23 images of the axial plane and 43 images of the sagittal plane. A researcher contributed to the annotation of the ground truth using the “polygon selections” and “fill” function of the software ImageJ (National Institutes of Health, Bethesa, MD, United States) to map the mask of the eye region in all CT images for eye region segmentation by U-Net++ networks.

2.2 Overall approach

In this study, we calculated the vertical distance from the apex of the anterior corneal surface to the lateral orbital rim of both eyes on the axial plane of the CT images, as shown in Figure 1. First, the neural network was trained based on the U-Net++ model for segmenting the eye region in the axial plane of the CT images, and then input the remaining CT images not used for training the model into the segmentation model to obtain a mask of the eye. Furthermore, the lateral orbital rim region of both eyes was obtained after a series of image processing steps. Next, the coordinates of the apex of the anterior corneal surface and the most protruding point of the lateral orbital rim of both eyes were extracted. Additionally, the vertical distance of the line from the apex of the anterior corneal surface to the most protruding point of the upper and lower orbital rim was calculated, as shown in Figure 2.

FIGURE 1
www.frontiersin.org

FIGURE 1. Process of calculating the vertical distance from the apex of the anterior surface of the cornea to the lateral orbital rim of both eyes in the axial plane of the CT image. (A) The original axial plane of the CT image, and after the U-Net++ model and Image processing, we can obtain the binary images of the eye region mask shown in (B) and the lateral orbital rim region of both eyes shown in (C), respectively. (D) The line from the apex of the anterior surface of the eye to the apex of the lateral orbital rim is represented by the blue line in (D), and the vertical distance from the apex of the anterior surface of the eye to the apex of the lateral orbital rim represented by the orange and green lines.

FIGURE 2
www.frontiersin.org

FIGURE 2. The vertical distance from the apex of the anterior surface of the cornea to the longest line between the superior and inferior orbital margins in the sagittal plane of the CT image. (A) Represents the original sagittal CT image, and after the U-Net++ model and Image processing, we obtain the binary images of the eye region mask shown in (B) and the upper and lower orbital rim regions shown in (C), and then after locating the coordinates of the anterior surface apex of the eye and the upper and lower orbital rims, we obtain the line from the apex of the upper and lower orbital rims represented by the blue line in (D) and the vertical distance of the line from the apex of the anterior surface of the cornea to the upper and lower orbital rim most protruding points represented by the orange line.

Similarly, the neural network must be trained to segment the eye region based on the U-Net++ model in the sagittal plane of the CT images while processing the CT images to obtain the upper and lower orbital rim regions. After extracting the coordinates of the apex of the anterior corneal surface and the most protruding point of the upper and lower orbital rim, the vertical distance of the line from the apex of the anterior corneal surface to the most protruding point of the upper and lower orbital rim was calculated.

2.3 Segmentation

In order to obtain the best segmentation results, we trained the models commonly used segmentation networks in clinical practice, including FCN32 (Long et al., 2015), SegNet (Badrinarayanan et al., 2017), U-Net (Ronneberger et al., 2015), U-Net++ (Zhou et al., 2018), and Res-U-Net (Zhang et al., 2018), respectively, using the dataset of this paper. The U-Net++ model that achieved the best segmentation performance on the test set was selected for the eye region segmentation.

We implemented the U-Net++ model for eye region segmentation. First, the learning rate and decay rate were set to 0.0001 and 0.99, respectively, Kaiming initialization (He et al., 2015) was used for initializing the weights of the model. Dice loss (Milletari et al., 2016) was used to compare the model segmentation results with the ground truth, and the Adam optimization method (Kingma and Ba, 2014) was applied to minimize the loss value of the network. We inputted the original images and ground truth masks from the training and validation sets to the U-Net++ model. After 200 training iterations of epochs on a server configured with the GPU NVIDIA GeForce GTX 3090TI and using the Pytorch (Paszke et al., 2017) framework, we obtained a network model for automatic eye region segmentation. The formula for Dice loss is shown below.

Diceloss=12XY/X+Y

X represents the eye area mask produced by the neural network segmentation, and Y represents the ground truth of the eye area mask input to the neural network.

In addition to the orbital region, we must segment the lateral orbital rim regions of both eyes in the axial plane as well as the superior and inferior orbital rim regions in the sagittal plane of the CT images to extract the coordinates of their most protruding points. The orbital rim, i.e., the human orbital bone, shows strong contrast in CT images compared to other tissues, so we can use traditional image processing methods to extract the skeletal region.

We first performed threshold segmentation separately with a grayscale values’ threshold of 200 for the CT images of the two planar views (as in Figure 3A) empirically. After extracting the structures with grayscale values greater than 200 (as in Figure 3B), we eliminated the residual watermark in the CT images, following threshold segmentation, by performing the morphological opening (as in Figure 3C). Finally, after eliminating the smaller connected domains in the images (considered to be noisy), a binary image containing the lateral orbital rims or the upper and lower orbital rims of both eyes was obtained (as in Figure 3D).

FIGURE 3
www.frontiersin.org

FIGURE 3. Image processing of the segmented orbital rim region. (A) The unprocessed CT image and the binary image in (B) are obtained after the threshold segmentation; the binary image shown in (C) is obtained after the morphological opening operation, and finally, the smaller connected domain is eliminated to obtain the binary image containing the outer edge of the orbit shown in (D). The first row represents the axial CT image and the second row the sagittal CT image.

2.4 Distance calculation

In the axial plane of the CT images, the lateral orbital rims of both eyes were located in the leftmost third of the CT image and the rightmost third of the CT image. Furthermore, the most protruding point can be regarded as the pixel point closest to the top, i.e., the pixel point with the smallest y-value in the image coordinate. Therefore, to extract the coordinates of the most protruding point of the lateral orbital rim of both eyes in the axial plane of the CT image, we divided the (d) image in Figure 4 into three subplots: left, middle, and right. First, the images were divided according to the direction of the x-axis in the image coordinates, and then the pixel points were traversed in the left and right images in turn. The pixel point with the smallest y-value in the “white” area of the two subplots was shortlisted as the coordinate of the most protruding point of the lateral orbital rim of both eyes, as shown in Figure 4. The entire process is shown in Figure 4.

FIGURE 4
www.frontiersin.org

FIGURE 4. The process of obtaining the coordinates of the most protruding point of the lateral orbital rim of both eyes in the axial plane of the CT image. (A) First row of Panel (B), which is the binary image containing the lateral orbital rim region of both eyes. Subsequently, the coordinates of the most protruding points (x1, y1), (x2, y2) of the left and right lateral orbital rims can be determined by traversing the two subgraphs separately, and the straight line passing through the two most protruding points can be visualized by the blue line in (C).

In the sagittal plane of the CT images, the upper and lower orbital margins are located in the upper and lower molecular maps of the CT image, respectively—their most protruding point can be regarded as the point closest to the left side of the image, i.e., the pixel point with the smallest x-value in the image coordinates. Therefore, we follow an operation similar to that of the axial plane–the pixel points in the “white” area in the upper and lower submaps are traversed, respectively, and the point with the smallest x-value is recorded as the coordinate of the most protruding point of the upper and lower orbital margins. The entire process is shown in Figure 5.

FIGURE 5
www.frontiersin.org

FIGURE 5. The process of obtaining the coordinates of the most protruding points of the upper and lower orbital rim in the sagittal plane of the CT image. (A) The second row of Figure 4B, which is the binary image containing the upper and lower orbital rim regions, and extracts the submaps of the upper and lower halves, respectively, to obtain the two submaps as shown in (B). Subsequently, the coordinates of the most protruding points of the superior and inferior orbital rims can be determined by traversing the two subgraphs separately, and then the straight line passing through the two most protruding points can be obtained—the straight line is shown in blue in (C).

After obtaining the coordinates of the lateral orbital rims of both eyes, we obtained the equation of the line passing through the two points of the lateral orbital rims of both eyes. Similarly, we obtained the equation of the line passing through the two points of the upper and lower orbital rims based on the coordinates of the most protruding points of the upper and lower orbital rims in the sagittal plane of the CT image. Subsequently, we traversed the pixel points of the eye region mask output by the eye region segmentation model and recorded the point with the smallest x-value among the mask pixel points as the coordinates of the most protruding point of the anterior corneal surface vertex. From this, we calculated the vertical distance from the apex of the anterior corneal surface to the lateral orbital rim of both eyes in the axial plane of the CT image and the vertical distance from the apex of the anterior corneal surface to the upper and lower orbital rims in the sagittal plane of the CT image. The results of our automated method and the manual measurements by the physicians were compared.

2.5 Statistical analysis

The Dice coefficient (Dice, 1945), Intersection Over Union (IOU), precision, and recall were used as metrics to evaluate the segmentation performance of the model. The metrics were calculated as shown below.

Dice=2TP/FP+2TP+FN
IOU=TP/TP+FN+FP
Precision=TP/TP+FP
Recall=TP/TP+FN

The intraclass correlation coefficient (ICC) and the concordance correlation coefficient (CCC) were used to demonstrate the concordance between the results of our automated method and the manual measurements of the physicians. The two-way mixture model and the absolute consistency type were chosen for the calculation of the intra-group correlation coefficients.

3 Results

3.1 Ocular segmentation

We used the segmentation results of 40 eyes in 23 horizontal CT images and 43 eyes in 43 sagittal CT images to test the ability of the model to segment eye region. Table 1 shows the segmentation performance of the five models on the test set data, and Figure 6 shows the visualization of the U-Net++ model segmentation results. From the results, we observe that the region of the eye was segmented accurately in both horizontal and sagittal CT images. This indicates that this method can accurately locate the vertex coordinates of the anterior surface of the eye.

TABLE 1
www.frontiersin.org

TABLE 1. The mean values (standard deviation) of the Dice Coefficient, IOU, Precision, Recall for segmenting the ocular region model in the axial plane and the sagittal plane of the CT images, respectively.

FIGURE 6
www.frontiersin.org

FIGURE 6. Visualization segmentation results of the model used to segment the eye region. Row (A) represents the visual segmentation results of the model in the axial plane of the test set, and row (B) represents the visual segmentation results of the model in the sagittal plane of the test set. The purple and green area represent the segmentation masks predicted by the manual and network, respectively. The light blue area is the overlapping part of both, i.e., the correctly predicted eye area.

3.2 Ocular prominence measurement

For testing, we used the vertical distance from the apex of the anterior corneal surface to the most protruding point of the lateral orbital rim of both eyes in 40 eyes from 23 CT images in the axial plane and the vertical distance from the apex of the anterior corneal surface to the most protruding point of the upper and lower orbital rims in 43 eyes from 43 CT images in the sagittal plane. Table 2 shows the mean values (standard deviation) of the Exophthalmometric values between the results measured by the proposed automated method and the manually measured results. The Bland-Altman plots and the scatter diagram between the computed results of the proposed automated method of ocular prominence measurement on the test set and its corresponding manual measurement by the physician is shown in Figure 7. Figure 8 shows the results of the proposed automated method and the manual measurement by the physician on an axial plane and one sagittal plane of the CT image, respectively. The results of the statistical analysis between the measurements using the two methods on all test sets are shown in Table 3.

TABLE 2
www.frontiersin.org

TABLE 2. The mean values (standard deviation) of the Exophthalmometric values between the results measured by the proposed automated method and the manually measured results.

FIGURE 7
www.frontiersin.org

FIGURE 7. The Bland-Altman plots and the scatter plots between the results on the test set using the automated method proposed in this paper. Row (A) represents the Bland-Altman plots and row (B) is the scatter plot, column (C) means the results on the axial plane while column (D) is the results on the sagittal plane. In the scatter plots, the y-axis represents the results calculated by the automated method we proposed in this paper while the x-axis means the results measured by the doctors.

FIGURE 8
www.frontiersin.org

FIGURE 8. Visualization of measurement results of the proposed automated method and the manual measurement method by the physician on the test set. Row (A) represents the visualization result of measurement using the proposed method, and row (B) is the visualization result of manual measurement by the physician on the same image.

TABLE 3
www.frontiersin.org

TABLE 3. Results of statistical analysis of the concordance correlation coefficient (CCC) and the intraclass correlation coefficient (ICC) between the results measured by the proposed automated method and the manually measured results.

From the results of the statistical analysis, it can be observed that although the Bland-Altman plots diagram as well as the mean values of the Exophthalmometric values show a stable error in the results measured by the proposed automated method and the manually measured results, our method is in good agreement with the results of the manual measurement by the physician for both vertical distances. Thus, the accuracy of this method can be verified.

4 Discussion

The degree of orbital protrusion is associated with a variety of orbital diseases, and its accurate quantification is important to diagnose certain orbital diseases and determine the effectiveness of their treatment. CT imaging has been used to measure the prominence of the eye because of its high-resolution accuracy and ability to analyze multiple views simultaneously (Kim and Choi, 2001; Nkenke et al., 2003; 2004; Fang et al., 2013). Some studies have shown that CT image-based ocular prominence measurements are more accurate (Hallin and Feldon, 1988; Segni et al., 2002; Nkenke et al., 2003; 2004; Ramli et al., 2015) and correlate well with measurements using the Hertel ocular prominence meter (Klingenstein et al., 2022). For special subjects such as children and those suffering from ptosis, using CT images to measure ocular prominence is the only feasible method. Currently, the most common clinical method is to manually measure the vertical distance from the apex of the anterior corneal surface to the lateral orbital rim of both eyes on an axial CT image as a measure of ocular prominence (as shown in Figure 9A ) (“Axial Globe Position Measurement: A Prospective Multicenter Study by the International Thyroid Eye Disease Society,” 2016; Nkenke et al., 2004). However, when the subject’s head is tilted, using only one plane of view may lead to large errors. In 2019, Na et al. (Na et al., 2019) proposed a method to represent ocular prominence in sagittal CT images by measuring the vertical distance of the longest line connecting the anterior surface apex of the cornea to the superior orbital rim to the inferior orbital rim (as shown in Figure 9B). The method proposed by Park et al. has been validated to be comparable to the Hertel exophthalmometer method with high correlation while being applicable to subjects with horizontal and vertical strabismus. Therefore, a exophthalmos measuring method that combines the two planar views described above would be applicable to a wider population with guaranteed accuracy.

FIGURE 9
www.frontiersin.org

FIGURE 9. Two commonly used clinical parameters for measuring the prominence of the eye based on CT images. (A) The axial plane of the CT image, (A and B) represent the vertical distance from the apex of the anterior corneal surface to the lateral orbital rim of both eyes, respectively; (B) The sagittal plane of the CT image, C represents the vertical distance from the apex of the anterior corneal surface to the longest line of the upper and lower orbital rims.

In this paper, we propose a method based on deep learning and image processing techniques to combine axial and sagittal CT images for the automatic measurement of exophthalmos. The experimental results show that our method can achieve accurate segmentation results with Dice coefficients of 0.976 ± 0.015 and 0.977 ± 0.0135 for the eye region in the axial and sagittal plane of the CT images, respectively, on the dataset used in this paper, as shown in Figure 6. We used image processing techniques to segment the orbital region to achieve accurate localization of the apex of the anterior surface of the eye and the most protruding point of the outer edge of the orbit. Based on the results obtained, the CCC and ICC between the two methods were 0.988 and 0.957 for the axial plane of the CT images, respectively, and 0.990 and 0.965 for the sagittal plane of the CT images, respectively—in our dataset of 23 axial and 43 sagittal CT images, which shows high consistency.

The deep learning and digital image processing methods used in our study can automatically segment the structures of the eye and orbital rim, and locate the apex of the anterior corneal surface and the most protruding point of the orbital rim. The process can then calculate the relevant parameters, ensuring the high accuracy and reproducibility of this method to a certain extent in our dataset. Furthermore, the suggested approach can determine the relevant exophthalmos measurements in both axial and sagittal planes of CT scans, offering medical professionals a multi-dimensional reference for diagnosing orbital disorders in patients displaying abnormal exophthalmos seen only in the axial or sagittal plane. After conducting a PubMed search using the keywords “proptosis” and “CT,” we discovered 57 relevant studies published in the past 20 years. However, all of these studies relied on manual drawings and measurements performed by clinicians or researchers, which can be remedied by implementing the proposed method. Additionally, the full automation of the process in this paper not only minimizes the impact of subjective factors on measurement results, but also enhances measurement efficiency. On average, the time required to calculate the vertical distance from the anterior corneal surface’s apex to the most protruding point of the lateral orbital rim in axial CT images and the vertical distance from the anterior corneal surface’s apex to the most protruding points of the upper and lower orbital rims in sagittal CT images is 0.9 s and 0.75 s, respectively. This automated method significantly reduces the time and effort required for eye protrusion measurement compared to manual methods.

The work in this paper was performed on 2D CT images, which meets the practical needs of current clinicians for diagnosis (Kim and Choi, 2001), especially in patients with eyelid exophthalmos and other conditions (Na et al., 2019). It is worth mentioning that some research teams have implemented the quantification of ocular prominence on 3D CT images (Guo et al., 2017; 2018; Huh et al., 2020; Willaert et al., 2020), but none of them have been fully automated. However, the mainstream methods in clinical practice are still dominated by the Hertel ocular prominence meter method and the lightweight-based 2D CT image method. The 3D CT image-based ocular prominence measurement method is complex and time-consuming, and we will explore other ocular prominence-related parameters (Kim and Choi, 2001; Campi et al., 2013; Guo et al., 2017; Afanasyeva et al., 2018; Choi and Lee, 2018) in our future work, for automatic measurement and validate their practical feasibility in a clinical setting.

However, the approach in this paper applies to both axial and sagittal CT images and requires the most protruding point of the outer edge of the orbit to determine the measurement of the exophthalmos, which is a limitation in case of some images with incomplete, missing or displaced outer orbit edges.

5 Conclusion

This study introduces an automated approach for assessing eye prominence in both axial and sagittal CT images of the orbit using deep learning and image processing techniques. This method eliminates the need for prior knowledge from clinicians, thereby reducing their workload. On the experimental dataset, the method shows satisfactory efficiency, accuracy, reliability, and reproducibility. This approach has the potential to support the diagnosis and treatment quantification of related orbital diseases.

Data availability statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Author contributions

YHZ conceived and designed the research and wrote the manuscript; JR and XW wrote the manuscript; YJZ, GL, and HZ designed the research, acquired the article information, and revised the manuscript. All authors contributed to the article and have approved the submitted version.

Funding

This study was supported by Shenzhen Fund for Guangdong Provincial High-level Clinical Key Specialties (SZGSP014), Sanming Project of Medicine in Shenzhen (SZSM202011015), and Science, Technology and Innovation Commission of Shenzhen (GJHZ20190821113605296).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Afanasyeva, D. S., Gushchina, M. B., Gerasimov, M. Y., and Borzenok, S. A. (2018). Computed exophthalmometry is an accurate and reproducible method for the measuring of eyeballs’ protrusion. J. Cranio-Maxillofacial Surg. 46 (3), 461–465. doi:10.1016/j.jcms.2017.12.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Alsuhaibani, A. H., Carter, K. D., Policeni, B., and Nerad, J. A. (2011). Orbital volume and eye position changes after balanced orbital decompression. Ophthalmic Plastic Reconstr. Surg. 27 (3), 158–163. doi:10.1097/IOP.0b013e3181ef72b3

PubMed Abstract | CrossRef Full Text | Google Scholar

Ameri, H., and Fenton, S. (2004). Comparison of unilateral and simultaneous bilateral measurement of the globe position, using the Hertel exophthalmometer. Ophthalmic Plastic Reconstr. Surg. 20, 448–451. doi:10.1097/01.iop.0000143712.42344.8c

PubMed Abstract | CrossRef Full Text | Google Scholar

Badrinarayanan, V., Kendall, A., and Cipolla, R. (2017). SegNet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Analysis Mach. Intell. 39 (12), 2481–2495. doi:10.1109/TPAMI.2016.2644615

CrossRef Full Text | Google Scholar

Bartalena, L., Pinchera, A., and Marcocci, C. (2000). Management of Graves’ ophthalmopathy: Reality and perspectives. Endocr. Rev. 21 (2), 168–199. doi:10.1210/edrv.21.2.0393

PubMed Abstract | CrossRef Full Text | Google Scholar

Bingham, C. M., Sivak-Callcott, J. A., Gurka, M. J., Nguyen, J., Hogg, J. P., Feldon, S. E., et al. (2016). Axial globe position measurement: A prospective multicenter study by the international thyroid eye disease society. Ophthalmic Plastic Reconstr. Surg. 32 (2), 106–112. doi:10.1097/IOP.0000000000000437

PubMed Abstract | CrossRef Full Text | Google Scholar

Burch, H. B., and Wartofsky, L. (1993). Graves’ ophthalmopathy: Current concepts regarding pathogenesis and management. Endocr. Rev. 6, 747–793. doi:10.1210/edrv-14-6-747

PubMed Abstract | CrossRef Full Text | Google Scholar

Campi, I., Vannucchi, G. M., Minetti, A. M., Dazzi, D., Avignone, S., Covelli, D., et al. (2013). A quantitative method for assessing the degree of axial proptosis in relation to orbital tissue involvement in Graves’ orbitopathy. Ophthalmology 120 (5), 1092–1098. doi:10.1016/j.ophtha.2012.10.041

PubMed Abstract | CrossRef Full Text | Google Scholar

Chang, A. A., Bank, A., Francis, I. C., and Kappagoda, M. B. (1995). Clinical exophthalmometry: A comparative study of the luedde and Hertel exophthalmometers. Aust. N. Z. J. Ophthalmol. 23 (4), 315–318. doi:10.1111/j.1442-9071.1995.tb00182.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Choi, K. J., and Lee, M. J. (2018). Comparison of exophthalmos measurements: Hertel exophthalmometer versus orbital parameters in 2-dimensional computed tomography. Can. J. Ophthalmol. 53 (4), 384–390. doi:10.1016/j.jcjo.2017.10.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Dice, L. R. (1945). Measures of the amount of ecologic association between species. Ecology 26 (3), 297–302. doi:10.2307/1932409

CrossRef Full Text | Google Scholar

Dunsky, I. L. (1992). Normative data for Hertel exophthalmometry in a normal adult black population. Optometry Vis. Sci. 69 (7), 562–564. doi:10.1097/00006324-199207000-00009

PubMed Abstract | CrossRef Full Text | Google Scholar

Fang, Z. J., Zhang, J. Y., and He, W. M. (2013). CT features of exophthalmos in Chinese subjects with thyroid-associated ophthalmopathy. Int. J. Ophthalmol. 6 (2), 146–149. doi:10.3980/j.issn.2222-3959.2013.02.07

PubMed Abstract | CrossRef Full Text | Google Scholar

Frueh, B. R., Garber, F., Grill, R., and Musch, D. C. (1985). Positional effects on exophthalmometer readings in Graves’ eye disease. Archives Ophthalmol. 103 (9), 1355–1356. doi:10.1001/archopht.1985.01050090107043

PubMed Abstract | CrossRef Full Text | Google Scholar

Guo, J., Qian, J., and Yuan, Y. (2018). Computed tomography measurements as a standard of exophthalmos? Two-dimensional versus three-dimensional techniques. Curr. Eye Res. 43 (5), 647–653. doi:10.1080/02713683.2018.1431285

PubMed Abstract | CrossRef Full Text | Google Scholar

Guo, J., Qian, J., Yuan, Y., Zhang, R., and Huang, W. (2017). A novel three-dimensional vector analysis of axial globe position in thyroid eye disease. J. Ophthalmol. 2017, 7253898–7253899. doi:10.1155/2017/7253898

PubMed Abstract | CrossRef Full Text | Google Scholar

Hallin, E. S., and Feldon, S. E. (1988). Graves’ ophthalmopathy: II. Correlation of clinical signs with measures derived from computed tomography. Br. J. Ophthalmol. 72 (9), 678–682. doi:10.1136/bjo.72.9.678

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2015). Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. IEEE Computer Society.

Google Scholar

Huang, X., Ju, L., Li, J., He, L., Tong, F., Liu, S., et al. (2022). An intelligent diagnostic system for thyroid-associated ophthalmopathy based on facial images. Front. Med. 9, 920716. doi:10.3389/fmed.2022.920716

PubMed Abstract | CrossRef Full Text | Google Scholar

Huh, J., Park, S. J., and Lee, J. K. (2020). Measurement of proptosis using computed tomography based three-dimensional reconstruction software in patients with Graves’ orbitopathy. Sci. Rep. 10 (1), 14554. doi:10.1038/s41598-020-71098-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Ji, Y., Chen, N., Liu, S., Yan, Z., Qian, H., Zhu, S., et al. (2022). Research progress of artificial intelligence image analysis in systemic disease-related ophthalmopathy. Dis. Markers 2022, 3406890–3406910. doi:10.1155/2022/3406890

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, I. T., and Choi, J. B. (2001). Normal range of exophthalmos values on orbit computerized tomography in Koreans. Ophthalmologica 215 (3), 156–162. doi:10.1159/000050850

PubMed Abstract | CrossRef Full Text | Google Scholar

Kingma, D. P., and Ba, J. (2014). “Adam: A method for stochastic optimization,” in International conference on learning representations.

Google Scholar

Klingenstein, A., Samel, C., Garip-Kübler, A., Hintschich, C., and Müller-Lisse, U. G. (2022). Cross-sectional computed tomography assessment of exophthalmos in comparison to clinical measurement via Hertel exophthalmometry. Sci. Rep. 12 (1), 11973. doi:10.1038/s41598-022-16131-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Lam, A. K. C., Lam, C. F., Leung, W. K., and Hung, P. K. (2010). Intra-observer and inter-observer variation of Hertel exophthalmometry. Ophthalmic & Physiological Opt. 29 (4), 472–476. doi:10.1111/j.1475-1313.2008.00617.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Long, J., Shelhamer, E., and Darrell, T. (2015). Fully convolutional networks for semantic segmentation. (arXiv:1411.4038). arXiv. Available at: http://arxiv.org/abs/1411.4038.

Google Scholar

Migliori, M. E., and Gladstone, G. J. (1984). Determination of the normal range of exophthalmometric values for black and white adults. Am. J. Ophthalmol. 98 (4), 438–442. doi:10.1016/0002-9394(84)90127-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Milletari, F., Navab, N., and Ahmadi, S. A. (2016). “V-Net: Fully convolutional neural networks for volumetric medical image segmentation,” in 2016 fourth international conference on 3D vision (3DV).

CrossRef Full Text | Google Scholar

Musch, D. C., Frueh, B. R., and Landis, J. R. (1985). The reliability of Hertel exophthalmometry. Observer variation between physician and lay readers. Ophthalmology 92 (9), 1177–1180. doi:10.1016/S0161-6420(85)33880-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Na, R. P., Moon, J. H., and Lee, J. K. (2019). Hertel exophthalmometer versus computed tomography scan in proptosis estimation in thyroid-associated orbitopathy. Clin. Ophthalmol. 13, 1461–1467. doi:10.2147/OPTH.S216838

PubMed Abstract | CrossRef Full Text | Google Scholar

Nkenke, E., Benz, M., Maier, T., Wiltfang, J., Holbach, L. M., Kramer, M., et al. (2003). Relative en- and exophthalmometry in zygomatic fractures comparing optical non-contact, non-ionizing 3D imaging to the Hertel instrument and computed tomography. J. Cranio-Maxillofacial Surg. 31 (6), 362–368. doi:10.1016/j.jcms.2003.07.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Nkenke, E., Maier, T., Benz, M., Wiltfang, J., Holbach, L. M., Kramer, M., et al. (2004). Hertel exophthalmometry versus computed tomography and optical 3D imaging for the determination of the globe position in zygomatic fractures. Int. J. Oral Maxillofac. Surg. 33 (2), 125–133. doi:10.1054/ijom.2002.0481

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Donnell, N. P., Virdi, M., and Kemp, E. G. (1999). Hertel exophthalmometry: The most appropriate measuring technique. Br. J. Ophthalmol. 83 (9), 1096b. doi:10.1136/bjo.83.9.1096b

PubMed Abstract | CrossRef Full Text | Google Scholar

Paszke, A., Gross, S., Chintala, S., Chanan, G., Yang, E., Devito, Z., et al. (2017). Automatic differentiation in PyTorch.

Google Scholar

Ramli, N., Kala, S., Samsudin, A., Rahmat, K., and Zainal Abidin, Z. (2015). Proptosis—correlation and agreement between Hertel exophthalmometry and computed tomography. Orbit 34 (5), 257–262. doi:10.3109/01676830.2015.1057291

PubMed Abstract | CrossRef Full Text | Google Scholar

Ronneberger, O., Fischer, P., and Brox, T. (2015). U-Net: Convolutional networks for biomedical image segmentation. (arXiv:1505.04597). arXiv. Available at: http://arxiv.org/abs/1505.04597.

Google Scholar

Segni, M., Bartley, G. B., Garrity, J. A., Bergstralh, E. J., and Gorman, C. A. (2002). Comparability of proptosis measurements by different techniques. Am. J. Ophthalmol. 133 (6), 813–818. doi:10.1016/s0002-9394(02)01429-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Sleep, T. J., and Manners, R. M. (2002). Interinstrument variability in hertel-type exophthalmometers. Ophthalmic Plastic Reconstr. Surg. 18 (4), 254–257. doi:10.1097/00002341-200207000-00004

PubMed Abstract | CrossRef Full Text | Google Scholar

Willaert, R., Shaheen, E., Deferm, J., Vermeersch, H., Jacobs, R., and Mombaerts, I. (2020). Three-dimensional characterisation of the globe position in the orbit. Graefe’s Archive Clin. Exp. Ophthalmol. 258 (7), 1527–1532. doi:10.1007/s00417-020-04631-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Z., Liu, Q., and Wang, Y. (2018). Road extraction by deep residual U-net. IEEE Geoscience Remote Sens. Lett. 15 (5), 749–753. doi:10.1109/LGRS.2018.2802944

CrossRef Full Text | Google Scholar

Zhao, J., Lu, Y., Zhu, S., Li, K., Jiang, Q., and Yang, W. (2022). Systematic bibliometric and visualized analysis of research hotspots and trends on the application of artificial intelligence in ophthalmic disease diagnosis. Front. Pharmacol. 13, 930520. doi:10.3389/fphar.2022.930520

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, Z., Rahman Siddiquee, M. M., Tajbakhsh, N., and Liang, J. (2018). “UNet++: A nested U-net architecture for medical image segmentation,” in Deep learning in medical image analysis and multimodal learning for clinical decision support. Editors D. Stoyanov, Z. Taylor, G. Carneiro, T. Syeda-Mahmood, A. Martel, L. Maier-Heinet al. (Springer International Publishing), 11045, 3–11. doi:10.1007/978-3-030-00889-5_1

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: CT images, deep learning, exophthalmos, orbital diseases, thyroid-associated ophthalmopathy

Citation: Zhang Y, Rao J, Wu X, Zhou Y, Liu G and Zhang H (2023) Automatic measurement of exophthalmos based orbital CT images using deep learning. Front. Cell Dev. Biol. 11:1135959. doi: 10.3389/fcell.2023.1135959

Received: 02 January 2023; Accepted: 13 February 2023;
Published: 24 February 2023.

Edited by:

Yanwu Xu, Baidu, China

Reviewed by:

Xiujuan Zheng, Sichuan University, China
Bo Zheng, Huzhou University, China
Peifang Ren, Zhejiang University, China

Copyright © 2023 Zhang, Rao, Wu, Zhou, Liu and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yongjin Zhou, eWp6aG91QHN6dS5lZHUuY24=; Guiqin Liu, bGl1Z3VpcWluMTA1OUAxNjMuY29t; Hua Zhang, emhhbmdodWF3aW5Ac29odS5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.