banner
Home / News / Visual field prediction using a deep bidirectional gated recurrent unit network model
News

Visual field prediction using a deep bidirectional gated recurrent unit network model

Aug 08, 2023Aug 08, 2023

Scientific Reports volume 13, Article number: 11154 (2023) Cite this article

320 Accesses

2 Altmetric

Metrics details

Although deep learning architecture has been used to process sequential data, only a few studies have explored the usefulness of deep learning algorithms to detect glaucoma progression. Here, we proposed a bidirectional gated recurrent unit (Bi-GRU) algorithm to predict visual field loss. In total, 5413 eyes from 3321 patients were included in the training set, whereas 1272 eyes from 1272 patients were included in the test set. Data from five consecutive visual field examinations were used as input; the sixth visual field examinations were compared with predictions by the Bi-GRU. The performance of Bi-GRU was compared with the performances of conventional linear regression (LR) and long short-term memory (LSTM) algorithms. Overall prediction error was significantly lower for Bi-GRU than for LR and LSTM algorithms. In pointwise prediction, Bi-GRU showed the lowest prediction error among the three models in most test locations. Furthermore, Bi-GRU was the least affected model in terms of worsening reliability indices and glaucoma severity. Accurate prediction of visual field loss using the Bi-GRU algorithm may facilitate decision-making regarding the treatment of patients with glaucoma.

Glaucoma, a leading cause of blindness worldwide, is characterized by irreversible loss of retinal ganglion cells1,2. Structural changes in retinal ganglion cells and the optic nerve head cause progressive deterioration of the visual field2. The prediction of future visual field is essential to preserve visual function. However, visual field test results are susceptible to random errors and fluctuations, particularly in patients with glaucoma, which hinders accurate prediction of visual field changes3.

Over the past several years, machine learning algorithms have demonstrated good performance in the prediction of glaucoma progression. Wang et al.4 classified and determined the progression of 16 archetypes of visual field defects. Murata et al.5 found superior prediction ability of variational Bayes linear regression, a type of machine learning algorithm, compared with pointwise linear regression (LR). Because of the recent development of artificial intelligence, deep learning algorithms have been used for various tasks with excellent performance. However, only a few studies have predicted the progression of visual field defects using deep learning algorithms. Wen et al.6 used a convolutional neural network to predict future visual fields, using a single visual field examination as input. Berchuck et al.7 used a variational autoencoder model to estimate the rate of visual field progression.

Recurrent neural network (RNN), an artificial network with recurrent connections, has been used for sequential time series with temporal dependence and for sequence modeling8. It can process current data, using previous data to make predictions, based on dependencies between sequential elements9,10. The two main variants of RNN, long short-term memory (LSTM)11 and gated recurrent unit (GRU)12, model long-term dependency into long sequences. In a previous study, we found that LSTM had superior abilities to predict future visual fields, compared with ordinary least-squares LR13. Dixit et al.14 found that LSTM networks can predict the longitudinal local and global trends in visual fields.

GRU uses gating units more efficiently and at a similar rate, compared with typical LSTMs15,16,17. Several studies have revealed that GRU has excellent performance for sequential data analysis, compared with other RNN types12,15,18,19. Recently, a bidirectional RNN method has been developed via simultaneous training with positive and negative time directions, which provides a better understanding of context20. Lynn et al.15 compared several RNN-based models for human identification using electrocardiogram-based biometrics from sequential time-series data. The bidirectional network with LSTM and GRU models was more effective than conventional RNN models, and the bidirectional-gated recurrent unit (Bi-GRU) model exhibited performance superior to the bidirectional LSTM model. Because visual field examinations provide sequential data with extensive interconnections, Bi-GRU may achieve better prediction of visual field progression, compared with the previous LSTM-based RNN model.

To our knowledge, this is the first study to use Bi-GRU to predict visual field damage. In a previous study, we evaluated the performance of LSTM in predicting visual field defects. Because the present study used a larger dataset than our previous work, we developed a computationally efficient RNN-based Bi-GRU model. We compared the performance of the Bi-GRU model with the performances of conventional LR and LSTM models.

This retrospective study was conducted in accordance with the tenets of the Declaration of Helsinki. Visual field data were collected from glaucoma clinics at Pusan National University Hospital, Kosin University Gospel Hospital, Dong-A University Hospital, Busan Paik Hospital, and Pusan National University Yangsan Hospital between June 2004 and January 2021. The study protocol was approved by the institutional review boards of Pusan National University Hospital (Approval No.: 2203-018-113), Kosin University Gospel Hospital (Approval No.: 2018-12-028), Dong-A University Hospital (Approval No.: 22-074), Busan Paik Hospital (Approval No.: 2021-03-014-002), and Pusan National University Yangsan Hospital (Approval No.: 05-2018-172). The requirement for patient consent was waived by the institutional review boards because of the retrospective study design. Sex and diagnostic data were retrospectively collected from medical records.

Participants who completed a minimum of six consecutive visual field examinations were included in the training and test datasets. There was no patient overlap between the two datasets. Eyes with an interval of ≥ 3 years between the first and sixth visual field examinations were included. For example, in an eye with 13 consecutive visual field examinations, the first through sixth examinations were considered the first dataset, the seventh through twelfth examinations were considered the second dataset, and the thirteenth examination was excluded from the dataset. The first five examinations were used as input data to predict the sixth examination, and the seventh through eleventh examinations were used as input data to predict the twelfth examination (Fig. 1).

Representative time displacement sequence of a patient who completed 13 visual field tests. Visual field test dates indicated in gray boxes were used for training, and dates in black boxes were used for prediction.

We obtained 6-cell data from 8323 visual fields of 6685 eyes and 4593 participants. Datasets from 7051 (85%) and 1,272 (15%) individuals were included in the training and test datasets, respectively. In total, 7051 records from the training dataset were randomly split into training and validation datasets at a ratio of 9:1. The validation dataset was used to determine the fitness of the neural network during training to prevent overfitting. All 8323 datasets included six visual field examinations, and the mean follow-up duration for the six examinations was 4.39 \(\pm\) 1.69 years. Table 1 presents the characteristics of each dataset.

Automated perimetry was conducted using a Humphrey Visual Field Analyzer 750i (Carl Zeiss Meditec, Inc., Dublin, CA, USA) and the 24-2 or 30-2 Swedish interactive threshold algorithm. Among the 54 test points of the 24-2 test pattern, the two points of physiological scotoma were excluded; the remaining 52 test points were used. The 30-2 test pattern was converted to the 24-2 test pattern using the overlapped test points. Reliable visual field tests were defined as a false positive rate < 33%, false negative rate < 33%, and fixation loss < 33%.

We used the LSTM and Bi-GRU neural network models. Python software (version 3.8) with TensorFlow 2.3 (Google, Mountain View, CA, USA) was used to predict visual field loss. Supplementary Fig. S1 illustrates the two model structures.

We built one-layer neural networks to learn the structural information of a specific dataset using preprocessed input. The LSTM cell-based neural networks were defined as follows:

where \({W}_{f},{W}_{i}, {W}_{o}, and {W}_{C}\) represent the weights and \({b}_{f}, {b}_{i}, {b}_{o}, { and b}_{C}\) represent the bias in the network, respectively, of the three gates and a memory cell. ⨂ is the elementwise product between two vectors. The sigmoid is the activation function used in the network, written as follows:

The input and output gates regulate the flow of memory cell inputs and outputs throughout the network, while the forget gate is incorporated into the memory cell to transmit output information with high weights from the previous neuron to the next one. The information residing in the memory depends on the high activation results. If the input unit has high activation, information is stored in the memory cell. On the other hand, if the output unit has high activation, it passes the information to the next neuron. Input information with a high weight resides in the memory cell. Sigmoid and tanh are employed as the active functions for the gates. Here, h(t-1) represents the prior hidden layer units that add the weights of the three gates in an elementwise manner. After processing Eq. (4), (C)t indicates the current memory cell unit. Equation (5) shows the elementwise multiplication of the prior hidden unit outputs and previous memory cell unit. Nonlinearity is introduced through the tanh and sigmoid activation functions as shown in Eqs. (1–5). Here, t − 1 and t are the previous and current time steps.

GRU is a simplified variant of LSTM that only has two gates: the update gate, which comprises the input and forget gates, and the reset gate. It has no additional memory cell to retain information and can only control information inside the unit.

The update gate in Eq. (6) determines the extent of information updating. In Eq. (7), the rest gate is similar to the update gate; if the gate is set to zero, GRU reads the input sequences and forgets the previously calculated state. Furthermore, \(\widetilde{{h}_{t}}\) exhibits functionality identical to the recurrent unit, and ℎt of the GRU at time t represents linear interpolation among the current \(\widetilde{{h}_{t}}\) and previous \({h}_{t-1}\) activation states in Eqs. (8) and (9).

A Bi-GRU layer was formed by combining a forward GRU with a reverse-direction GRU. Both GRUs receive the same input but train in opposite directions, and their results are concatenated to produce the output. Deep hierarchical neural networks effectively capture specific functions and model dependencies of varying lengths21. Our experiments revealed that Bi-GRU outperformed other models on our datasets.

In our proposed method, the deep learning model comprises input data, a one-time series neural network layer used for sequential predictions, and a dense layer. The neural network structures for LSTM and Bi-GRU are shown in Fig. 2.

Architectures of the (a) long short-term memory (LSTM) method and (b) bidirectional gated recurrent unit (Bi-GRU) method. The input layers of both models consisted of time displacement values in days, reliability data, and visual field data. Reliability data consisted of false-positive (FP) rate, false-negative (FN) rate, and fixation loss (FL) percentage. Visual field data consisted of 52 pattern deviation values (PDVs) and 52 total deviation values (TDVs) on the 24-2 visual field test (two points of physiological scotoma were excluded). The last cell contained a positive time displacement value and 107 zeros as input because all other values were set to zero. These unique inputs can specify the exact date the user wants to predict. LSTM = long short-term memory; Bi-GRU = bidirectional gated recurrent unit; TDV = total deviation value.

The single-layer time-series neural network consists of six parallel and connected LSTM or Bi-GRU cells. The detailed structures of the LSTM and GRU cells are presented in Supplementary Fig. S1a, b, respectively.

Each of the first five cells uses 108 features as input, including 52 total deviation values (TDVs), 52 pattern deviation values (PDVs), reliability data (such as false-negative and false-positive rates, fixation loss percentage), and time displacement value. To improve the performance of the deep learning model, the input data were normalized to a reasonable range. The TDV, PDV, and time displacement values were divided into sets of 50, 50, and 1000, respectively. Time displacement indicated the number of days from the most recent visual field examination. For example, if the most recent visual field examination has a time displacement of “0,” the visual field examination performed 1 month (− 31 days) prior to “0” has a time displacement of “ − 31.” A negative sign in the time displacement value indicates that the examination was performed in the past. With respect to the 6 consecutive visual field input data elements, the last input data element used a unique format with positive time displacement (i.e., the point in the future that the user wishes to predict) and 107 zeros. Since the other data were set to 0, these unique inputs can specify the exact date which the user wishes to predict. A series of input data was arranged by reducing the time displacement value (i.e., from future to past) and then supplying this information to the neural network. Subsequently, the neural network layer was connected to the next single fully connected layer (dense layer) with 52 neurons. These neurons generated a final output of 52 TDVs, such that one neuron generated a single visual field test point.

The root mean square error (RMSE) and mean absolute error (MAE) of the TDV were used as accuracy metrics. The RMSE was calculated for each eye using the following equation:

The MAE was calculated for each test point in the visual field of all eyes using the following equation:

The RMSE and MAE of the LR, LSTM, and Bi-GRU models were calculated using the above formulas. Repeated measures one-way analysis of variance was performed to compare accuracy metrics among LR, LSTM, and Bi-GRU models. P < 0.05 (single comparison) and p < 0.017 (multiple comparisons) were considered indicative of statistical significance. Parametric and nonparametric tests (Spearman’s correlation and simple LR analyses) were performed to compare variables. These tests were used to investigate prediction error trends according to various factors, including false positive rate, false negative rate, fixation loss percentage, and visual field mean deviation (MD).

Table 2 shows the demographic characteristics of the test dataset. The most common diagnosis was primary open-angle glaucoma (47.68%). The mean prediction time (time interval between prediction and final visual field examination) was 1.00 ± 0.84 years (Table 1). The mean RMSE and pointwise mean absolute error (PMAE) are shown in Table 3. Figure 3 presents representative examples of the PMAE in the visual field test.

Representative examples of visual field prediction according to mean deviation (MD) of the first visual field examination. Five consecutive input visual field examinations are shown in chronological order from left to right, followed by the sixth examination (regarded as the true value). Columns 7–9 indicate the prediction results of LR, LSTM, and Bi-GRU models, respectively. LR = linear regression; LSTM = long short-term memory; Bi-GRU = bidirectional gated recurrent unit.

Bi-GRU exhibited better prediction performance, compared with LR and LSTM. The RMSEs of Bi-GRU, LR, and LSTM were 3.71 ± 2.42, 4.81 ± 3.89, and 4.06 ± 2.61 dB, respectively. There were statistically significant differences in prediction errors among the three models (F = 42.94, p < 0.001). The RMSE was significantly lower for Bi-GRU than for the other two models (both p < 0.001).

The number of eyes binned according to RMSE prediction error is shown in Fig. 4. More than 50% of eyes had Bi-GRU prediction errors of ≤ 2 dB (530 eyes, 41.67%) and 2–3 dB (175 eyes, 13.76%). The corresponding LR prediction errors were ≤ 2 dB (329 eyes, 25.86%) and 2–3 dB (254 eyes, 19.97%), and the corresponding LSTM prediction errors were ≤ 2 dB (505 eyes, 39.70%) and 2–3 dB (165 eyes, 12.97%).

Number of eyes binned according to prediction error (RMSE, root mean squared error).

Figure 5 shows the PMAE in the visual field. With respect to the 52 TDV points, Bi-GRU exhibited the lowest prediction error among the three models. Bi-GRU showed significantly better performance at 29 (red dots) and 49 (blue dots) points compared with LR and LSTM, respectively.

Pointwise mean absolute error (PMAE) of predicted total deviation value (TDV). Bi-GRU had the lowest prediction error (PMAE) for all 52 points. Darker colors indicate higher error. Red dots indicate significant differences between LR and Bi-GRU; blue dots indicate significant differences between LSTM and Bi-GRU (paired t-test). LR = linear regression; LSTM = long short-term memory; Bi-GRU = bidirectional gated recurrent unit.

Table 4 shows the mean prediction error (RMSE) according to sectors of the visual field examination (Fig. 6). The 24-2 visual field was divided into the six sectors proposed by Garway-Heath et al.,22 based on optic nerve head anatomy (superotemporal, superonasal, temporal, nasal, inferotemporal, and inferonasal) [Fig. 6b] and two sectors (central and peripheral) [Fig. 6c]. The prediction errors of Bi-GRU were significantly lower than the errors of LR and LSTM for all sectors (p ≤ 0.001).

Division of the optic nerve head (a) and visual field (b, c). (b) The visual field was divided into six sectors proposed by Garway-Heath et al.22 (c) The visual field was divided into the central and peripheral zones. ST = superotemporal; SN = superonasal; T = temporal; N = nasal; IT = inferotemporal; IN = inferonasal; P = peripheral; C = central.

The mean RMSE values binned according to various factors are listed in Table 5 and Fig. 7. The prediction error was significantly lower for Bi-GRU than for the other two models in terms of the false-positive rate, false-negative rate, and fixation loss percentage (p ≤ 0.025). As the visual field MD increased, the RMSE prediction errors of all three models decreased.

Average prediction error (RMSE) binned according to various factors. RMSE vs. (a) false positive rate, (b) false negative rate, (c) fixation loss percentage, and (d) visual field mean deviation (MD). Bi-GRU showed the lowest prediction error. LR = linear regression; LSTM = long short-term memory; Bi-GRU = bidirectional gated recurrent unit; RMSE = root mean squared error.

The correlation coefficients and LR analyses between the prediction error and various factors are presented in Table 6 and Fig. 8. For all models, RMSE was positively correlated with the false-negative rate and fixation loss percentage, whereas it was negatively correlated with visual field MD (all p ≤ 0.029) (Fig. 8).

Linear regression analysis between prediction error (RMSE) and various factors. RMSE vs. (a) false positive rate, (b) false negative rate, (c) fixation loss percentage, and (d) visual field mean deviation (MD). LR = linear regression; LSTM = long short-term memory; Bi-GRU = bidirectional gated recurrent unit; RMSE = root mean squared error.

To the best of our knowledge, this study is the first to utilize the Bi-GRU architecture for predicting visual field loss. We compared the prediction of visual field loss using the Bi-GRU, LR, and LSTM models. The Bi-GRU model demonstrated the highest predictive accuracy among the three models. The overall prediction errors (RMSEs) of the LR, LSTM, and Bi-GRU models were 4.81 ± 3.89, 4.06 ± 2.61, and 3.71 ± 2.42 dB, respectively. The RMSE significantly differed between Bi-GRU and the other models (p < 0.001).

In the six sectors of the visual fields according to optic nerve head anatomy, as well as the central and peripheral visual field areas, Bi-GRU exhibited superior performance compared with the other two models (all p < 0.001).

The predictive performance was negatively correlated with the false-negative rate and fixation loss percentage in all three models; however, Bi-GRU was least affected by reliability indices. A decrease in MD was associated with lower prediction performance in all three models. The RMSE was lowest for Bi-GRU among the three models; Bi-GRU performed better even in patients with advanced glaucoma.

Several studies have used artificial intelligence to detect glaucoma and its progression. Asaoka et al.23 built a deep feed-forward neural network to detect preperimetric glaucoma. The area under the receiver operating characteristic curve (AUROC) of the model was 92.6%, indicating better performance than other machine learning methods (e.g., random forest, gradient boosting, support vector machine, and neural network). Although that study was the first to use deep learning for the evaluation of preperimetric glaucoma, only a small quantity of data from preperimetric visual fields of patients with glaucoma (53 eyes) were analyzed. Elze et al.24 classified visual fields into 16 archetypes and found that the archetypes were closely correlated with the clinical features of glaucoma25. However, these studies classified visual fields, rather than predicting visual field changes. Yousefi et al.26 compared various machine learning algorithms in terms of detecting glaucoma progression, using the retinal nerve fiber layer on optical coherence tomography and the MD and pattern standard deviation on visual field examination as input. The random forest classifier showed the best performance, with an AUROC of 0.88. Wang et al.4 assessed the predictive ability for visual field changes using archetypes; they found that the mean hit and correct rejection rates were 0.77 and 0.77, suggesting that the predictive ability of the archetype approach was higher than the abilities of other methods, such as MD slope, advanced glaucoma intervention study scoring, collaborative initial glaucoma treatment study scoring, and the permutation of pointwise linear regression. However, unlike our study, previous studies did not predict visual field changes.

Dixit et al.14 found that the progression of visual field changes using a deep learning algorithm based on LSTM architecture could be predicted with an accuracy of 91–93%. The AUROC was 0.89–0.93 when using multiple visual field examinations and baseline clinical data as input. Additionally, the use of clinical data to supplement the visual field data led to improved model performance. Murata et al.5 found that variational Bayes linear regression more accurately predicted the progression of visual field changes in patients with glaucoma, compared with conventional least-squares LR. Wen et al.6 used Cascade-Net, a type of convolutional neural network architecture, to predict future Humphrey visual field findings using only a single visual field input. The models showed excellent predictive abilities; the overall PMAE and RMSE were 2.47 and 3.47 dB, respectively. The PMAE and RMSE of the Bi-GRU model were slightly higher than the PMAE and RMSE of the Cascade-Net model. However, this model may not reflect true progression because the authors used single visual field examination as input. Berchuck et al.7 used a generalized variational autoencoder algorithm to estimate progression rates and predict future visual fields. The overall MAE was 1.89–2.33 dB, comparable with the MAE of our model. Park et al.13 used an RNN to predict the sixth visual field examination; they found that the RMSE was 4.31 ± 2.4 dB, indicating that RNN predicted future visual field better than LR.

In a previous study, we used the LSTM model to analyze time-sequential input consisting of visual field examinations13. In the present study, we built a deep learning architecture based on a Bi-GRU network. Both GRU and LSTM are variants of RNN, a state-of-the-art deep learning architecture that processes sequential data for sequence recognition and prediction27. Cho et al.16 presented a GRU architecture that allowed each recurrent unit to adaptively capture dependencies of different time scales. Both GRU and LSTM have recurrent units in sequence modeling. However, GRU has gating units that modulate the flow of information inside the unit without separate memory cells8,12,16. Chung et al.12 reported that GRU was comparable with LSTM for polyphonic music modeling and speech signal modeling. Khandelwal et al.17 found that GRU outperformed LSTM in terms of shorter computation time and lower word error rate for automatic speech recognition.

Conventional RNN only considers the previous context of training data. To overcome the limitations of a conventional RNN, Shuster et al.20 proposed a bidirectional RNN that considers both past and future input sequences to estimate the output vector. Several studies have shown that Bi-GRU outperforms LSTM15,17,18. Bi-GRU achieved the highest classification accuracy among deep neural network-based models for human identification based on electrocardiogram biometrics15.

In the present study, Bi-GRU exhibited better predictive performance than LR and LSTM for the entire visual field, as well as the central area; this area is important because the preservation of central visual function has a strong effect on quality of life in patients with glaucoma28,29. Bi-GRU was least affected by reliability indices. The false-negative rate and fixation loss affected visual field prediction in all models. However, there was poor correlation between fixation loss and visual field prediction, indicating a small effect of fixation loss. Previous studies showed that false-negative rates, but not fixation loss, were associated with visual field assessment13,30,31. Moreover, previous studies revealed that false-negative rates were the most common cause of unreliable visual field classification32,33.

Our study had several limitations. First, the study results cannot be fully generalized to patients with different degrees of glaucoma severity. The study included a greater number of patients with early glaucoma (MD >  − 6 dB) in the training and test datasets, compared with patients who had advanced glaucoma. Although this difference may have affected the performance of Bi-GRU model learning, it reflects the distribution of glaucoma severity observed in clinical practice.

Second, we did not include clinical data for training, in contrast to the work by Dixit et al.14 Future studies should improve deep learning architecture by adding clinical characteristics to the input data.

Third, we trained and tested the model using five consecutive visual field data elements as input. Glaucoma specialists recommend that at least five serial visual field examinations are used to detect glaucoma progression. The Glaucoma Progression Analysis included in the Humphrey Visual Field Analyzer requires at least five reliable visual field examinations and a follow-up period of 2 years34. Previous studies also used five visual field data elements as input to predict visual field progression in glaucoma35,36. Additionally, sequential pointwise LR was performed with at least four visual field examinations because regression analysis is unlikely to detect a trend when fewer data are available37. We predicted the sixth visual field examination using the previous five examinations to compare the predictive performances of Bi-GRU and LR models. Glaucoma requires lifelong periodic visual field examinations38,39. Thus, five consecutive visual field examinations over 3 years are not an excessively frequent number, and the prediction of subsequent examinations based on the initial five examinations may enhance patient convenience.

On further analysis, we predicted future visual field based on four consecutive visual field data elements using the Bi-GRU model. The mean prediction errors were 3.84 ± 2.48 and 2.91 ± 1.96 dB for RMSE and PMAE, respectively. Although there were statistically significant differences in prediction errors (both p < 0.001) between the models using five and four visual field data elements, the difference was not clinically significant.

Fourth, the model could only predict the sixth visual field examinations. Future studies should collect additional patient data with a greater number of visual field examinations and evaluate the performance of our model in terms of predicting the seventh through tenth visual field examinations, using the first five visual field examinations as input. However, our model can forcast visual fields at future time points. For example, the model can predict the visual fields at 4, 8, and 12 months after the fifth visual field examination.

In summary, a deep learning architecture using the Bi-GRU model, a variant of RNN, predicts future visual field examinations significantly better than the pointwise LR and LSTM models. The Bi-GRU model is less affected by the reliability indices of visual field input data. This model may facilitate decision-making by accurately predicting future visual field examinations in clinical practice, particularly for patients who experience difficulty with repeated examinations.

The data generated or analyzed during this study are available from the corresponding author (J.R.P.) upon reasonable request.

Resnikoff, S. et al. Global data on visual impairment in the year 2002. Bull. World Health Organization 9 (2004).

Weinreb, R. N., Aung, T. & Medeiros, F. A. The pathophysiology and treatment of glaucoma: A review. JAMA 311, 1901 (2014).

Article PubMed PubMed Central Google Scholar

Henson, D. B., Chaudry, S., Artes, P. H., Faragher, E. B., & Ansons, A. Response variability in the visual field: Comparison of optic neuritis, glaucoma, ocular hypertension, and normal eyes. 41, 5 (2000).

Wang, M. et al. An artificial intelligence approach to detect visual field progression in glaucoma based on spatial pattern analysis. Invest. Ophthalmol. Vis. Sci. 60, 365 (2019).

Article PubMed PubMed Central Google Scholar

Murata, H., Araie, M. & Asaoka, R. A new approach to measure visual field progression in glaucoma patients using variational bayes linear regression. Invest. Ophthalmol. Vis. Sci. 55, 8386–8392 (2014).

Article PubMed Google Scholar

Wen, J. C. et al. Forecasting future Humphrey Visual Fields using deep learning. PLoS ONE 14, e0214875 (2019).

Article CAS PubMed PubMed Central Google Scholar

Berchuck, S. I., Mukherjee, S. & Medeiros, F. A. Estimating rates of progression and predicting future visual fields in glaucoma using a deep variational autoencoder. Sci Rep 9, 18113 (2019).

Article CAS PubMed PubMed Central ADS Google Scholar

Salehinejad, H., Sankar, S., Barfett, J., Colak, E., Valaee, S. Recent advances in recurrent neural networks. 21.

Liu, S., Yang, N., Li, M. & Zhou, M. A recursive recurrent neural network for statistical machine translation. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 1491–1500 (Association for Computational Linguistics, 2014). https://doi.org/10.3115/v1/P14-1140.

Young, T., Hazarika, D., Poria, S. & Cambria, E. Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag. 13, 55–75 (2018).

Article Google Scholar

Hochreiter, S. & Schmidhuber, J. Long short-term memory. Neural Comput. 9, 1735–1780 (1997).

Article CAS PubMed Google Scholar

Chung, J., Gulcehre, C., Cho, K., & Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. (2014) https://doi.org/10.48550/ARXIV.1412.3555.

Park, K., Kim, J. & Lee, J. Visual field prediction using recurrent neural network. Sci Rep 9, 8385 (2019).

Article PubMed PubMed Central ADS Google Scholar

Dixit, A., Yohannan, J. & Boland, M. V. Assessing glaucoma progression using machine learning trained on longitudinal visual field and clinical data. Ophthalmology 128, 1016–1026 (2021).

Article PubMed Google Scholar

Lynn, H. M., Pan, S. B. & Kim, P. A deep bidirectional gru network model for biometric electrocardiogram classification based on recurrent neural networks. IEEE Access 7, 145395–145405 (2019).

Article Google Scholar

Cho, K. et al. Learning phrase representations using RNN encoder–decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 1724–1734 (Association for Computational Linguistics, 2014). https://doi.org/10.3115/v1/D14-1179.

Khandelwal, S., Lecouteux, B. & Besacier, L. Comparing GRU and LSTM for Automatic Speech Recognition. 7.

Li, X. et al. Time-series production forecasting method based on the integration of Bidirectional Gated Recurrent Unit (Bi-GRU) network and Sparrow Search Algorithm (SSA). J. Petrol. Sci. Eng. 208, 109309 (2022).

Article CAS Google Scholar

Darmawahyuni, A., Nurmaini, S., Rachmatullah, M. N., Firdaus, F. & Tutuko, B. Unidirectional-bidirectional recurrent networks for cardiac disorders classification. TELKOMNIKA 19, 902 (2021).

Article Google Scholar

Schuster, M. & Paliwal, K. K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45, 2673–2681 (1997).

Article ADS Google Scholar

Pascanu, R., Gulcehre, C., Cho, K. & Bengio, Y. How to construct deep recurrent neural networks. (2013). https://doi.org/10.48550/ARXIV.1312.6026.

Garway-Heath, D. F., Poinoosawmy, D., Fitzke, F. W. & Hitchings, R. A. Mapping the visual field to the optic disc in normal tension glaucoma eyes. 107, 7 (2000).

Asaoka, R., Murata, H., Iwase, A. & Araie, M. Detecting preperimetric glaucoma with standard automated perimetry using a deep learning classifier. Ophthalmology 123, 1974–1980 (2016).

Article PubMed Google Scholar

Elze, T. et al. Patterns of functional vision loss in glaucoma determined with archetypal analysis. J. R. Soc. Interface. 12, 20141118 (2015).

Article PubMed PubMed Central Google Scholar

Cai, S. et al. Clinical correlates of computationally derived visual field defect archetypes in patients from a glaucoma clinic. Curr. Eye Res. 42, 568–574 (2017).

Article PubMed Google Scholar

Yousefi, S. et al. Detection of longitudinal visual field progression in glaucoma using machine learning. Am. J. Ophthalmol. 193, 71–79 (2018).

Article PubMed Google Scholar

Bengio, Y., Simard, P. & Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 5, 157–166 (1994).

Article CAS PubMed Google Scholar

Johnson, C. A. & Nelson-Quigg, J. M. A prospective three-year study of response properties of normal subjects and patients during automated perimetry. Ophthalmology 100, 269–274 (1993).

Article CAS PubMed Google Scholar

Katz, J., Sommer, A. & Witt, K. Reliability of visual field results over repeated testing. Ophthalmology 98, 70–75 (1991).

Article CAS PubMed Google Scholar

Murata, H. et al. Identifying areas of the visual field important for quality of life in patients with glaucoma. PLoS ONE 8, e58695 (2013).

Article CAS PubMed PubMed Central ADS Google Scholar

Abe, R. Y. et al. The impact of location of progressive visual field loss on longitudinal changes in quality of life of patients with glaucoma. Ophthalmology 123, 552–557 (2016).

Article PubMed Google Scholar

Rao, H. L. et al. Role of visual field reliability indices in ruling out glaucoma. JAMA Ophthalmol 133, 40 (2015).

Article PubMed Google Scholar

Raman, P., Khy Ching, Y., Sivagurunathan, P. D., Ramli, N. & Mohd. Khalid, K. H. The Association between visual field reliability indices and cognitive impairment in glaucoma patients. J. Glaucoma 28, 685–690 (2019).

Casas-Llera, P. et al. Visual field index rate and event-based glaucoma progression analysis: Comparison in a glaucoma population. Br. J. Ophthalmol. 93, 1576–1579 (2009).

Article CAS PubMed Google Scholar

Crabb, D. P., Fitzke, F. W., McNaught, A. I., Edgar, D. F. & Hitchings, R. A. Improving the prediction of visual field progression in glaucoma using spatial processing. Ophthalmology 104, 517–524 (1997).

Article CAS PubMed Google Scholar

Bengtsson, B. Prediction of glaucomatous visual field loss by extrapolation of linear trends. Arch Ophthalmol 127, 1610 (2009).

Article PubMed Google Scholar

Nouri-Mahdavi, K. Comparison of methods to predict visual field progression in glaucoma. Arch Ophthalmol 125, 1176 (2007).

Article PubMed Google Scholar

European Glaucoma Society Terminology and Guidelines for Glaucoma, 5th Edition. Br. J. Ophthalmol. 105, 1–169 (2021).

Prum, B. E. et al. Primary open-angle glaucoma preferred practice pattern® guidelines. Ophthalmology 123, P41–P111 (2016).

Article PubMed Google Scholar

Download references

This research was supported by grants from Medical Big Data; AI-based early detection of visual dysfunction, funded by Busan and managed by Busan Techno Park; Patient-Centered Clinical Research Coordinating Center, funded by the Ministry of Health & Welfare, Republic of Korea (Grant Nos.: HI19C0481 and HC19C0276); and National Research Foundation (NRF) Korea, funded by the Korean government (Grant Nos.: NRF-2021R1I1A1A01057767, NRF-2021R1A2B5B03087097, NRF-2017R1A5A1015722, and NRF-2022R1A5A1033624). We thank Textcheck (textcheck.com) for English language editing.

These authors contributed equally: Hwayeong Kim and Jiwoong Lee

Department of Ophthalmology, Pusan National University College of Medicine, Busan, Korea

Hwayeong Kim, Jiwoong Lee & Sangwoo Moon

Biomedical Research Institute, Pusan National University Hospital, Busan, Korea

Jiwoong Lee

Department of Mathematics, Pusan National University, Busan, Republic of Korea

Sangil Kim, Taehyeong Kim & Yuanmeng Hu

Department of Ophthalmology, Dong-A University College of Medicine, Busan, Korea

Sang Wook Jin

Department of Ophthalmology, Busan Paik Hospital, Inje University College of Medicine, Busan, Korea

Jung Lim Kim

Department of Ophthalmology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan, Korea

Jonghoon Shin

Department of Ophthalmology, Kosin University College of Medicine, Busan, Korea

Seung Uk Lee

Nonlinear Dynamics and Mathematical Application Center, Kyungpook National University, Daegu, Korea

Geunsoo Jang

Department of Mathematics, Kyungpook National University, 80, Daehak-ro, Buk-gu, Daegu, 41566, Republic of Korea

Jeong Rye Park

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

J.R.P. built the artificial intelligence architecture, analyzed the data, and wrote the manuscript. H.Y.K. and J.W.L. also analyzed the data and wrote the manuscript. T.H.K. and Y.U. built the artificial intelligence architecture and wrote the manuscript. S.I.K. and G.S.J. analyzed the data. S.W.J., J.L.K., J.H.S., S.U.L., and S.W.M. collected the data. All authors discussed the results and commented on the manuscript.

Correspondence to Jeong Rye Park.

The authors declare no competing interests.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

Kim, H., Lee, J., Moon, S. et al. Visual field prediction using a deep bidirectional gated recurrent unit network model. Sci Rep 13, 11154 (2023). https://doi.org/10.1038/s41598-023-37360-1

Download citation

Received: 15 July 2022

Accepted: 20 June 2023

Published: 10 July 2023

DOI: https://doi.org/10.1038/s41598-023-37360-1

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.