Abstract
Introduction
Visual acuity (VA) testing is a critical screening examination for the assessment of visual function. This study describes the development and validation of a smartphone-based VA test: the K-VA test.
Methods
A total of 171 patients with various ocular diseases were examined in our outpatient unit at Aristotle University of Thessaloniki, School of Medicine in Greece. Participants underwent VA examination using the standard Early Treatment Diabetic Retinopathy Study (ETDRS) charts and the K-VA smartphone-based test. The K-VA test was performed by participants themselves. The Bland–Altman method was employed to assess the agreement between the ETDRS charts and the new test for the examination of VA at 1 m and 40 cm. Test–retest reliability was also calculated. A questionnaire regarding the participants’ feedback on the K-VA test was completed.
Results
No significant bias was observed between the gold standard ETDRS charts and the K-VA test measurements. The mean difference (95% limits of agreement, LoA) between the K-VA test at 1 m and the ETDRS chart at 4 m was −0.006 (95% LoA −0.129 to 0.117) logarithm of the minimal angle of resolution (logMAR). The agreement of the K-VA test at 40 cm with the near ETDRS chart was also high with a mean difference of −0.007 (95% LoA −0.105 to 0.090) logMAR. Test–retest reliability was found to be high with a mean difference of 0.003 (95% LoA −0.045 to 0.033) logMAR and 0.005 (95% LoA −0.065 to 0.076) logMAR for the K-VA test at 1 m and 40 cm, respectively. A total of 97 participants answered the questionnaire and 71 (73.2%) stated that the test was easy to very easy to use for self-performance.
Conclusions
The study demonstrated that the K-VA application performed well compared with the ETDRS charts and provides reliable and repeatable measurements of VA across a wide range of VA.
Plain Language Summary
Visual acuity (VA) is the most significant measure of visual function and an accurate examination of VA is of high importance for clinicians. Traditionally, the Snellen and the Early Treatment Diabetic Retinopathy Study (ETDRS) acuity charts are the most commonly used VA tests. However, the examination requires physical presence, which is not always feasible for elderly, mobility-impaired, rural patients, or patients in quarantine due to coronavirus disease 2019 (COVID-19) pandemic. Currently, there are hundreds of vision-testing applications available, with only a few having been validated, especially against the standard ETDRS chart. This study describes the development and investigates the validity of a new smartphone application, the K-VA test, for VA testing in 171 participants with various ocular diseases. Participants performed the K-VA test themselves with the supervision of a healthcare provider. The K-VA test was shown to perform well when compared with the ETDRS charts. We also showed that the test is well accepted by patients familiar with the Android operating system. The test should be performed following the instructions of the attending physician. The K-VA application provides repeatable and reliable measurements of VA across a wide range of VA. This tool offers users a secure system for the tele-transmission of results to physicians for further consultation.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Several barriers to physical attendance for visual acuity examination, such as increased age, difficulty in transportation, mobility issues, or pandemic crisis, could be overcome with the development of accurate vision-testing applications. |
We designed and examined the validity of a new smartphone-based visual acuity test based on the standard ETDRS principles: the K-VA test. |
The K-VA application performed well when compared with the standard ETDRS chart and provides repeatable and reliable measurements of VA across a wide range of VA. |
The results of the K-VA test are automatically recorded and can be transmitted to the attending physician for further evaluation. |
Introduction
Visual acuity (VA) is the most frequently performed ophthalmic examination of the visual function. The accurate evaluation of VA provides quantification of changes in central vision caused by various ocular diseases. Hence, VA examination is considered the most common primary outcome measure of ophthalmic clinical trials [1].
Vision impairment (VI) has a great impact on the quality of life of affected individuals [2] and is considered a major public health issue [3]. According to the World Health Organization (WHO), 285 million people worldwide suffer from some grade of VI, the majority of whom have poor access to eye care services [4]. The definition of VI is usually in terms of reduced VA, highlighting the central role of this test in the evaluation of ophthalmological patients.
There are several VA charts in use with the Snellen chart since it was developed in the 1860s, being the most widely available and subsequently used vision chart worldwide [5]. However, several characteristics of this chart, including the nongeometric progression in the size of the displayed letters and the variable number of letters per line, limit its use for research purposes and the evaluation of patients with low vision (LV) [6, 7]. Recently, the gold standard for the accurate evaluation of VA has been considered to be the Early Treatment Diabetic Retinopathy Study (ETDRS) acuity chart that is widely used in ophthalmic research [8]. This type of chart is recommended since it improves the precision and reproducibility of VA measurement at the penalty of significantly longer test duration [8]. The examination with these charts requires the physical presence of the examinee, which is not always possible for elderly, mobility-impaired, or rural patients, or patients in quarantine due to coronavirus disease 2019 (COVID-19) pandemic [9].
In ophthalmology, a promising approach includes the use of smart devices such as smartphones and tablet computers to perform portable assessments, screening, and evaluation of the visual function to address accessibility issues [9]. Currently, there are hundreds of vision-testing applications available; however, only a few have been validated and compared against the standard ETDRS chart [10,11,12,13]. Most of the existing validated applications for VA testing, such as the Eye Chart pro, the Eye Hand Book, and the Odysight, lack a significant index of validity, that is, the test–retest analysis [14,15,16,17]. The Peek Acuity test is a successfully implemented application for VA screening in school children and low-income areas; however, the testing procedure requires a tester to translate the gestures of the examinee [18]. The Easee eye test is an interactive, web-based VA test with proved accuracy in various study groups; though, the tool has not been evaluated in older patients, who are less familiar with digital devices [19, 20]. Furthermore, the inability of all patients to navigate such technology themselves highlights the need for appropriate instructions and remote consultation by their attending physician [10,11,12,13].
Within this context, a valid, repeatable, and interactive smartphone-based VA test could serve as a beneficial tool for both patients and their physicians. This study aimed to describe the development and validation of an interactive, smartphone-based VA test for self-performance, following the design principles of the standard ETDRS chart.
Methods
Study Participants
A total of 171 Greek-speaking participants (n = 250 eyes) with VI due to various ocular diseases were recruited from our outpatient unit at Aristotle University of Thessaloniki, School of Medicine in Greece. Participants were randomly selected among individuals who were referred to our unit and attended the LIFE4LV research program for patients with VI. The trial is registered at clinicaltrials.gov (NCT05184036). Patients were previously diagnosed and referred with their medical record. Most of them had undergone cataract surgery or were free of significant lenticular changes, based on their medical record. Participants with Best Corrected Visual Acuity (BCVA) levels ranging from logMAR 0.0 to 1.0 were included. Informed written consent was obtained from all subjects after an explanation of the nature of the study. Exclusion criteria included severe cognitive impairment as assessed by an expert psychologist with the Mini-Mental State Examination [21] and severe systemic comorbidities (Parkinson’s disease, etc.). Illiterate individuals, patients faced linguistic issues, or those who had a non-Greek primary language were also excluded from the present study. However, the selected letters could also be potentially used in English (already available by the application) or other languages with the appropriate modification. All study procedures were approved by the Committee for Bioethics and Ethics, Medical Department, Aristotle University of Thessaloniki (code#1.60/21.11.2018) and adhered to the principles embodied in the Declaration of Helsinki Code of Ethics of the World Medical Association.
Digital Visual Acuity Test
The K-VA test was developed as a smartphone-based application and is available for devices running Android OS 5.0 or later. The application has been created with Android Studio V4.1 and the software provided various options on the smartphone screen. The application has been demonstrated with a Samsung A30S smartphone (display: Super AMOLED, size 6.4 inches, resolution 720 × 1560 pixels, ratio 19.5:9, density ~ 268 ppi; GPU: Mali-G71 MP2) running Android OS 9.0 with maximal brightness of 489 cd/m2 (personal communication with Samsung technical support), and the VA testing was conducted using full brightness (100%). The device was switched on at least 5 min before each experimental session to allow its output to stabilize.
The K-VA test follows the standard ETDRS chart design [7, 22]. Monocular VA can be measured, with the appropriate correction, at 1 m and 40 cm, recorded, and saved following the General Data Protection Regulation (GDPR), and the physician has access to the results for further evaluation (Fig. 1).
The device should have a stable internet connection, and participants were instructed to read aloud letter by letter and were encouraged to guess the letter if not sure, as per the ETDRS protocol [1]. Each line contains five letters and the letter that should be spelled is pointed to guide the patient (Fig. 1). The letter contrast of the neighboring letters to the letter that should be pronounced was slightly lower to guide the user and to avoid losing their focus (Fig. 1). According to Chung and Mansfield, altering the contrast polarity of neighboring letters does not reduce the crowding effect since local uncrowding of individual letters seems to be counteracted by the grouping of letters [23,24,25]. Once the letter is read, either wrong or right, patients feel a vibration (40 cm examination) of the phone to better understand that they should read the next letter in the row. A speech recognizer records if the letter was recognized correctly or not. The crowding effect of the standard ETDRS chart was achieved by using a crowding bar at the right and left sides of each row of letters (Fig. 1).
As mentioned before, appropriate instructions written in a font corresponding to VA of 1.0 logMAR are shown on the landing page of the application, informing patients on how to perform the test correctly. Participants held the smartphone at 40 cm distance or supported the device with the back of a book at 1 m distance at a perpendicular angle, at the line of sight, to avoid reflection of light. In the case of an interruption, the application asks the users if they wish to continue or to repeat the examination. There is also an opportunity for font selection in cases of patients with visual acuity below 1.0 logMAR for the examination at 40 cm.
A clear statement is also shown before testing to inform users that this application is a nonclinical setting and should not replace the VA examination by an eye care professional. Instead, this application could be used as a useful tool by physicians to perform an immediate and accessible, remote examination.
For this test, a single-letter scoring methodology [26] was utilized and the score was converted to the logarithm of the minimal angle of resolution (logMAR) visual acuity. The application contains a set of letters that are commonly used among most European alphabets and can also be recognized by patients who are not familiar with some letters of the Roman alphabet [19, 27,28,29,30]. These are: A, E, H, Z, N, O, P, T, Y, X. The new letters were constructed on Snellen’s 5 × 5 grid and are compatible with the specification of the Sloan letters [31]. The Sloan letters originate from the Roman alphabet and cannot be used with individuals who are unfamiliar with this alphabet [19, 27, 28]. The new set of letters used in the K-VA application was previously evaluated for similarity with the standard ETDRS chart in 55 individuals; 15 controls and 40 with visual impairment and low vision. No obvious bias was detected between the proposed set of letters and the standard ETDRS chart [32].
A staircase algorithm was used to enhance testing efficiency as follows: if three out of five letters of each sentence are correctly identified, the next sentence representing the next size of letters is automatically presented. This sequence continues until the smallest optotype line is recognized or the logMAR 0.0 is reached. If less than three letters are correctly identified, then the test ends and the result is calculated according to the ETDRS algorithm [7, 22]. Results are displayed in logMAR form after testing and can be saved and shared via a secured backend system with the physician. The K-VA test was designed to be used for VA measurements and the monitoring of these measurements by the attending physician remotely. Within this frame, a single user would repeat the examination with the K-VA test in predetermined intervals, according to the physician’s instructions. Thus, letters are randomly presented during each time of examination to minimize the risk of learning and memory effects.
Testing Protocol
All study participants underwent near and distance VA testing twice using the ETDRS charts and the K-VA test for both distances on the same day, using a standardized protocol in a controlled examination environment. The K-VA test was performed by the participants themselves. Participants were supervised by a trained healthcare provider during the K-VA testing to keep the testing distance and to occlude correctly the eye with their palm, avoiding the application of pressure on the eye. An expert ophthalmologist examined participants with their best correction in the ETDRS charts, without knowing the score of the K-VA, which was automatically calculated by the application. Both eyes of the participants were examined.
For conventional VA testing, an externally illuminated 4-m ETDRS VA chart (CAT. No. 2111, LOT. 113529, Precision Vision Inc.) was used for the examination of distance VA at 4 m, and an ETDRS near VA card (CAT. No. 2106B, LOT110326, Precision Vision Inc.) was used at 40 cm distance for the evaluation of near VA. The testing distance for all tests was precisely measured with a measuring string before each examination and a trained healthcare provider observed that participants remained at the appropriate distance throughout the examination.
VA was examined using the K-VA application at 1 m and 40 cm. The participants were asked to follow the instructions presented on the smartphone’s screen and perform the test themselves, monocularly, occluding the other eye with the palm of their hand in the following sequence: right eye first and then left eye (Fig. 2). It takes approximately 2.5 min to perform the test for each eye (time also depends on the visual status of the patient). The K-VA test was performed in office lighting conditions and it is recommended to avoid excessive ambient lighting (sunlight) during the performance of the test [18]. The results were automatically recorded by the application.
For ETDRS testing, participants were asked to read aloud the letters until the smallest row of letters in which they could identify less than three out of five letters, and results were recorded by the ophthalmologist. All participants were retested on the same day.
In total, participants conducted eight VA examinations: four with the K-VA test, at 1 m and 40 cm (test and retest for each distance), and four with the standard ETDRS at 4 m and 40 cm (test and retest for each distance). Whether ETDRS or the K-VA was tested first was decided at random, using an online randomization program (www.randomizer.org).
Questionnaire Feedback
A total of 97 participants, all Android users, were asked to answer a short questionnaire (Supplementary Material) regarding the ease of use of the test and their familiarity with such technologies (smartphones, tablet computers, etc.). Participants were asked to report the kind of mobile phone they used before answering the feedback questionnaire; those who were non-Android users were excluded from the feedback section to avoid possible effects of the unfamiliarity with the Android operating system in the feedback [33, 34]. The non-responders were users of different devices; in specific, 64 of them were iOS users and 10 were feature phone users.
Statistical Analysis
All VA measurements were converted to logMAR units [26]. Bland–Altman plots were employed to assess the pairwise agreement between the ETDRS charts and the K-VA test at 1 m and 40 cm examination distance. For most individuals, measurements for both eyes were taken and used in the analysis. The SimplyAgree [35] package in R, version 4.1.2 (R Foundation of Statistical Computing, Vienna, Austria) was used for estimating reliability and agreement between the different measurements while accounting for dependencies within individuals having both eyes included in the analysis.
Regarding the analysis of the feedback questionnaire, the Shapiro–Wilk test was used for the normality assessment of age. Continuous variables were described using median (IQR) while categorical variables were described using frequencies and relative frequencies. The chi-squared test was used to investigate for any dependencies between categorical variables. The Kruskal–Wallis H test was used to explore the differences between the groups of continuous variables. Bonferroni correction was used to adjust for pairwise comparisons. p-Values less than 0.05 were considered statistically significant. SPSS version 27.0 (IBM Corp., Armonk, NY) was used for the statistical analysis of the feedback questionnaire.
Results
In total, 171 patients (n = 250 eyes) were enrolled in the study. The mean age was 67.57 years [standard deviation (SD) 15.51 years, range 18–89 years] and 52% (n = 89) of the participants were female. A total of 53 subjects (31%) had Age-related Macular Degeneration (AMD), 40 (23.4%) had Diabetic Retinopathy (DR), 18 (10.5%) had glaucoma, 10 (5.8%) had Retinitis Pigmentosa (RP), 10 (5.8%) had treated retinal detachment, 9 (5.2%) had retinal vein occlusion, 6 (3.5%) had Stargardt disease, 6 (3.5%) had optic nerve atrophy, 5 (2.9%) had macular hole, 4 (2.3%) had epiretinal membrane, 3 (1.7%) had myopic maculopathy, 2 (1.2%) had cystoid macular edema, 2 (1.2%) had amblyopia, 1 (0.58%) had choroideremia, 1 (0.58%) had Fuchs’ dystrophy, and 1 (0.58%) had Birdshot Chorioretinopathy. The demographic characteristics of participants are presented in Table 1.
The mean differences (95% confidence intervals) for the pairwise comparisons between ETDRS charts and K-VA test at both distances are shown in Fig. 3. The median VA for all measurements (ETDRS chart and K-VA test at both distances) are presented in Supplementary Table S1. The mean values for both eyes were also calculated and are presented in the same table.
The pairwise comparisons between ETDRS charts and K-VA application in measuring distance and near VA across participants are presented in Table 2. Bland–Altman summary statistics of mean bias and 95% Limits of Agreement (LoA) were calculated and are presented in the same table. Results suggest no significant bias between the methods studied. The mean difference between the K-VA test and the ETDRS charts was 0.006 logMAR (95% LoA −0.129 to 0.117) and −0.007 (95% LoA −0.105 to 0.090) for measurement at 1 m and 40 cm, respectively. Bland–Altman scatter plots were constructed to represent the graphical spread of each of the pairs of measurement results. No evidence of systematic association was found between the level of agreement and the underlying acuity (Fig. 4).
The K-VA test–retest reliability was found to be high (Table 2). Bland–Altman summary statistics of mean bias and 95% Limits of Agreement were calculated and are shown in Supplementary Figure S1.
Feedback Questionnaire
Overall, 97 participants, all Android users, answered the questionnaire; 52 were female (53.6%) with a mean age of 67.42 ± 12.97 years. In total, 71 (73.2%) stated that the test was easy to very easy to use, and 40 (41.3%) stated having high to very high familiarity with new technologies (smartphones, tablet computers, etc.). There was a significant association between ease of use and familiarity with smartphones (p < 0.001). Age was found to significantly correlate with ease of use (p = 0.007). These correlations are illustrated in Fig. 5.
Discussion
This study aimed to validate the performance of a new, smartphone-based VA test, the K-VA application, which was designed for the remote monitoring of patients by their physician. Overall, the K-VA test performed well and was proven to be repeatable and well-accepted. The mean difference between the K-VA test and the ETDRS charts was 0.006 logMAR (95% LoA −0.129 to 0.117) and −0.007 (95% LoA −0.105 to 0.090) for measurement at 1 m and 40 cm, respectively, and were within the accepted deviation of ± 0.15 logMAR [15, 16, 36]. The distribution of differences, expressed as 95% LoA, were also within the previously reported limits and indicate acceptable variability of the VA measurements with the K-VA test [16]. When comparing the repeatability of the K-VA test with that of the ETDRS charts, the LoA for the K-VA test–retest reliability were very close to the previously published estimates of test–retest variability (TRV) for the standard ETDRS charts, which ranges between 0.07 and 0.18 logMAR [28, 37,38,39].
The use of smartphones and other wireless technologies is rapidly increasing in clinical practice [40]. The role of mobile applications in delivering and measuring health-related outcomes was especially illustrated due to the coronavirus disease 2019 (COVID-19) pandemic [11, 41, 42]. Currently, there are many vision-testing applications available; however, only a minority of them have undergone rigorous validation, especially against the standard ETDRS chart [43, 44]. Even fewer have undergone test–retest analysis, a fundamental index of validity [15]. Unfortunately, healthcare applications with questionable validity could be misleading both for patients and healthcare providers [45].
In a recent systematic review evaluating the performance of other previously validated VA applications to reference standards, the mean differences ranged from −0.08 to 0.10 logMAR while the distribution of the differences, as expressed by the 95% LoA, ranged from ± 0.08 logMAR (lowest variability) to ± 0.47 logMAR (highest variability) from the mean difference [16]. Furthermore, test–retest variability of logMAR charts ranged from ± 0.07 to ± 0.18 logMAR from the mean difference [16]. Similarly, the K-VA test performed within these ranges, compared with the standard ETDRS charts.
Two important parameters must be taken into account when interpreting results of studies comparing VA assessments. Firstly, even when VA is stable, there is always variation of the repeated VA measurements of the same individual. This type of variation also depends on the visual status of the examinee; use of appropriate correction and better VA lead to more accurate results. For instance, the accuracy of the Easee VA test differed for various VA ranges and study populations (corrected VA versus uncorrected VA) [19]. Reports on test–retest variabilities of commonly used charts show a wide range of 95% LoA, with logMAR charts providing more reliable results than Snellen charts (95% LoA ranging from ± 0.18 to ± 0.33 logMAR) [46]. The Snellen chart illustrates only one or two letters in the poor vision lines and a line assignment method is used for scoring; this means that losing a letter in these lines makes a great impact in the score. However, time and convenience seem to outweigh the accuracy in clinical settings.
Secondly, another source of variation in VA measurements is the type of charts used for comparison [18, 47]. This phenomenon contributes partly to the obtained differences in outcomes between VA applications and the reference VA chart. This was obvious in the case of the Peek Acuity test when different reference charts were used (Snellen versus ETDRS) [18]. Siderov et al. proposed that differences of less than 0.15 logMAR do not indicate an actual change in VA, since they reflect the variability of VA measurements among different settings (different charts, different examiners, etc.) [36]. The high accuracy of the K-VA when compared with the standard ETDRS chart could be attributed to the design of the K-VA test that was based on the standard ETDRS chart’s principles. Specifically, the optotype presented by the K-VA application is very similar to the conventional ETDRS chart. Thus, a conversion effect was not detected in our study.
The use of digital tools for VA assessments may be limited due to the lack of appropriate instructions and remote consultation by the attending physician. The K-VA offers some advantages for self-performance: clear instructions are present prior to examination, the test interacts with the examinee, providing vocal recognition of the letter and a feeling of vibration (40 cm examination) to inform the patients to move to the next letter. The results are automatically recorded in logMAR (Fig. 1) units, minimizing possible technician-based errors. The data are saved and can be securely sent to the attending physician for further consultation. Importantly, since data from every examination are saved, they could be compared with those of a following examination by the physician. Using the K-VA test, the physician may detect changes remotely, indicative of the need for further evaluation of the patient. This could be of high clinical significance for patients who require frequent monitoring of their VA and face accessibility issues.
It is important to note that discrepancies when comparing examination in clinical setting with that in an unsupervised environment may be detected, especially if the test distance has not been accurately measured. In the case of the K-VA test, patients are strongly instructed to measure the testing distance and to maintain the same distance when performing the test (placing a distinctive mark). Clear instructions are provided prior to the test to ensure the consistency of testing conditions. In addition, a large print manual with more detailed instructions for the correct performance of the test is provided with the application (the manual can be sent by e-mail or downloaded with the application). The test distance is fundamental to the accuracy of the results, along with the refractive correction. The fact that the results can be evaluated directly by the attending physician and appropriate consultation can be provided accordingly aims to reduce the likelihood of inaccurate results. The accuracy of another well-established VA test, the Easee VA test, was recently evaluated in unsupervised home environments. The test provided reliable measures of VA in the majority of patients and the study highlighted important aspects on how to ensure the accuracy of the results in home settings [48]. Importantly, in suspicion of invalid outcomes, patients should be instructed to retest under optimal testing conditions following the instructions of the application and the attending physician.
Our study has certain limitations that warrant consideration. By convention, we chose to measure VA at a 1 m distance, a method used before for distance VA examination using smart devices [25]. In some patients, the 4 m and 1 m scores are not directly interchangeable; cataract, amblyopia, and inappropriate refractive correction in patients with high myopia could cause a disagreement between 4 m and 1 m VA scores [49,50,51]. However, a high degree of correlation between these VA measurements has been previously observed [52, 53]. VA less than 1.0 logMAR can only be measured at 40 cm and the examination is limited to 1.3 logMAR at this distance. Moreover, the accuracy of the VA assessment is often dependent on the competency level of the examiner, leading to inter-observer variability [1, 54]. In our study, a single researcher conducted all the measurements with the traditional ETDRS charts [54]. Furthermore, the study was conducted using only one type of smartphone and the application is available for devices running Android OS 5.0 or later. Thus, in the feedback section we included only patients who were Android users to avoid possible effects of unfamiliarity with the Android operating system in the feedback, since users of different operating systems have been previously found to differ in the usage of mobile health applications [55]. We are currently developing the application for the iOS operating system. According to recent data, Android was the leading mobile operating system worldwide with close to 71.8% of users, while iOS accounted for around 27.6% of the mobile operating system market [56, 57]. In our study, these percentages were found to be different; this may reflect differences in the sociodemographic composition of our sample [33, 34].
We should also acknowledge that the testing distance should be kept constant for accurate results. Moreover, previous studies suggest the susceptibility of smartphone-based tests to glare [58]. Keeping the device at a perpendicular angle at the line of sight to avoid reflection of light decreases the effect. Furthermore, the K-VA test was validated in a Greek-speaking population, which limits the generalizability of the results; however, the test utilizes a set of letters easily adaptable to other European languages. The K-VA application was assessed for self-performance at clinical settings. Thus, self-performance at home settings remains to be evaluated.
Conclusions
The study demonstrated that the smartphone-based K-VA test provides reliable and repeatable measurements of VA. This test seems to be a promising tool for quick and easily accessible measurements of VA. Patients who have difficulty accessing specialized healthcare services may benefit from applications such as the K-VA test, and the attending ophthalmologist has an additional tool for the follow-up of patients with accessibility issues. Future research should investigate the accuracy of the tool when performed independently by patients at home.
References
Kaiser PK. Prospective evaluation of visual acuity assessment: a comparison of Snellen versus ETDRS charts in clinical practice (An AOS Thesis). Trans Am Ophthalmol Soc. 2009;107:311–24.
Crews JE, Chou C-F, Zhang X, Zack MM, Saaddine JB. Health-related quality of life among people aged 65 years with self-reported visual impairment: findings from the 2006–2010 behavioral risk factor surveillance system. Ophthalmic Epidemiol. 2014;21:287–96.
Eckert KA, Carter MJ, Lansingh VC, et al. A simple method for estimating the economic cost of productivity loss due to blindness and moderate moderate to severe visual impairment. Ophthalmic Epidemiol. 2015;22:349–55.
Bourne RR, Flaxman SR, Braithwaite T, et al. Magnitude, temporal trends, and projections of the global prevalence of blindness and distance and near vision impairment: a systematic review and meta-analysis. Lancet Glob Health. 2017;5:e888–97.
Hetherington R. The Snellen chart as a test of visual acuity. Psychol Forsch. 1954;24:349–57.
Ferris FL, Kassoff A, Bresnick GH, Bailey I. New visual acuity charts for clinical research. Am J Ophthalmol. 1982;94:91–6.
Bailey JL, Lovie-Kitchin JE. Visual acuity testing. From the laboratory to the clinic. Vis Res. 2013;90:2–9. https://doi.org/10.1016/j.visres.2013.05.004. (ISSN 0042-6989).
Shamir RR, Friedman Y, Joskowicz L, Mimouni M, Blumenthal EZ. Comparison of Snellen and Early Treatment Diabetic Retinopathy Study charts using a computer simulation. Int J Ophthalmol. 2016;9(1):119–23. https://doi.org/10.18240/ijo.2016.01.20.PMID:26949621;PMCID:PMC4768517.
Gower EW, Silverman E, Cassard SD, Williams SK, Baldonado K, Friedman DS. Barriers to attending an eye examination after vision screening referral within a vulnerable population. J Health Care Poor Underserved. 2013;24:1042–52.
Han X, Scheetz J, Keel S, Liao C, Liu C, Jiang Y, Muller A, Meng W, He M. Development and validation of a smartphone-based visual acuity test (vision at home). Trans Vis Sci Tech. 2019;8(4):27. https://doi.org/10.1167/tvst.8.4.27.
Tofigh S, Shortridge E, Elkeeb A, Godley B. Effectiveness of a smartphone application for testing near visual acuity. Eye. 2015;29:1464.
Pathipati AS, Wood EH, Lam CK, Sáles CS, Moshfeghi DM. Visual acuity measured with a smartphone app is more accurate than Snellen testing by emergency department providers. Graefes Arch Clin Exp Ophthalmol. 2016;254:1175–80.
Beck RW, Moke PS, Turpin AH, et al. A computerized method of visual acuity testing: adaptation of the early treatment of diabetic retinopathy study testing protocol. Am J Ophthalmol. 2003;135:194–205.
Hazari H, Curtis R, Eden K, Hopman WM, Irrcher I, Bona MD. Validation of the visual acuity iPad app Eye Chart Pro compared to the standard Early Treatment Diabetic Retinopathy Study chart in a low-vision population. J Telemed Telecare. 2022;28(9):680–6. https://doi.org/10.1177/1357633X20960640. (Epub 2020 Sep 26).
Samanta A, Mauntana S, Barsi Z, Yarlagadda B, Nelson PC. Is your vision blurry? A systematic review of home-based visual acuity for telemedicine. J Telemed Telecare. 2020. https://doi.org/10.1177/1357633x20970398.
Claessens JLJ, Geuvers JR, Imhof SM, Wisse RPL. Digital tools for the self-assessment of visual acuity: a systematic review. Ophthalmol Ther. 2021;10(4):715–30. https://doi.org/10.1007/s40123-021-00360-3. (Erratum in: Ophthalmol Ther. 2021 Dec;10(4):731-732).
Steren BJ, Young B, Chow J. Visual acuity testing for telehealth using mobile applications. JAMA Ophthalmol. 2021;139(3):344–7.
Bastawrous A, Rono HK, Livingstone IA, Weiss HA, Jordan S, Kuper H, Burton MJ. Development and validation of a smartphone-based visual acuity test (peek acuity) for clinical practice and community-based fieldwork. JAMA Ophthalmol. 2015;133(8):930–7. https://doi.org/10.1001/jamaophthalmol.2015.1468. (Erratum in: JAMA Ophthalmol. 2015 Sep;133(9):1096).
Wisse RPL, Muijzer MB, Cassano F, Godefrooij DA, Prevoo YFDM, Soeters N. Validation of an independent web-based tool for measuring visual acuity and refractive error (the manifest versus online refractive evaluation trial): prospective open-label noninferiority clinical trial. J Med Internet Res. 2019;21(11): e14808. https://doi.org/10.2196/14808.
Claessens J, van Egmond J, Wanten J, Bauer N, Nuijts R, Wisse R. The accuracy of a web-based visual acuity self-assessment tool performed independently by eye care patients at home: method comparison study. JMIR Form Res. 2023;25(7): e41045. https://doi.org/10.2196/41045.
Fountoulakis K, Tsolaki M, Chantzi H, Kazis A. Mini Mental State Examination (MMSE): a validation study in Greece. Am J Alzheimers Dis Other Dem. 2000;15:342–5. https://doi.org/10.1177/153331750001500604.
Bailey IL, Lovie JE. New design principles for visual acuity letter charts. Am J Optom Physiol Opt. 1976;53(11):740–5. https://doi.org/10.1097/00006324-197611000-00006.
Chung ST, Mansfield JS. Contrast polarity differences reduce crowding but do not benefit reading performance in peripheral vision. Vis Res. 2009;49(23):2782–9.
Rosen S, Pelli DG. Crowding by a repeating pattern. J Vis. 2015;15(6):10.
Manassi M, Sayim B, Herzog MH. Grouping, pooling, and when bigger is better in visual crowding. J Vis. 2012;12(10):13.
Khoshnood B, Mesbah M, Jeanbat V, Lafuma A, Berdeaux G. Transforming scales of measurement of visual acuity at the group level. Ophthalmic Physiol Opt. 2010;30(6):816–23.
Plainis S, Tzatzala P, Orphanos Y, Tsilimbaris MK. A modified ETDRS visual acuity chart for European-wide use. Optom Vis Sci. 2007;84(7):647–53. https://doi.org/10.1097/OPX.0b013e3180dc9a60.
Plainis S, Kontadakis G, Feloni E, Giannakopoulou T, Tsilimbaris MK, Pallikaris IG, Moschandreas J. Comparison of visual acuity charts in young adults and patients with diabetic retinopathy. Optom Vis Sci. 2013;90(2):174–8. https://doi.org/10.1097/OPX.0b013e31827ce251.
European Committee for Standardization. Repertoire of characters used for writing the indigenous languages of Europe. Report No.: CEN/TC304, Project 11. Brussels: European Committee for Standardization; 1997.
Grigas G, Juskeviciene A. Letter frequency analysis of languages using Latin alphabet. Int Linguist Res. 2018. https://doi.org/10.30560/ilr.v1n1p18.
Sloan L, Rowland WM, Altman A. Comparison of three types of test target for measurement of visual acuity. Q Rev Ophthalmol Otorhinolaringol. 1952;8:4–17.
https://vision2022dublin.com/wpcontent/uploads/2022/07/Vision2022_AbstractBook_Final-1.pdf
Ravindra R, Amarasuriya T. Role of brand identity in developing global brands: a literature based review on case comparison between Apple iPhone vs Samsung smartphone brands. Res J Bus Manage. 2015;2:430–40.
Götz FM, Stieger S, Reips UD. Users of the main smartphone operating systems (iOS, Android) differ only little in personality. PLoS ONE. 2017;12(5): e0176921. https://doi.org/10.1371/journal.pone.0176921.
Aaron Caldwell. SimplyAgree: flexible and robust agreement and reliability analyses. R package version 0.0.3. 2022. https://CRAN.R-project.org/package=SimplyAgree.
Siderov J, Tiu AL. Variability of measurements of visual acuity in a large eye clinic. Acta Ophthalmol Scand. 1999;77:673–6.
Laidlaw DA, Tailor V, Shah N, Atamian S, Harcourt C. Validation of a computerised logMAR visual acuity measurement system (COMPlog): comparison with ETDRS and the electronic ETDRS testing algorithm in adults and amblyopic children. Br J Ophthalmol. 2008;92(2):241–4. https://doi.org/10.1136/bjo.2007.121715. (Epub 2007 Nov 9).
Rosser DA, Cousens SN, Murdoch IE, Fitzke FW, Laidlaw DA. How sensitive to clinical change are ETDRS log MAR visual acuity measurements? Invest Ophthalmol Vis Sci. 2003;44:3278–81.
Rosser DA, Murdoch IE, Cousens SN. The effect of optical defocus on the test-retest variability of visual acuity measurements. Invest Ophthalmol Vis Sci. 2004;45:1076–9.
Deloitte. Mobile consumer survey 2017: The Australian cut. 2017. Retrieved from http://landing.deloitte.com.au/rs/761-IBL-328/images/tmt-mobile-consumer-survey-2017_pdf, pdf#targetText=The%202017%20study%20comprises%20more,become%20the% 20smartphone%20space%2Drace. Accessed Dec 5, 2018.
Sujarwoto S, Augia T, Dahlan H, Sahputri RAM, Holipah H, Maharani A. COVID-19 mobile health apps: an overview of mobile applications in Indonesia. Front Public Health. 2022;10: 879695. https://doi.org/10.3389/fpubh.2022.879695.
Saleem SM, Pasquale LR, Sidoti PA, Tsai JC. Virtual ophthalmology: telemedicine in a COVID-19 era. Am J Ophthalmol. 2020;216:237–42. https://doi.org/10.1016/j.ajo.2020.04.029.
Abdulhussein D, Abdul Hussein M, Szymanka M, Farag S. A systematic review of the current availability of mobile applications in eyecare practices. Eur J Ophthalmol. 2022. https://doi.org/10.1177/11206721221131397. (Epub ahead of print).
Brady CJ, Eghrari AO, Labrique AB. Smartphone-based visual acuity measurement for screening and clinical assessment. JAMA. 2015;314:2682–3.
Mathews SC, McShea MJ, Hanley CL, Ravitz A, Labrique AB, Cohen AB. Digital health: a path to validation. NPJ Digit Med. 2019;2:38.
Rosser DA, Laidlaw DAH, Murdoch IE. The development of a “reduced logMAR” visual acuity chart for use in routine clinical practice. Br J Ophthalmol. 2001;85(4):432–6. https://doi.org/10.1136/bjo.85.4.432.
Mataftsi A, Koutsimpogeorgos D, Brazitikos P, Ziakas N, Haidich AB. Is conversion of decimal visual acuity measurements to logMAR values reliable? Graefes Arch Clin Exp Ophthalmol. 2019;257(7):1513–7. https://doi.org/10.1007/s00417-019-04344-9. (Epub 2019 May 8).
Claessens J, van Egmond J, Wanten J, Bauer N, Nuijts R, Wisse R. The accuracy of a web-based visual acuity self-assessment tool performed independently by eye care patients at home: method comparison study. JMIR Form Res. 2023;25(7): e41045. https://doi.org/10.2196/41045.
Mercado CL, Doroslovački P, Wang J, Siddiqui AA, Kolker AF, Kolker RJ. Comparison of distance and near visual acuity in patients with vision loss due to cataract. Int Ophthalmol. 2017;37(1):169–72. https://doi.org/10.1007/s10792-016-0247-y. (Epub 2016 May 10).
Chun H, Lee YC. Comparison between distance and near visual acuity in amblyopes. Medicine (Baltimore). 2020;99(46): e23111. https://doi.org/10.1097/MD.0000000000023111.
Jin P, Zhu J, Zou H, Lu L, Zhao H, Li Q, He X. Screening for significant refractive error using a combination of distance visual acuity and near visual acuity. PLoS ONE. 2015;10(2): e0117399. https://doi.org/10.1371/journal.pone.0117399.
Peh KK, Agelis L, Chen FK. Are ETDRS Scores at 1m and 4m interchangeable? Invest Ophthalmol Vis Sci. 2012;53(14):4786.
Naor J, Hnik P, Hao Y, Wolin M. Comparison of visual acuity scores obtained with an EDTRS chart at 2 and 4 m. Invest Ophthalmol Vis Sci. 2005;46(13):245.
Forshaw TRJ, Parpounas AK, Sørensen TL. Correlation of macular sensitivity measures and visual acuity to vision-related quality of life in patients with age-related macular degeneration. BMC Ophthalmol. 2021;21(1):149. https://doi.org/10.1186/s12886-021-01901-x.
Ubhi HK, Kotz D, Michie S, van Schayck OCP, West R. A comparison of the characteristics of iOS and Android users of a smoking cessation app. Transl Behav Med. 2017;7(2):166–71. https://doi.org/10.1007/s13142-016-0455-z.
Black JM, Jacobs RJ, Phillips G, et al. An assessment of the iPad as a testing platform for distance visual acuity in adults. BMJ Open. 2013;3:2730.
Acknowledgements
We would like to thank Gkioka Mara, Ph.D. for administering the Mini-Mental test to the study participants.
Funding
This research has been co‐financed by the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH – CREATE – INNOVATE (project code: T1EDK-03742). The authors funded the journal’s Rapid Service Fee.
Author Contributions
Conceptualization: Vasileios Karampatakis; Methodology: Vasileios Karampatakis; Formal analysis and investigation: Vasileios Karampatakis, Stavroula Almpanidou and Diamantis Almaliotis; Writing-original draft preparation: Vasileios Karampatakis and Stavroula Almpanidou; Writing- review and editing: Vasileios Karampatakis; Resources: Vasileios Karampatakis, and Diamantis Almaliotis; Supervision: Vasileios Karampatakis; Validation: Karampatakis Vasileios, and Stavroula Almpanidou; Data Curation: Diamantis Almaliotis, Stavroula Almpanidou, and Persefoni Talimtzi, Formal Analysis: Persefoni Talimtzi.
Prior Presentation
Preliminary results of the study were presented at the 13th International Low Vision Conference, 5–8 July 2022, The Convention Center Dublin, Ireland.
Disclosures
Vasileios Karampatakis, Diamantis Almaliotis, Persefoni Talimtzi and Stavroula Almpanidou have nothing to disclose.
Compliance with Ethics Guidelines
All study procedures were approved by the Committee for Bioethics and Ethics, Medical Department, Aristotle University of Thessaloniki (code#1.60/21.11.2018) and adhered to the principles embodied in the Declaration of Helsinki Code of Ethics of the World Medical Association. Informed written consent was obtained from all subjects after an explanation of the nature of the study.
Data Availability
The datasets generated during and/or analyzed during the current study are available from the corresponding author upon reasonable request.
Author information
Authors and Affiliations
Corresponding author
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License, which permits any non-commercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc/4.0/.
About this article
Cite this article
Karampatakis, V., Almaliotis, D., Talimtzi, P. et al. Design and Validation of a Novel Smartphone-Based Visual Acuity Test: The K-VA Test. Ophthalmol Ther 12, 1657–1670 (2023). https://doi.org/10.1007/s40123-023-00697-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40123-023-00697-x