Abstract
Background
Cognitive impairment occurs in up to 70% of people with MS (pwMS) and has a large impact on quality of life and working capacity. As part of the development of a smartphone-app (dreaMS) for monitoring MS disease activity and progression, we assessed the feasibility and acceptance of using cognitive games as assessment tools for cognitive domains.
Methods
We integrated ten cognitive games in the dreaMS app. Participants were asked to play these games twice a week for 5 weeks. All subjects underwent a battery of established neuropsychological tests. User feedback on acceptance was obtained via a five-point Likert-scale questionnaire. We correlated game performance measures with predetermined reference tests (Spearman’s rho) and analyzed differences between pwMS and Healthy Controls (rank biserial correlation).
Results
We included 31 pwMS (mean age 43.4 ± 12.0 years; 68% females; median Expanded Disability Status Scale score 3.0, range 1.0–6.0) and 31 age- and sex-matched HC. All but one game showed moderate–strong correlations with their reference tests, (|rs|= 0.34–0.77). Performance improved in both groups over the 5 weeks. Average ratings for overall impression and meaningfulness were 4.6 (range 4.2–4.9) and 4.7 (range 4.5–4.8), respectively.
Conclusion
Moderate–strong correlations with reference tests suggest that adaptive cognitive games may be used as measures of cognitive domains. The practice effects observed suggest that game-derived measures may capture change over time. All games were perceived as enjoyable and meaningful, features crucial for long-term adherence. Our results encourage further validation of adaptive cognitive games as monitoring tools for cognition in larger studies of longer duration.
Study Register
ClinicalTrials.gov: NCT04413032.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Multiple Sclerosis (MS) a chronic inflammatory and degenerative autoimmune disease of the central nervous system affecting approximately 2.4 million people worldwide is one of the most prevalent neurological causes of permanent disability in young adults [1, 2]. Up to 70% of people with Multiple Sclerosis (pwMS) show signs of cognitive impairment (CI) [3]. These symptoms affect emotional well-being, working capacity and quality of life (QoL) [4, 5]. Still, because comprehensive neuropsychological test batteries are time-consuming, require specialized examiners, and are not always well accepted by patients’ systematic cognitive assessments are not widely used in clinical practice, even in specialized centers [6]. To better understand disease progression and optimize treatment strategies, a more complete and detailed monitoring of cognitive functions is desirable.
In recent years, there have been many advances in digitization of cognitive assessment using the potential of digital devices such as computers, tablets, and smartphones [7,8,9,10,11,12,13,14]. Especially smartphone apps have many advantages: they are easily available, user-friendly and cheap, making them the ideal tool to reach a large range of the population [15]. Importantly, recent research has shown that digital games can offer valuable information about cognitive functions [7]. Thus, we hypothesized that adaptive smartphone-games may have relevant advantages over the standard neuropsychological tests regarding acceptance and motivation, accuracy (e.g. more accurate measurement through touchscreen-sensors vs. test-rater using stopwatch), standardization and objectivity (no rater-bias) and convenience (accessibility of smartphones) [12, 13, 16].
In this proof-of-concept study, we correlate measures derived from adaptive smartphone-based cognitive training games with results of predefined corresponding established cognitive paper and pencil tests. The primary objective is to identify cognitive game measures that correlate with established neuropsychological tests and identify games that are well-accepted and meaningful for pwMS. Additional exploratory objectives were (a) investigating correlations of all reference test scores with all game measure scores; (b) determining whether the games and tests could be attributed to specific cognitive domains, and c) analyzing differences between pwMS and HC regarding game performance and game ratings.
Methods
The Research Center for Clinical Neuroimmunology and Neuroscience Basel (RC2NB) is currently developing a smartphone app „dreaMS “, in cooperation with the medical device software manufacturer Healios AG [17, 18]. This app aims to allow monitoring of a large variety of potential digital biomarkers by having patients repeatedly perform short tasks via their smartphone. To reach the largest population possible, the app works on a wide range of smartphone models with both android (5.0 (API 21) or later) and iOS (11.0 or later) operating systems. The tasks included in the dreaMS app cover five different functional domains: dexterity, walking ability, balance, cognition, and vision. Additionally, the app includes patient-reported outcome measures (PROs) for fatigue, walking ability, general symptoms, and quality of life [19]. As part of a study to investigate the technical feasibility, reliability, and acceptance of the dreaMS app (NCT04413032), we integrated 10 adaptive cognitive training games from the commercially available brain training app Peak [20] in the dreaMS app via deep link.
Here, we report on the performance of these cognitive games as measures of cognitive domains including their correlation with established neuropsychological tests, acceptance by users, and assessment of their meaningfulness by pwMS.
Standard protocol approvals, registrations, and patient consents
This study was approved by the local ethics committee (Ethikkommission Nordwest- und Zentralschweiz (EKNZ), Basel, Switzerland, on July 17th 2020/project-ID 2020-01515). All participants gave their written informed consent. This study conforms with World Medical Association Declaration of Helsinki and was registered at ClinicalTrials.gov: NCT04413032.
Study procedures
Participants attended three visits, all of them taking place at the MS Center at the University Hospital Basel: 1. Screening visit (S), 2. Baseline visit (BL) and 3. End-of-study visit (EoS). A written informed consent was obtained from all participants at the screening visit. At BL, all participants underwent a neurological examination and completed a battery of established standardized neuropsychological tests (Table 1). The participants were instructed to download the dreaMS app which included the Peak games. A study nurse went through every task with the participants to ensure their understanding. During the 6-week study period, the participants were asked to play each game twice a week at home during week 1–5, according to a given schedule. To ensure maximum adherence, the participants received automated messages multiple times a week, reminding them to complete the tasks. After completion of a game, the data were uploaded and stored on a secure server of the University Hospital Basel. Every completion of a game was registered in the study portal, where study nurses were able to check adherence and, if necessary, contact the participant personally to remind them to complete the tasks. The participants then had the opportunity to complete the games behind schedule, provided they did so in the same week. In week 6, participants were asked to fill out multiple PROs via the app. At EoS, we collected participants’ feedback regarding acceptance and meaningfulness via a five-point Likert scale questionnaire. The complete feedback questionnaire can be found in the supplementary material (S1: Feedback questionnaire).
Participants
Inclusion criteria for pwMS comprised a diagnosis of MS (RRMS, SPMS, PPMS, CIS) according to the revised McDonald criteria (2017) [21] and an Expanded Disability Status Scale (EDSS) of ≤ 6.5. Further, clinical stability at the time of inclusion and during the whole study period was a prerequisite (pwMS who experienced a relapse or clinical progression during the study period would be excluded from the study). Both pwMS and HC were required to be between 18 and 70 years old, to own and be capable of using a smartphone device fulfilling defined minimum technical standards, have sufficient dexterity and visual functions, and be able to follow the study procedures. The complete list of in- and exclusion criteria can be found in the supplementary material (S2: Inclusion and exclusion criteria).
Instruments and measurements
10 cognitive training games from Peak were included in the dreaMS app via deep link [20]. Game selection was based on domains known to be relevant in pwMS [4]: short-term and working memory, mental flexibility and processing speed, inhibition, language, and visuo-construction. All games were structured with multiple difficulty levels, which would adapt to the performance. This difficulty-adaptation is based on the sequence of correct/false answers: after X consecutive correct answers, the difficulty increases, after X consecutive errors, the difficulty decreases. Furthermore, a ranking-system is included: if the user reaches a certain score-threshold in two consecutive gaming sessions, the following session will start from a higher difficulty. Equally, the same is the case for leveling down. For each game, we determined quantifiable features as measures of game performance. Features are measures derived from the results of a test. Typically, in cognitive tests, these are: Number of correct answers within a given time frame but also number of errors per se. Depending on the structure of a cognitive game additional features like qualitative and temporal patterns of errors might also be derived or a combination of such features. In adaptive games, the level of difficulty reached by the participant could also by itself be a valid measure. We prospectively assigned each game to a cognitive domain according to information provided by the game developer and expert consensus. Every Peak game was intended to train a specific cognitive domain, which was developed under supervision of certified neuropsychologists at Peak [20]. To ensure the correctness of this categorization, P. Calabrese and S. Pless analyzed the tasks of each game and confirmed that they represent the domain proposed by the developer. Established cognitive tests covering the same cognitive domains were then selected out of a comprehensive neuropsychological test battery used routinely in the assessment of pwMS participating in the Swiss MS Cohort Study (SMSC) [22]. Regarding the selection of game-features, we focused on quantifiable measures most similar to those used in established cognitive assessments e.g. the number of correct answers in a given time interval, since it includes both speed and accuracy. However, the games have the additional feature of difficulty-adaptation, according to a scoring system based on the number of correct answers. Since the difficulty level adapts, the measure number of correct answers can be misleading for some games. In these cases, taking the change in difficulty level itself as the measure was the preferred option (table S1: Description of cognitive games and measures used for statistical analyses).
As a general screening test of cognitive impairment we used the MUSIC (Multiple Sclerosis Inventory of Cognition) [23]. The cognitive games chosen, their corresponding reference tests, and the cognitive domain they refer to, are shown in Table 1. The game-derived measures used for the statistical analyses and brief game descriptions are listed in the supplementary material (table S1: Description of cognitive games and measures used for statistical analyses). Exemplary Screenshots of the Peak Games are shown in the supplementary material (figures S1–S6: Exemplary Screenshots of Peak Games).
At end of study (EoS), all participants were asked to provide feedback. For each game, participants rated four questions on a five-point Likert scale. The questions addressed the participant’s overall impression, whether the game was appropriately challenging (including an additional elaboration question), whether the participant would be willing to play the games regularly in the future, and whether they thought the games were relevant for MS (meaningfulness). The question about meaningfulness was only asked to pwMS. Mean ratings of those four questions were used to calculate overall acceptance.
Objectives, outcomes and statistical analyses
The primary objective was to identify cognitive game measures that correlated with established neuropsychological tests and identify games that are well-accepted and meaningful for pwMS. Hence, the primary study outcomes were (a) correlation coefficient between game measures and the corresponding reference test scores. The average of the game measures over all ten sessions (2x/week for 5 weeks) was correlated with the corresponding established neuropsychological reference test score using spearman’s rank correlation coefficient. As an acceptable threshold, we targeted an at least moderate (rs ≥ 0.3) correlation coefficient [30]; (b) user acceptance and meaningfulness of the games for pwMS as assessed via a 5-point Likert scale questionnaire. Regarding user acceptance (overall impression, meaningfulness, and willingness to use in the future) we aimed for a mean Likert scale score of ≥ 3.
Additional exploratory objectives were the correlations of all reference test scores (mean) with all game measure scores (mean), using spearman’s rank correlation coefficient in correlation matrices. Further, we investigated whether the games and tests could be attributed to specific cognitive domains by performing a factor analysis including all game measures (mean) and test scores (mean), using maximum likelihood estimation and the rotation method “varimax”. Differences between pwMS and HC regarding game performance were assessed by means of rank biserial correlations between the average game scores over all ten sessions and the group variable. Since the analysis of differences between the two groups was exploratory, we used the more conservative analysis method: rank biserial correlations, in order to avoid misinterpretation. Spaghetti plots of all participants’ game performances over the ten sessions were created to better visualize group differences over time. Further, we compared the mean Likert scale game ratings across pwMS and HC using rank biserial correlation.
Statistical analyses were performed using the Statistical Package for the Social Sciences (SPSS).
Version 28.0 and R version 4.2.0 (2022-04-22, R Core Team, 2022). The study protocol and statistical analysis plan are available as supplementary material (S3: Study protocol).
Data access and availability statement
Ludwig Kappos, Silvan Pless, and Andrea Wiencierz take full responsibility for the data, the analyses and interpretation, and the conduct of the research, have full access to all of the data and have the right to publish any data separate and apart from any sponsor. The data supporting the findings of this study are available from the corresponding author upon reasonable request.
Results
Between October 5th 2020 and February 28th 2021, we recruited 31 pwMS from the MS Center, University Hospital Basel with a mean age of 43.4 ± 12.0 years, 68% females with a median Expanded Disability Status Scale score of 3.0 (range 1.0–6.0) and 31 age- and sex-matched HC. Table 2 provides an overview of the participants’ demographics. Three participants (2 HC, 1 pwMS) were excluded from the performance analyses due to incomplete data, however, they were included in the acceptance rating analysis.
According to the results of MUSIC cognitive impairment (CI) was present in 7/30 (22.5%) pwMS (3 (9.7%) moderate and 4 (12.9%) mild). No CI was observed in the HC group. When comparing the mean MUSIC score between pwMS (24 ± 4.9) and HC (26 ± 3.9), no clear differences were found (|rrb|= − 0.24, p = 0.07). We did not find a strong correlation between the MUSIC and the EDSS (rs = − 0.33, 95%CI − 0.62 to 0.05, p = 0.08).
CIS clinically isolated syndrome; RRMS relapse remitting MS; SPMS secondary progressive MS; PPMS primary progressive MS; EDSS expanded disability status scale; MUSIC multiple sclerosis inventory of cognition; CI cognitive impairment
Primary outcomes
Correlation of game-derived measures with established neuropsychological tests
All but the language-game Word Hunt (rs = − 0.25, 95%CI − 0.48 to 0.01) reached the preset level of moderate spearman correlation coefficients (|rs|≥ 0.3) with their predefined reference tests. Six games (Must Sort (rs = 0.77, 95%CI 0.64–0.86), Spin Cycle (rs = 0.53, 95%CI 0.31–0.70), Rush Back (rs = 0.66, 95%CI 0.48–0.79), Face Switch (rs =− 0.51, 95%CI − 0.68 to − 0.28), Perilous Path (rs = 0.51, 0.28–0.68), and Low Pop (rs = − 0.72, 95%CI − 0.82 to − 0.56) correlated strongly with their reference tests. Three (Zap Gap (rs =− 0.34, 95%CI − 0.55 to − 0.08), Baggage Claim (rs = 0.48, 95%CI 0.25–0.67), and Puzzle Blox (rs = 0.44, 95%CI 0.20–0.63)) showed moderate correlation coefficients. An overview of the correlation coefficients, their 95% confidence intervals and p-values is shown in Table 3.
Acceptance and meaningfulness
Over all games, the mean Likert scale rating was 4.63 (range 4.15–4.9) for overall impression, 4.40 (range 4.13–4.58) for perceived difficulty, 4.50 (range 4.15–4.77) for willingness to use in the future and 4.69 (range: 4.52–4.84) for meaningfulness for pwMS. All categories in all games met the predefined target of ≥ 3. Table 4 shows the means and standard deviations of acceptance ratings by game.
Exploratory outcomes
Correlation matrix and factor analysis
Of the 70 possible correlations (10 cognitive games * 7 reference tests), we found 2 near zero (|rs|< 0.1), 17 weak (|rs|≥ 0.1), 28 moderate (|rs|≥ 0.3), and 23 strong (|rs|≥ 0.5) correlations as shown in Table 5. Regarding correlations of the established reference tests among each other (21 possible combinations), 7 weak, 6 moderate, and 8 strong (Table 5). The exploratory factor analysis showed that separating the games and tests into four factors was sufficient. The SDMT showed equal factor loadings for two factors (factor 1 and 3). Attributing every game and test to the factor with its highest factor loading, factor 1 includes 2 tests (SDMT and VLMT), and 3 games (Face Switch, Must Sort, and Rush Back). One Test (ROCF recall) and 5 games (Baggage Claim, Puzzle Blox, Perilous Path, Spin Cycle, and Zap Gap) showed the highest factor loadings with factor 2. Four tests: ROCF copy, Stroop, TMT b, SDMT, and the game Word Hunt had the highest association with factor 3. Lastly, the test RWT and the game Low Pop showed the highest factor loadings with factor 4 (Table 6).
Group differences in game performance
When comparing the means of game scores (mean pwMS, mean HC, rank biserial correlation coefficient), no clear differences between pwMS (n = 30) and HC (n = 29) were found: Low Pop (11.6, 11.9, |rrb|= 0.188), Baggage Claim (2.14, 2.51, |rrb|= 0.378), Face Switch (26.8, 28.5, |rrb|= 0.11), Must Sort (80.7, 85.7, |rrb|= 0.125), Perilous Path (6.8, 7.0, |rrb|= 0.053), Puzzle Blox (3.0, 3.2, |rrb|= 0.096), Rush Back (43.6, 47.6, |rrb|= 0.193), Spin Cycle (1.5, 1.7, |rrb|= 0.182), Word Hunt (2.3, 2.0, |rrb|= 0.255), Zap Gap (4.3, 4.8, |rrb|= 0.174). When the average scores of each session for each participant are plotted, an increase in game-scores over the 5-week study period is visible, with the control-group showing steeper increases in some, and higher scores and or difficulty levels reached in most games (Figs. 1, 2, 3, 4, 5, 6, 7, 8, 9, 10). The average score of all but two games (Must Sort and Word Hunt) showed a continuous increase over the ten sessions. When we plotted the data obtained from these two games using mean difficulty level instead of number of successful rounds and completion time as game measures, an increase of average performance was visible as well (Figs. 11 and 12).
Group differences in game acceptance ratings
There were no clear differences regarding acceptance ratings between pwMS and HC. An overview of the game acceptance ratings can be found in the supplementary material (Table S2: Mean acceptance ratings on Likert scale (1–5) by group and topic (HC vs pwMS)).
Discussion
Correlation with established neuropsychological assessments
All but one of the adaptive cognitive games showed a moderate–strong correlation with their respective pre-assigned established reference tests and met the predefined correlation level (rs ≥ 0.3) supporting their utility as cognitive assessment tools. The fact that the language-game Word Hunt did not reach a correlation coefficient of rs ≥ 0.3 with RWT may be the effect of different assessment approaches: whereas Word Hunt relies on visual detection of words in a letter grid, the RWT is a purely verbal semantic fluency test. The overall correlation matrix, that shows moderate correlations of Word Hunt with all other reference tests (range rs = 0.36–0.69), most of which include visual tasks, supports this assumption. Similarly, the only other predefined game-test correlation below rs = 0.4: Zap Gap – Stroop, might have also been affected by the issue of different assessment approaches: while Stroop is a purely verbal test, in Zap Gap small orbs have to be touched at the right moment. Ergo visual-timing and dexterity might have influenced the comparison. Of course, other games also include a dexterity aspect which does usually not play a role in most paper pencil tests. However, dexterity required in the other games are very basic (touching large buttons to respond by yes/no). When omitting these two tests due to this effect, the other 8/10 correlations, where assessment methods are more similar, are distributed with a rather small range (rs = 0.44–0.77).
Amongst the established reference tests, the correlation of SDMTs with the other tests stood out. Probably this relates to the fact that SDMT targets processing speed most and games are time-limited and, therefore, dependent on cognitive speed. This interpretation is supported both by the factor analysis, where the SDMT showed equal loadings for two of the four factors, and by the literature where the value of the SDMT as a practicable overall measure of CI in MS, is increasingly recognized [31]. The fact that speed is one of the main measures in many neuropsychological tests and cognitive games might have contributed to the many moderate–strong correlations, we were able to observe. Considering that many cognitive functions are based on information processing, this seems likely. However, while speed certainly is a factor shared by many games and neuropsychological tests, it alone did not explain all of our results. The tests and their corresponding games: VLMT, ROCF, Baggage Claim, and Perilous Path all have no time-factor and did not differ from the speed-based tests or games in the correlation matrix.
The many moderate–strong correlations with not only the preassigned reference tests but between the vast majority of games and established cognitive tests, depicted by the correlation matrix (Table 5), suggests that neither the games nor the chosen established neuropsychological reference tests are highly specific for single defined cognitive domains. To further investigate domain-affinity, we conducted a factor analysis with all cognitive games and reference tests. By categorizing each game and test into the factor with the strongest loading, we distinguished the following four factors by cognitive functions: 1. mental speed; 2. memory; 3. visual perception/recognition, and 4. mental flexibility. Whereas this categorization fits most games and tests, only 3/10 of the preassigned pairs of games and tests matched the same factor. These results show that even though a certain domain affinity was observable, an exclusive domain-based categorization was not possible. The fact that not only the games but also the reference tests, which are designed to assess specific cognitive domains, did not show clear domain-specificity suggests that this is more likely a genuine effect of the interdependence of cognitive domains and their measures rather than an issue of the game design only. Both the investigated cognitive games and the established reference tests seem to cover a broader spectrum of interdependent cognitive domains [32]. Furthermore, both test and game results depend on the measurement method chosen (e.g. mental processing, mental flexibility, and inhibition are mostly measured by speed). The observed correlation might therefore—at least in part—reflect the shared assessment measure, rather than an overlap of the cognitive domains. In any case, the main aim of this study was not to prove high domain-specificity, but rather to investigate whether cognitive games can reliably measure levels of cognitive performance.
The improvements observed in the scores achieved in the games reflect the practice effect which occurs when repeating a similar task multiple times [33]. Only in the games Must Sort (processing speed), and Word Hunt (language), we did not observe a clear increase in performance. For these two games, the predefined measures were raw scores (number of correct answers and completion time). We assume that the scores of these games stagnated or declined because the level-, and therefore difficulty change interfered with the measurement of raw scores. As performance improved, the difficulty level of these adaptive games increased, leading to less increase or even relative decrease in the raw scores. When we measured the performance over time by the difficulty level reached, a continuous increase in performance was seen for these two games as well despite the lower number of possible levels (Figs. 11 and 12). That training effects were consistently depicted by the games indirectly supports the value of cognitive games in measuring changes in performance but underlines the need to control for practice effects when assessing disease evolution over time. This task is not trivial as it is further complicated by evidence supporting a relation of practice effects in cognitive testing with disability progression as well as brain volume loss in pwMS [34]. Our study was not powered to detect group differences between pwMS and HCs. It is, therefore, not surprising that no clear group differences were found. Nevertheless, HC scored higher across all games and showed mostly steeper improvements during the study.
Overall, our results regarding correlation with established tests and change over time support the potential of cognitive games as measures of cognitive function in pwMS and HC. Our findings are in accordance with those of studies investigating gamified digital cognitive assessment methods in elderly people with CI, in patients post-stroke and in healthy individuals [12, 13] and one study using the tablet-based game EVO-Monitor to distinguish pwMS with CI from pwMS without CI, and from HC [7].
Acceptance by participants
Our results clearly show that the games were not only appealing, but they were also recognized as beneficial to the management of their disease by the participating pwMS. We suspect that the gamification factor plays a large role in the enjoyment and motivation to complete the cognitive games. This assumption is supported by the participant’s feedback provided in a semi-structured interview conducted at EoS. Similarly, studies by Cerrato et al. (2017) and Wiley et al. (2020) also describe advantages of gamification elements regarding motivation, positive affect, enjoyment, felt challenge, meaning, and even performance in cognitive tasks [16, 35]. Such features are key for the ascertainment of good adherence, a critical requirement for a long-term monitoring tool of a chronic disease.
Limitations
Since this was a sub study of the dreaMS feasibility study, the sample size, schedule, and prevalence of cognitive impairment were not specifically set up for the assessment of cognitive games [19]. The wide range of age and disability grades allowed by the inclusion criteria may have contributed to higher variability of the results and thus have reduced the power of the analysis. Since we are primarily interested in the potential of adaptive cognitive games as a monitoring tool, our aim is to show changes within one individual, rather than to compare to a population. Therefore, the wide age range should not have a major impact. Our study was not designed to further characterize amount and time course of the practice effects observed in this study, a known impediment of use as a monitoring tool [33, 36, 37]. Approaches such as varying the frequency of testing, providing multiple versions of the same test/game, or having an intense practice period with the goal of reaching the ceiling of the learning curve, need to be evaluated in further studies. Another inherent limitation of our study is related to the performance-dependent adaptation of the difficulty-levels. To prevent floor- and ceiling effects and help to avoid boredom (too easy) or frustration (too difficult) adapting the difficulty level is an important motivator [38]. On the flip side, adaptive levels interfere with interpretation of raw scores as long as these are not weighted by difficulty level. Establishing well-defined difficulty-levels and weighting the scores obtained according to difficulty level is a necessary prerequisite for the use of adaptive games as measures of cognitive performance. There might have been a recruitment bias towards people with technical affinity which might have influenced general performance and acceptance ratings.
Conclusion
In this feasibility study, we demonstrate that smartphone games can provide reliable measures of cognitive function both in pwMS and HC. Although most game-derived measures correlated with their established cognitive reference tests, domain-affinity needs to be further explored in larger and more diverse populations. Practice effects were clearly depicted in both HCs and pwMS, suggesting that cognitive game measurements are sensitive to change over time and learning curves have to be taken into account in data analysis. All participants found the games appealing and meaningful and were motivated to use such a monitoring tool on a regular basis for longer periods. Further studies with longer duration in larger populations are warranted to validate such cognitive games as monitoring tools of cognition in pwMS. While the novel method explored in this study focused on monitoring disease evolution in pwMS, we acknowledge the great potential it shows as a rehabilitation tool in both MS and other medical fields.
References
Lassmann H (2018) Multiple sclerosis pathology. Cold Spring Harb Perspect Med 8(3):1–16. https://doi.org/10.1101/cshperspect.a028936
Walton C, King R, Rechtman L et al (2020) Rising prevalence of multiple sclerosis worldwide: Insights from the Atlas of MS, third edition. Mult Scler J 26(14):1816–1821. https://doi.org/10.1177/1352458520970841
Benedict RHB, Bobholz JH (2007) Multiple sclerosis. Semin Neurol 27(1):78–85. https://doi.org/10.1055/s-2006-956758
Opara JA, Jaracz K, Brola W (2010) Quality of life in multiple sclerosis. J Med Life 3(4):352–358. https://doi.org/10.15844/pedneurbriefs-34-14
Strober LB, Rao SM, Lee JC, Fischer E, Rudick R (2014) Cognitive impairment in multiple sclerosis: an 18 year follow-up study. Mult Scler Relat Disord 3(4):473–481. https://doi.org/10.1016/j.msard.2014.03.004
Lambros M, Mary HK, Epameinondas L, Panagiotis P (2010) Assessment and rehabilitation of cognitive impairment in multiple sclerosis. Int Rev Psychiatry 22(1):22–34. https://doi.org/10.3109/09540261003589372
Hsu WY, Rowles W, Anguera J et al (2021) Application of an adaptive, digital, game-based approach for cognitive assessment in multiple sclerosis: observational study. J Med Internet Res. https://doi.org/10.2196/24356
Maillart E, Labauge P, Cohen M et al (2020) MSCopilot, a new multiple sclerosis self-assessment digital solution: results of a comparative study versus standard tests. Eur J Neurol 27(3):429–436. https://doi.org/10.1111/ene.14091
Rubin LH, Severson J, Marcotte TD et al (2021) Tablet-based cognitive impairment screening for adults with hiv seeking clinical care: observational study. JMIR Ment Heal. https://doi.org/10.2196/25660
Papp KV, Samaroo A, Chou H et al (2021) Unsupervised mobile cognitive testing for use in preclinical Alzheimer’s disease. Alzheimer’s Dement Diagn Assess Dis Monit 13(1):1–10. https://doi.org/10.1002/dad2.12243
Lam KH, van Oirschot P, den Teuling B et al (2021) Reliability, construct and concurrent validity of a smartphone-based cognition test in multiple sclerosis. Mult Scler J. https://doi.org/10.1177/13524585211018103
Jung HT, Daneault JF, Lee H et al (2019) Remote assessment of cognitive impairment level based on serious mobile game performance: an initial proof of concept. IEEE J Biomed Heal Inf 23(3):1269–1277. https://doi.org/10.1109/JBHI.2019.2893897
Bonnechère B, Van Vooren M, Bier JC et al (2018) The use of mobile games to assess cognitive function of elderly with and without cognitive impairment. J Alzheimers Dis 64(4):1285–1293. https://doi.org/10.3233/JAD-180224
Petrova-Antonova D, Spasov I, Petkova Y, Manova I, Ilieva S (2020) Cognisoft: a platform for the automation of cognitive assessment and rehabilitation of multiple sclerosis. Computers 9(4):1–13. https://doi.org/10.3390/computers9040093
Brewster PWH, Rush J, Ozen L, Vendittelli R, Hofer SM (2021) Feasibility and psychometric integrity of mobile phone-based intensive measurement of cognition in older adults. Exp Aging Res 47(4):303–321. https://doi.org/10.1080/0361073x.2021.1894072
Cerrato A, Ponticorvo M (2017) Enhancing neuropsychological testing with gamification and tangible interfaces: the baking tray task, 10338th edn. Springer International Publishing. https://doi.org/10.1007/978-3-319-59773-7
RC2NB-dreaMS (2023). https://rc2nb.unibas.ch/en/research/#c520. Accessed 21 Mar 2023
Healios (2023). https://healios.io/healios-science/. Accessed 21 Mar 2023
Woelfle T, Pless S, Reyes O et al (2022) Reliability and acceptance of dreaMS, a software application for people with multiple sclerosis: a feasibility study. J Neurol. https://doi.org/10.1007/s00415-022-11306-5
Peak (2023). https://www.peak.net/about/. Accessed 21 Mar 2023
Thompson AJ, Banwell BL, Barkhof F et al (2018) Diagnosis of multiple sclerosis: 2017 revisions of the McDonald criteria. Lancet Neurol 17(2):162–173. https://doi.org/10.1016/S1474-4422(17)30470-2
Disanto G, Benkert P, Lorscheider J et al (2016) The Swiss Multiple Sclerosis Cohort-Study (SMSC): a prospective Swiss wide investigation of key phases in disease evolution and new treatment options. PLoS ONE 11(3):1–13. https://doi.org/10.1371/journal.pone.0152347
Calabrese P, Kalbe E, Kessler J (2004) Das Multiple Sklerose Inventarium Cognition (MUSIC). Psychoneuro 30(7):384–388
Aschenbrenner S, Tucha O, Lange KW (2001) Regensburger Wortflüssigkeitstest (RWT). Göttingen: Hogrefe
Smith A (1973) Symbol Digit Modalities Test. Los Angeles, CA: Western Psychological Services
Spreen O, Strauss E (1998) A compendium of neuropsychological tests, 2nd edn. Oxford Univ Press, Cham
Helmstaedter C, Durwen HF (1990) VLMT: Verbaler Lern- und Merkfähigkeitstest: Ein praktikables und differenziertes Instrumentarium zur Prüfung der verbalen Gedächtnisleistungen [VLMT: A useful tool to assess and differentiate verbal memory performance]. Schweizer Archiv für Neurologie, Neurochirurgie und Psychiatrie 141(1):21–30
Rey A, Osterrieth PA (1941) Rey-Osterrieth complex figure copying test. Psychol Assess. https://doi.org/10.1037/t07717-000
Reitan RM (1955) The relation of the trail making test to organic brain damage. J Consult Psychol. https://doi.org/10.1037/h0044509
Cohen J (1988) Statistical power analysis for the behavioral sciences, 2nd edn. Lawrence Erlbaum Associates
Benedict RHB, Deluca J, Phillips G, LaRocca N, Hudson LD, Rudick R (2017) Validity of the symbol digit modalities test as a cognition performance outcome measure for multiple sclerosis. Mult Scler 23(5):721–733. https://doi.org/10.1177/1352458517690821
Harvey PD (2019) Domains of cognition and their assessment. Dialog Clin Neurosci 21(3):227–237. https://doi.org/10.31887/DCNS.2019.21.3/pharvey
Woelfle T, Pless S, Wiencierz A, Kappos L, Naegelin Y, Lorscheider J (2021) Practice effects of mobile tests of cognition, dexterity, and mobility on patients with multiple sclerosis: data analysis of a smartphone-based observational study. J Med Internet Res 23(11):1–16. https://doi.org/10.2196/30394
Sormani MP, De Stefano N, Giovannoni G et al (2019) Learning ability correlates with brain atrophy and disability progression in RRMS. J Neurol Neurosurg Psychiatry 90(1):38–43. https://doi.org/10.1136/jnnp-2018-319129
Wiley K, Vedress S, Mandryk RL (2020) How points and theme affect performance and experience in a gamified cognitive task. Conf Hum Factors Comput Syst. https://doi.org/10.1145/3313831.3376697
Duff K (2012) Current topics in science and practice evidence-based indicators of neuropsychological change in the individual patient: relevant concepts and methods. Arch Clin Neuropsychol 27(3):248–261. https://doi.org/10.1093/arclin/acr120
Thorgusen SR, Suchy Y, Chelune GJ, Baucom BR (2016) Neuropsychological practice effects in the context of cognitive decline: contributions from learning and task novelty. J Int Neuropsychol Soc 22(4):453–466. https://doi.org/10.1017/S1355617715001332
Cowley B, Charles D, Black M, Hickey R (2008) Toward an understanding of flow in video games. Comput Entertain 6(2):1–27. https://doi.org/10.1145/1371216.1371223
Acknowledgements
We thank all study participants and PopReach Incorporated (Peak) for providing access to a selection of their cognitive training games for this study and permission to publish our findings. We also thank Prof. Anthony Feinstein for his advice in planning this study.
Funding
Open access funding provided by University of Basel. This study was performed in collaboration with Healios AG, Basel, Switzerland and received funding from the Swiss Innovation Agency (Innosuisse, project ID 33535.1 IP-ICT).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
The research activities of RC2NB (Research Center for Clinical Neuroimmunology and Neuroscience Basel) are supported by the University Hospital and the University of Basel and by grants from Novartis, Roche and Neurostatus-UHB AG. One of the main projects of RC2NB is the development of a new comprehensive MS Digital solution. This study was performed in collaboration with Healios AG, Basel, Switzerland and received funding from the Swiss Innovation Agency (Innosuisse, project ID 33535.1 IP-ICT). PopReach Incorporated (Peak) provided access to selected cognitive training games without any influence on the study design, analysis, and interpretation. Silvan Pless has nothing to disclose. Tim Woelfle has nothing to disclose. Yvonne Naegelin’s institution (University Hospital Basel) has received financial support for lectures from Teva and Celgene and grant support from Innosuisse (Swiss Innovation Agency). Johannes Lorscheider’s institution has received research grants from Novartis, Biogen and Innosuisse as well as honoraria for advisory boards and/or speaking fees from Novartis, Roche and Teva. Oscar Reyes is Lead Data Scientist of Healios AG. Andrea Wiencierz has nothing to disclose. Pasquale Calabrese has received honoraria for speaking at scientific meetings, serving at scientific advisory boards and consulting activities from: Abbvie, Actelion, Almirall, Bayer-Schering, Biogen Idec, Celgene, EISAI, Genzyme, Lundbeck, Merck Serono, Novartis, Pfizer, Teva, and Sanofi-Aventis. His research is also supported by the Swiss Multiple Sclerosis Society and the Swiss National Research Foundation. Ludwig Kappos has received no personal compensation. His institution (University Hospital Basel/Foundation Clinical Neuroimmunology and Neuroscience Basel) has received the following exclusively for research support: steering committee, advisory board and consultancy fees (Abbvie, Actelion, AurigaVision AG, Biogen, Celgene, Desitin, Eli Lilly, EMD Serono, Genentech, Genzyme, Glaxo Smith Kline, Janssen, Japan Tobacco, Merck, Minoryx, Novartis, Roche, Sanofi, Santhera, Senda, Shionogi, Teva, and Wellmera; speaker fees (Celgene, Janssen, Merck, Novartis, and Roche); support for educational activities (Biogen, Desitin, Novartis, Sanofi and Teva); license fees for Neurostatus products; and grants (European Union, Innosuisse, Novartis, Roche Research Foundation, Swiss MS Society and Swiss National Research Foundation).
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Pless, S., Woelfle, T., Naegelin, Y. et al. Assessment of cognitive performance in multiple sclerosis using smartphone-based training games: a feasibility study. J Neurol 270, 3451–3463 (2023). https://doi.org/10.1007/s00415-023-11671-9
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00415-023-11671-9