Zum Inhalt

Verbal and non-verbal recognition memory assessment: validation of a computerized version of the Recognition Memory Test

  • Open Access
  • 22.12.2023
  • Original Article
Erschienen in:

Abstract

Background

The use of computerized devices for neuropsychological assessment (CNADs) as an effective alternative to the traditional pencil-and-paper modality has recently increased exponentially, both in clinical practice and research, especially due to the pandemic. However, several authors underline that the computerized modality requires the same psychometric validity as "in-presence" tests. The current study aimed at building and validating a computerized version of the verbal and non-verbal recognition memory test (RMT) for words, unknown faces and buildings.

Methods

Seventy-two healthy Italian participants, with medium–high education and ability to proficiently use computerized systems, were enrolled. The sample was subdivided into six groups, one for each age decade. Twelve neurological patients with mixed aetiology, age and educational level were also recruited. Both the computerized and the paper-and-pencil versions of the RMT were administered in two separate sessions.

Results

In healthy participants, the computerized and the paper-and-pencil versions of the RMT showed statistical equivalence for words, unknown faces and buildings. In the neurological patients, no statistical difference was found between the performance at the two versions of the RMT. A moderate-to-good inter-rater reliability between the two versions was also found in both samples. Finally, the computerized version of the RMT was perceived as acceptable by both healthy participants and neurological patients at System Usability Scale (SUS).

Conclusion

The computerized version of the RMT can be used as a reliable alternative to the traditional version.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Introduction

In the past few decades, the use of computerized neuropsychological assessment devices (CNADs) has received increasing attention in both clinical and research settings, because it is considered as a more advantageous alternative to conventional examiner-based tests [1]. The interest in CNADs has exponentially increased in response to the COVID-19 pandemic, which has forced significant changes in clinical practice. Indeed, one of the properties of CNADs is to allow remote encounters between clinicians and patients [see 2 for review]. Furthermore, focusing on the fact that neuropsychological tests on digital platforms are sensitive and cost-effective, several studies have already used these tools with many clinical populations [3]. However, to date, "paper-and-pencil" tests remain the gold standard for neuropsychological assessment [4, 5]. The American Academy of Clinical Neuropsychology and the National Academy of Neuropsychology, in a joint document, defined CNADs as any tool that administers, evaluates, or interprets tests of brain function using a computer, portable devices (such as digital tablets or smartphones), or any other digital interface instead of a human examiner [3]. However, such a position statement, that emphasizes the administration of cognitive tests by a computer "instead of" a human examiner, can be extended to all computerized neuropsychological platforms, including a broader spectrum of ever-growing technology and, therefore, include tools that may allow a better presentation of the stimuli and data recording through computer presentation (such as immersive and non-immersive Virtual Reality) [4]. As aforementioned, CNADs allow to present with extreme accuracy a wider range of tasks including, for example, those involving multi-tasking, multiple measurements, divided attention, speed processing and response times for multiple measurements. Furthermore, CNADs provide new parameters, more easily and quickly available [as, for instance, recording of response latencies, see 6], leading to greater accuracy in detecting cognitive changes [79], in both the clinical and research fields [10, 11]. Finally, computerized tests can be administered to many subjects in shorter times and with lower costs; the evaluation of the responses can be more precise, enabling automatic administration and scoring procedures, and thus reducing the probability of human error [3]. On the other hand, in a conventional neuropsychological assessment, there can be a great deal of variation between different examiners or between the same examiner along an assessment session, in the manual use of tools (such as a stopwatch), or in the manual presentation of material (such as stimulus cards) [12]. Furthermore, while in an examiner-centred paper-and-pencil assessment the patient interacts with an examiner who presents stimuli, records responses, notes key behavioral observations and also understands the patient’s level of effort and motivation, in CNADs, the patient interacts with a computerized test station or a tablet through one or more alternative input/output devices (keyboard, voice, mouse, joystick, touchscreen, head-mounted display), sometimes without any supervision or observation. Additionally, computerized assessment does not permit examiners to introduce flexibility into their evaluation or provide any structured encouragements to the examinee. For those reasons, the use of CNADs may require familiarity with technical skills that can significantly impact on relevant evaluation parameters. Patients, especially the elderly, may have limited familiarity with computer interfaces, as well as negative attitudes and anxiety about computers [13]: as a result, their performance may differ significantly from pencil-and-paper measurements [1416]. On these assumptions, computerized tests are not simply the replacement of paper-and-pencil for a computer screen and electronic response capture and they could not be considered tout court directly comparable to an examiner-administered evaluation [1719]. Rather, a traditional test administered by computer becomes a new test, different from the conventional one both qualitatively and technically. Indeed, several studies have proved that there are substantial differences between the computerized measurements and the corresponding examiner administered tests in different samples [2023]. In addition, the "replication crisis" [24] has increased the need for appropriate and robust statistical inference: when examiner-based tests are proposed in a computerized version, new psychometric data should be provided according to the same standards of psychometric tests. It cannot be assumed a priori that the normative data for a test administered by the examiner apply equally well to a computerized version of the same test, due to changes in the administration method and patients’ familiarity with the computer. Thus, if a computerized test adapted from a conventional test is not simply a slightly different format for an existing test, it becomes essential that equivalence testing between the paper-and-pencil and the computerized versions of that test is provided. New normative data for computerized tests must be collected when no equivalence with the paper-and-pencil version of the tests is proved. Several tests have already been adapted for computerized presentation, such as the Montreal Cognitive Assessment – MoCA [25], the Clock Drawing Test, the Pentagon Drawing Test [26], the Rey Complex Figure copy task [27], the Bells Cancellation, Line Bisection and Five Elements Drawing Tests [28]: these tests are, of course, critical in the assessment of cognitive functioning.
Among neuropsychological patients, memory problems are one of the most reported symptoms and, due to growing life expectancy, differentiating early neurodegenerative disorders from normal aging is for sure one of the greatest challenges for the future. Notably, despite the extensive body of literature supporting the use of recognition memory paradigms to enhance differential diagnosis between neurological patients, depressed ones and healthy controls [29], its clinical assessment is still limited due to the paucity of validated, clinical tools available. As for the Italian population, only the verbal and non-verbal recognition memory test (RMT) [30] and the visual long-term recognition memory test [31] are available. The RMT represents a relatively fast screening tool, which may be useful for the early detection of memory disorders; two parallel versions are also available for multiple testing [32]. However, despite its promising features, the original examiner-centred version of the RMT may be prone to human error: clinicians are indeed required to control for several factors, namely the duration of the stimuli presentation and the registration of patients’ responses and behaviours [see 30, methods section]. For all these reasons, we decided to provide a validation of a computerized version of the verbal and non-verbal recognition memory test (RMT—Form A) for words, unknown faces, and buildings [32], comparing it to the results of the original paper-and-pencil test. To this end, a computerized version of the RMT was built and we administered both the pencil-and-paper and the computerized version to different age-balanced groups of healthy participants, with medium–high education. Furthermore, in order to evaluate the possible impact of different neurological conditions on patients' ability to interact with the computer interface, the two version of the RMT were also administered to a small sample of neurological patients with mixed aetiology.

Materials and methods

Materials

The previously published battery of verbal and non-verbal recognition memory tests (RMT) [32] and its computerized version were used. The battery evaluates different components of recognition memory for words, unfamiliar faces, and unknown buildings. Stimuli selection and test construction are described elsewhere [30]. Briefly, the original, pencil-and-paper version had the same administration procedure for the three subtasks (i.e., for words, faces and buildings): during the study phase, the 30 target stimuli were individually displayed at a 3-s. interval per item and, for each stimulus, participants were asked to judge whether they like it or not to enhance an adequate level of attention. In the recognition phase, performed immediately after the end of the study phase, three stimuli were presented simultaneously: the target stimulus and two distractors. Participants were asked to recognize, among the three items, the previously seen one. Targets were distributed randomly to the top, bottom, or centre of the sheet, ensuring an equal number of placements in the three positions. The number of correct answers corresponds to the total score of each subtask (range: 0–30). Overall, the test lasts about 20 min.
Here we present a computerized version of the RMT. The test was built on PsychoPy and the stimuli were shown to the participants using a MacBook Pro 13" (with a set screen resolution of 1680 × 1050). The stimuli were the same as in Smirni et al.’s [30] original version and, in the study phase, each stimulus was displayed at a 3-s. interval. To ensure an adequate level of attention on each stimulus, participants were asked to press "1" on the keyboard if it produced any emotion or feeling (e.g. happiness, sadness, etc.), "0" if it did not produce any particular emotion or feeling (Fig. 1). With respect to the original one, in this computerized version, administration and scoring procedures are fully automatic: indeed, instructions are shown on the PC/laptop monitor and participants are invited to provide responses using some keys on the keyboard. Furthermore, after each subtest completion, an electronic sheet with all the answers provided by the participant, the raw score and its conversion into adjusted and equivalent scores is available. The system automatically applies the regression formulas of Smirni et al. [30] to the raw scores obtained by the participants, even in the case of minimum or maximum raw score (0/30 and 30/30 respectively). In agreement with Novelli et al. [33], in the former case the equivalent score is always zero, in the latter, the equivalent score is equal to four, disrespectfully from the foreseen automatic correction.
Fig. 1
Example of the study and recognition phases for the computerized version of the RMT
Bild vergrößern
In addition, the participants were administered the Montreal Cognitive Assessment (MoCA) [34] to assess the general cognitive level, and the System Usability Scale (SUS) [35], which evaluates the subjective opinion from a user regarding a particular software. The SUS consists of 10 items in which even-numbered questions are posed in negative form. All the items need to be rated on a Likert scale ranging from 1 (completely disagree), to 5 (completely agree).

Methods

The two versions of the RMT (i.e., the original pencil-and-paper and the computerized versions) were administered by one of the authors (BF) to a sample of 72 healthy participants, in two testing sessions over two separate days, at least one week apart and within 10 days. The examiner was always present in the same room during the computerized administration of the test, but no further instructions or clarifications were requested from the examinees. The administration order of the two versions (first pencil-and-paper; first computerized), was counterbalanced across subjects as well as the administration order of the three subtasks. Inclusion criteria for participants were: i) age between 20 and 79 years; ii) at least 5 years of education. Exclusion criteria were: i) a history or clinical evidence of neurological or neuropsychiatric disease; ii) subjective complaint of memory or other cognitive impairments; iii) a Montreal Cognitive Assessment (MoCA) adjusted score below the Italian cut-off score [34]. The sample was then subdivided into six groups, one for each age decade; the number of males and females was matched across the six groups.
Furthermore, a continuous series of 12 neurological patients with mixed aetiology, age and educational level was recruited at the Centre for Cognitive Rehabilitation (CeRiN) between May and August 2023. The two versions of the test RMT were administered by two psychologists expert in neuropsychology (EB and ST), to evaluate the possible impact of different neurological conditions on patients' ability to interact with the computer interface. The administration order of the two versions was counterbalanced across the patients as well as the administration order of the three subtasks.
Written informed consent was obtained from all participants and the study was conducted in accordance with the Declaration of Helsinki, with the approval of the local ethics committee.
Socio-demographic characteristics of both healthy participants and neurological patients are shown in Table 1, together with the SUS scores.
Table 1
Socio-demographic characteristics, MoCA’s adjusted score and SUS’s raw score from the healthy participants (top of the table), and the neurological patients (bottom of the table)
Healthy participants
Decade
N
M/F
Age (years ± sd)
Education (years ± sd)
MoCA (adjusted score ± sd)
SUS (raw score ± sd)
  20–29
12
6/6
26.60 (1.63)
17.25 (1.14)
26.03 (2.52)
96.46 (5.48)
  30–39
12
6/6
33.06 (2.43)
15.50 (2.39)
26.60 (2.06)
95.00 (8.05)
  40–49
12
6/6
44.09 (2.32)
15.58 (3.99)
27.10 (2.53)
93.96 (9.62)
  50–59
12
6/6
56.99 (2.19)
11.67 (2.67)
26.75 (1.68)
98.54 (1.98)
  60–69
12
6/6
64.58 (3.05)
13.92 (3.75)
25.92 (1.25)
93.75 (5.69)
  70–79
12
6/6
74.47 (2.66)
10.33 (3.42)
25.12 (2.25)
96.67 (2.89)
Total
72
36/36
49.97 (17.26)
14.04 (3.80)
26.25 (2.13)
97.52 (6.23)
Neurological patients
ID
Aetiology
Sex
Age
Education
MoCA (adjusted score)
SUS (raw score)
  1
Traumatic brain injury
M
75
8
19.837
92.5
  2
Amnesic Mild Cognitive Impairment
M
76
13
19.668
92.5
  3
Haemorrhagic stroke
M
31
13
19.168
100
  4
Ischemic stroke
M
66
13
12.918
57.5
  5
Posterior cortical atrophy
F
52
13
11.168
87.5
  6
Ischemic stroke
F
31
18
22.729
75
  7
Brain tumour
M
20
11
23.337
92.5
  8
Ischemic stroke
F
29
18
22.168
92.5
  9
Ischemic stroke
F
59
11
20.216
92.5
  10
Ischemic stroke
M
68
8
18.087
87.5
  11
Body Lewy dementia
M
71
18
18.354
57.5
  12
Multi-domain Mild Cognitive Impairment
M
68
17
21.354
85
Total
-
8 M/4F
50.83 (20.48)
13.41 (3.65)
19.08 (3.69)
84.4. (13.9)
N number; M males; F females; MoCA Montreal Cognitive Assessment; SUS System Usability Scale

Statistical analysis

Learning effect

A series of linear mixed effects models were run to check for possible effects of learning. The raw scores at the three RMT subtasks were considered as a continuous dependent variable. As a factorial variable, the version of test (Version, two levels: paper-and-pencil; computerized), the administration order of the two versions (Version order, two levels: first paper-and-pencil; first computerized), and the administration order of the three subtasks (Task order; six levels: faces-words-buildings; faces-buildings-words; buildings-faces-words; buildings-words-faces; words-faces-buildings; words-buildings-faces), were added to the model. A by subject random intercept was also added to account for inter-subject variability. The same analysis was separately performed for healthy participants and neurological patients.

Inter-rater reliability

Inter-rater reliability between the computerized and the paper-and-pencil versions of the RMT was separately assessed for healthy participants and neurological patients through intra-class correlations. Results were interpreted according to Koo and Li’s guidelines [36].

Healthy participants

Equivalence between the two versions of the RMT for faces, buildings and words was assessed through a two one-sided test procedure (TOST) for dependent samples [37]. Indeed, while the lack of a significant difference between scores on a paired t test cannot be taken as an evidence of agreement, equivalence tests demonstrate that scores from different tests are equal within a predetermined boundary. Normality assumptions on raw variables were checked by the evaluation of skewness and kurtosis values (judged as abnormal if ≥|1| and |3|, respectively [38]). Given the non-normality of the raw scores, data were reflected and log transformed prior to analyses with equivalence tests. The TOST procedure involves an a priori determination of an equivalence boundary (± delta), within which score difference can be considered trivial. Delta was set a priori at 1.1 raw score points, which log transforms to 0.1 and results in an equivalence interval of − 0.1 and 0.1. α level was set at 0.005.

Neurological patients

The presence of any difference in the performance between the two versions of the RMT for faces, buildings and words was assessed through paired samples t-tests.

Results

Learning effect

When testing for possible effects of learning in healthy participants, ANOVA on the run final models did not showed a significant 3-way interaction effect (Version*Version order*Task order) for any of the three RMT subtasks (words: F(5, 60) = 1.102, p = 0.369; faces: F(5, 60) = 0.154, p = 0.987; buildings: F(5, 60) = 0.543, p = 0.743).
Similarly, no interaction effect (Version*Version order*Task order), was found when checking for a possible learning effect in the neurological patients for any of the three RMT subtasks (words: F(3,2) = 3.379, p = 0.237; faces: F(3,2) = 8.063, p = 0.112; buildings: F(3,2) = 0.573, p = 0.686).

Inter-rater reliability

In the healthy participants, inter-rater reliability between the two versions of the RMT was moderate for faces (ICC = 0.699), buildings (ICC = 0.644) and words (ICC = 0.746). In the neurological patients inter-rater reliability was moderate for buildings (ICC = 0.539), and good for faces (ICC = 0.756) and words (ICC = 0.836).

Healthy participants

When comparing the two versions of the RMT for faces, no significant differences were found (t(71) = 0.298, p = 0.767, d = 0.003). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.95 ± 2.11; paper-and-pencil version mean = 27.93 ± 2.20). Also when comparing the two versions of the RMT for buildings, no significant effect was found (t(71) = 895, p = 0.374, d = 0.105). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.87 ± 2.46; paper-and-pencil version mean = 28.23 ± 1.98). Finally, no significant differences were found even when comparing the two versions of the RMT for words (t(71) = 1.88, p = 0.064, d = 0.221). Consistently, they also showed statistical equivalence on the TOST procedure (both upper and lower equivalence bound yielding a p < 0.005; computerized version mean = 27.25 ± 2.56; paper-and-pencil version mean = 27.66 ± 2.26) (Fig. 2). Table 2 reports means and standard deviations scores obtained by the sample at the two versions of the RMT (paper-and-pencil, computerized), in all the three subtests (words, unfamiliar faces and unknown buildings), for each decade of age.
Fig. 2
Mean and standard error of the scores obtained by the healthy participants (top of the image), and the neurological patients (bottom of the image), in the three subtasks of the two version of the RMT
Bild vergrößern
Table 2
Mean scores and standard deviations of the two versions of the RMT in the validation sample divided by decades of age (top of the table), and in the neurological patients (bottom of the table)
Healthy participants
Version → 
Paper-and-pencil
Computerized
Decade ↓
Faces
Words
Buildings
Faces
Words
Buildings
  20–29
28.75 (1.54)
29.42 (.99)
28.92 (1.08)
29.08 (.99)
29.17 (.71)
28.83 (1.80)
  30–39
28.00 (2.17)
28.08 (1.31)
28.58 (1.37)
27.42 (2.06)
27.25 (2.13)
28.58 (1.44)
  40–49
28.08 (2.10)
27.92 (2.15)
29.08 (.90)
28.08 (1.92)
27.42 (3.05)
28.58 (1.24)
  50–59
28.17 (2.32)
27.17 (2.69)
28.50 (1.78)
28.83 (.93)
26.67 (3.14)
28.75 (1.05)
  60–69
27.92 (2.35)
26.42 (2.46)
27.92 (1.31)
27.92 (2.71)
25.83 (2.72)
27.17 (2.28)
  70–79
26.73 (2.61)
26.92 (2.50)
26.42 (3.39)
26.42(2.57)
26.08 (2.87)
25.33 (3.91)
Total
27.93 (2.20)
27.62 (2.26)
28.23 (1.98)
27.95 (2.11)
27.06 (2.72)
27.87 (2.46)
Neurological patients
Version → 
Paper-and-pencil
Computerized
Mean (sd)
22.83 (3.77)
20.91 (5.50)
21.83 (4.96)
21.41 (5.44)
19.66 (6.80)
20.50 (4.10)

Neurological patients

When comparing the two version of the RMT, no significant differences were found for faces (computerized version mean = 21.41 ± 5.44; paper-and-pencil version mean = 22.83 ± 3.79; t(11) = 1.48, p = 0.167, d = 0.427), buildings (computerized version mean = 20.50 ± 4.10; paper-and-pencil version mean = 21.83 ± 4.96; t(11) = 1.07, p = 0.309, d = 0.308), or words (computerized version mean = 19.66 ± 6.80; paper-and-pencil version mean = 20.91 ± 5.50; t(11) = 1.29, p = 0.224, d = 0.372). (Fig. 2). Table 2 reports means and standard deviations scores obtained by the neurological patients at the two versions of the RMT.

System usability scale – SUS

The SUS total median value from healthy participants was 97.5 (IQR = 5.63), with a minimum rating score of 72.5 and a maximum rating score of 100 (Fig. 3). The mean score of 95.7 (± 6.23), shows that the system is perceived as "excellent" and "acceptable" according to the adjective scale developed by Bangor et al. [39]. When evaluating the user experience of the neurological patients, the SUS total median value was 90.0 (IQR = 10.0), with a minimum rating score of 57.5 and a maximum rating score of 100 (Fig. 3). The mean score of 84.4 (± 13.9), shows that the system is perceived as "good" and "acceptable" [39].
Fig. 3
Distribution of the SUS scores from the healthy participants (on the left), and the neurological patients (on the right). Horizontal solid lines represent median and non-solid lines interquartile range
Bild vergrößern

Discussion

In recent years, a large body of literature has shown an increasing interest in the use of computerized modalities in neuropsychological assessment both in clinical practice and in research [3, 7, 8, 12, 40]. CNADs have been proved to be highly effective, compared to the traditional assessment. On the other hand, this considerable enthusiasm is reduced by others studies arguing that the computerized tests cannot be considered tout court the equivalent of the conventional pencil-and-paper test [17, 21]. The same studies underlined that a computerized version requires the same psychometric validity of the tests directly administered by an examiner. In addition, these studies pointed out that the use of CNADs creates a specific evaluation setting in which the patient’s performance can be conditioned by the peculiar modality and then it may provide information that do not correspond to the real subject’s ability [17, 21].
According to this theoretical-applicative context, the present study aimed to create and validate a computerized version of the pencil-and-paper verbal and non-verbal recognition memory test (RMT—Form A) for words, unfamiliar faces and unknown buildings [32]. Both the pencil-and-paper and computerized versions were administered to different age-balanced groups of healthy participants, with medium–high education (mean education: 14.04 years; range: 8–25), and ability to use a high computerized system, proficiently and without performance slowdowns (SUS mean score: 97.52/100; range: 72.5–100). Our results showed a moderate inter-rater reliability between the two versions of the RMT for words, unknown faces and buildings. Furthermore, when comparing the two versions, no significant differences were found, and according to the TOST procedure, the three tests can be considered equivalent in both modalities (i.e., computerized and paper-and-pencil). We also enrolled a small sample of neurological patients with mixed aetiology, age (range: 20–76) and educational level (range: 8–18) to evaluate the possible impact of different neurological conditions on the patients’ ability to interact with the computerized version of the RMT. Results showed a moderate-to-good inter-rater reliability between the two versions of the RMT. Furthermore, no difference was found in patient’s performance between the two versions of the RMT for words, unknown faces and buildings. Finally, results on SUS showed that the system is perceived as "good" and "acceptable" [39] (SUS mean score: 84.4/100; range: 57.5–100). Moreover, none of the patients referred feeling of discomfort while interacting with the computer interface, nor did they need further instructions from the experimenter, others then those provided by the computerized test. These findings support the possibility to use, in clinical practice, the computerized version as an alternative to the traditional one.
The computerized version of the RMT can offer several advantages in the assessment procedures, such as, for example, the reduction of the examiner’s influence, higher consistency administration, presentation of the stimuli, and scoring. It can also support a higher number of screening and follow-up activities on large populations [7, 40, 41], due to the ease of data collection and the scoring procedure, the immediate availability of results and a significant reduction in costs linked to administering the tests. Furthermore, another strength is that, in the spirit of Open Science, test materials and instructions are freely available on the web under a Creative Common licence (https://osf.io/yuqbh/?view_only=1edfc3db31b74ac581f83dfbc4b7cced). Therefore, interested neuropsychologists can use them in both clinical and research settings. Finally, this computerized version of the RMT may be suitable for online assessment, which has recently been acknowledge as a reliable alternative to the "in-presence", traditional, neuropsychological assessment for diagnosing neurocognitive and degenerative disorders [42, 43], such as preclinical dementia [44], as well as in healthy population assessment and in Alzheimer’s disease [45]. Online assessment makes it certainly easier for patients to access medical care permitting to save time and money even if several potential limitations must be considered (e.g., technological devices malfunctioning during the session; video or audio quality variability during the assessment).
The computerized version of the RMT is not exempt from limitations. For example, this test cannot be administered to poorly educated people over 80 years, because they usually are not familiar with technological devices. In fact, our validation sample included participants aged from 20 to 79 years, with medium–high education, and their rating of usability for the computerized version of the RMT was almost at ceiling. However, technological devices are expected to become usual tools for the whole population.

Acknowledgment

We are grateful to all volunteers who took part in the study. We thank Dr. Tommaso Mastropasqua for his support with test creation.

Declarations

Ethical approval

The study was conducted in accordance with the Declaration of Helsinki, with the approval of the local ethics committee.

Conflict of interest

All authors declare no conflict of interest.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Download
Titel
Verbal and non-verbal recognition memory assessment: validation of a computerized version of the Recognition Memory Test
Verfasst von
Elena Baruzzo
Stefano Terruzzi
Beatrice Feder
Costanza Papagno
Daniela Smirni
Publikationsdatum
22.12.2023
Verlag
Springer International Publishing
Erschienen in
Neurological Sciences / Ausgabe 5/2024
Print ISSN: 1590-1874
Elektronische ISSN: 1590-3478
DOI
https://doi.org/10.1007/s10072-023-07171-3
1.
Zurück zum Zitat Rabin LA, Spadaccini AT, Brodale DL et al (2014) Utilization rates of computerized tests and test batteries among clinical neuropsychologists in the United States and Canada. Prof Psychol Res Pract 45:368. https://doi.org/10.1037/a0037987CrossRef
2.
Zurück zum Zitat Marra DE, Hamlet KM, Bauer RM, Bowers D (2020) Validity of teleneuropsychology for older adults in response to COVID-19: A systematic and critical review. Clin Neuropsychol 34:1411–1452. https://doi.org/10.1080/13854046.2020.1769192CrossRefPubMedPubMedCentral
3.
Zurück zum Zitat Bauer RM, Iverson GL, Cernich AN et al (2012) Computerized neuropsychological assessment devices: joint position paper of the American Academy of Clinical Neuropsychology and the National Academy of Neuropsychology. Arch Clin Neuropsychol 27:362–373. https://doi.org/10.1093/arclin/acs027CrossRefPubMedPubMedCentral
4.
Zurück zum Zitat Parsons TD (2016) Clinical Neuropsychology and Technology. Springer International Publishing, ChamCrossRef
5.
Zurück zum Zitat Terruzzi S, Albini F, Massetti G, Etzi R, Gallace A, Vallar G (2023) The neuropsychological assessment of unilateral spatial neglect through computerized and virtual reality tools: A scoping review. Neuropsychol Rev 1-39. https://doi.org/10.1007/s11065-023-09586-3
6.
Zurück zum Zitat Aşkar P, Altun A, Cangöz B et al (2012) A comparison of paper-and-pencil and computerized forms of line orientation and enhanced cued recall tests. Psychol Rep 110:383–396. https://doi.org/10.2466/03.22.PR0.110.2.383-396CrossRefPubMed
7.
Zurück zum Zitat Woo E (2008) Computerized neuropsychological assessments. CNS Spectr 13:14–17. https://doi.org/10.1017/S1092852900026985CrossRefPubMed
8.
Zurück zum Zitat Crook TH, Kay GG, Larrabee GJ, Grant I, Adams K (2009) Computer-based cognitive testing. Neuropsychological assessment of neuropsychiatric and neuromedical disorders, pp 84–100
9.
Zurück zum Zitat Russell EW (2000) The application of computerized scoring programs to neuropsychological assessment. Clinician’s guide to neuropsychological assessment, pp 483–515
10.
Zurück zum Zitat Gur RC, Ragland JD, Moberg PJ et al (2001) Computerized neurocognitive scanning: I. Methodology and validation in healthy people. Neuropsychopharmacology 25:766–776. https://doi.org/10.1016/S0893-133X(01)00278-0CrossRefPubMed
11.
Zurück zum Zitat Schatz P, Browndyke J (2002) Applications of computer-based neuropsychological assessment. J Head Trauma Rehabil 17:395–410. https://doi.org/10.1097/00001199-200210000-00003CrossRefPubMed
12.
Zurück zum Zitat Sheppard LD, Vernon PA (2008) Intelligence and speed of information-processing: A review of 50 years of research. Pers Individ Dif 44:535–551. https://doi.org/10.1016/j.paid.2007.09.015CrossRef
13.
Zurück zum Zitat Iverson GL, Brooks BL, Ashton VL et al (2009) Does familiarity with computers affect computerized neuropsychological test performance? J Clin Exp Neuropsychol 31:594–604. https://doi.org/10.1080/13803390802372125CrossRefPubMed
14.
Zurück zum Zitat Cernich AN, Brennana DM, Barker LM, Bleiberg J (2007) Sources of error in computerized neuropsychological assessment. Arch Clin Neuropsychol 22:S39–S48. https://doi.org/10.1016/j.acn.2006.10.004CrossRefPubMed
15.
Zurück zum Zitat Hoskins LL, Binder LM, Chaytor NS et al (2010) Comparison of oral and computerized versions of the word memory test. Arch Clin Neuropsychol 25:591–600. https://doi.org/10.1093/arclin/acq060CrossRefPubMed
16.
Zurück zum Zitat Williams JE, McCord DM (2006) Equivalence of standard and computerized versions of the Raven Progressive Matrices Test. Comput Human Behav 22:791–800. https://doi.org/10.1016/j.chb.2004.03.005CrossRef
17.
Zurück zum Zitat Feldstein SN, Keller FR, Portman RE et al (1999) A comparison of computerized and standard versions of the Wisconsin Card Sorting Test. Clin Neuropsychol 13:303–313. https://doi.org/10.1076/clin.13.3.303.1744CrossRefPubMed
18.
Zurück zum Zitat French CC, Beaumont JG (1987) The reaction of psychiatric patients to computerized assessment. Br J Clin Psychol 26:267–278. https://doi.org/10.1111/j.2044-8260.1987.tb01360.xCrossRefPubMed
19.
Zurück zum Zitat Ozonoff S (1995) Reliability and validity of the Wisconsin card sorting test in studies of autism. Neuropsychology 9:491. https://doi.org/10.1037/0894-4105.9.4.491CrossRef
20.
Zurück zum Zitat Berger SG, Chibnall JT, Gfeller JD (1997) Construct validity of the computerized version of the Category Test. J Clin Psychol 53:723–726. https://doi.org/10.1002/(SICI)1097-4679(199711)53:73.0.CO;2-ICrossRefPubMed
21.
Zurück zum Zitat Campbell KA, Rohlman DS, Storzbach D et al (1999) Test-retest reliability of psychological and neurobehavioral tests self-administered by computer. Assessment 6:21–32. https://doi.org/10.1177/107319119900600103CrossRefPubMed
22.
Zurück zum Zitat Choca J, Morris J (1992) Administering the Category Test by computer: Equivalence of results. Clin Neuropsychol 6:9–15. https://doi.org/10.1080/13854049208404112CrossRef
23.
Zurück zum Zitat Tien AY, Spevack TV, Jones DW et al (1996) Computerized Wisconsin Card Sorting Test: comparison with manual administration. Kaohsiung J Med Sci 12:479–485PubMed
24.
Zurück zum Zitat Gelman A, Geurts HM (2017) The statistical crisis in science: How is it relevant to clinical neuropsychology? Clin Neuropsychol 31:1000–1014. https://doi.org/10.1080/13854046.2016.1277557CrossRefPubMed
25.
Zurück zum Zitat Yu K, Zhang S, Wang Q et al (2015) Development of a computerized tool for the chinese version of the montreal cognitive assessment for screening mild cognitive impairment. Int psychogeriatrics 27:213–219. https://doi.org/10.1017/S1041610214002269CrossRef
26.
Zurück zum Zitat Garre-Olmo J, Faúndez-Zanuy M, López-de-Ipiña K et al (2017) Kinematic and pressure features of handwriting and drawing: preliminary results between patients with mild cognitive impairment, Alzheimer disease and healthy controls. Curr Alzheimer Res 14:960–968. https://doi.org/10.2174/1567205014666170309120708CrossRefPubMedPubMedCentral
27.
Zurück zum Zitat Petilli MA, Daini R, Saibene FL, Rabuffetti M (2021) Automated scoring for a Tablet-based Rey Figure copy task differentiates constructional, organisational, and motor abilities. Sci Rep 11:1–19. https://doi.org/10.1038/s41598-021-94247-9CrossRef
28.
Zurück zum Zitat Massetti G, Albini F, Casati C et al (2023) Validation of “Neurit. Space”: Three Digital Tests for the Neuropsychological Evaluation of Unilateral Spatial Neglect. J Clin Med 12:3042. https://doi.org/10.3390/jcm12083042CrossRefPubMedPubMedCentral
29.
Zurück zum Zitat Coughlan AK, Hollows SE (1984) Use of memory tests in differentiating organic disorder from depression. Br J Psychiatry 145:164–167. https://doi.org/10.1192/bjp.145.2.164CrossRefPubMed
30.
Zurück zum Zitat Smirni D, Turriziani P, Oliveri M et al (2010) Standardizzazione di tre nuovi test di memoria di riconoscimento verbale e non verbale: uno studio preliminare. G Ital di Psicol 37:227–248. https://doi.org/10.1421/32178CrossRef
31.
Zurück zum Zitat Barvas E, Mattavelli G, Meli C et al (2022) Standardization and normative data for a new test of visual long-term recognition memory. Neurol Sci 43:2491–2497. https://doi.org/10.1007/s10072-021-05642-zCrossRefPubMed
32.
Zurück zum Zitat Smirni D, Smirni P, Di Martino G et al (2018) Standardization and validation of a parallel form of the verbal and non-verbal recognition memory test in an Italian population sample. Neurol Sci 39:1391–1399. https://doi.org/10.1007/s10072-018-3433-zCrossRefPubMed
33.
Zurück zum Zitat Novelli G, Papagno C, Capitani E, Laiacona M (1986) Tre test clinici di memoria verbale a lungo termine: taratura su soggetti normali. Archivio di psicologia, neurologia e psichiatria
34.
Zurück zum Zitat Conti S, Bonazzi S, Laiacona M et al (2015) Montreal Cognitive Assessment (MoCA)-Italian version: regression based norms and equivalent scores. Neurol Sci 36:209–214. https://doi.org/10.1007/s10072-014-1921-3CrossRefPubMed
35.
Zurück zum Zitat Brooke J (1996) SUS-A quick and dirty usability scale. Usability Eval Ind 189:4–7
36.
Zurück zum Zitat Koo TK, Li MY (2016) A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med 15:155–163. https://doi.org/10.1016/j.jcm.2016.02.012CrossRefPubMedPubMedCentral
37.
Zurück zum Zitat Lakens D, Scheel AM, Isager PM (2018) Equivalence Testing for Psychological Research: A Tutorial. Adv Methods Pract Psychol Sci 1:259–269. https://doi.org/10.1177/2515245918770963CrossRef
38.
Zurück zum Zitat Kim H-Y (2013) Statistical notes for clinical researchers: assessing normal distribution (2) using skewness and kurtosis. Restor Dent Endod 38:52–54. https://doi.org/10.5395/rde.2013.38.1.52CrossRefPubMedPubMedCentral
39.
Zurück zum Zitat Bangor A, Kortum P, Miller J (2009) Determining what individual SUS scores mean: Adding an adjective rating scale. J usability Stud 4:114–123
40.
Zurück zum Zitat Bilder RM (2011) Neuropsychology 3.0: Evidence-based science and practice. J Int Neuropsychol Soc 17:7–13. https://doi.org/10.1017/S1355617710001396CrossRefPubMed
41.
Zurück zum Zitat Parsey CM, Schmitter-Edgecombe M (2013) Applications of technology in neuropsychological assessment. Clin Neuropsychol 27:1328–1361. https://doi.org/10.1080/13854046.2013.834971CrossRefPubMed
42.
Zurück zum Zitat Adams JL, Myers TL, Waddell EM et al (2020) Telemedicine: a valuable tool in neurodegenerative diseases. Curr Geriatr reports 9:72–81. https://doi.org/10.1007/s13670-020-00311-zCrossRef
43.
Zurück zum Zitat Parikh M, Grosch MC, Graham LL et al (2013) Consumer acceptability of brief videoconference-based neuropsychological assessment in older individuals with and without cognitive impairment. Clin Neuropsychol 27:808–817. https://doi.org/10.1080/13854046.2013.791723CrossRefPubMedPubMedCentral
44.
Zurück zum Zitat Chaytor NS, Barbosa-Leiker C, Germine LT et al (2021) Construct validity, ecological validity and acceptance of self-administered online neuropsychological assessment in adults. Clin Neuropsychol 35:148–164. https://doi.org/10.1080/13854046.2020.1811893CrossRefPubMed
45.
Zurück zum Zitat Loh P, Donaldson M, Flicker L et al (2007) Development of a telemedicine protocol for the diagnosis of Alzheimer’s disease. J Telemed Telecare 13:90–94. https://doi.org/10.1258/135763307780096159CrossRefPubMed

Kompaktes Leitlinien-Wissen Neurologie (Link öffnet in neuem Fenster)

Mit medbee Pocketcards schnell und sicher entscheiden.
Leitlinien-Wissen kostenlos und immer griffbereit auf ihrem Desktop, Handy oder Tablet.

Neu im Fachgebiet Neurologie

Verzögert eine hochdosierte Influenza-Vakzine eine Demenz?

  • 10.04.2026
  • Demenz
  • Nachrichten

Eine hochdosierte Influenza-Vakzine geht mit einer verzögerten Demenzdiagnose einher. Darauf deutet eine Auswertung von US-Gesundheitsdaten hin. Besonders auffällig sind die Effekte in den ersten Monaten nach der Impfung.

Intensive Drucksenkung zeigt U-förmigen Nutzen nach Hirnblutung

Intensive Senkung eines erhöhten Blutdrucks kann nach einer intrazerebralen Blutung die funktionelle Erholung verbessern – mutmaßlich über eine Reduktion der Hämatomausdehnung. Offenbar hängt das aber vom Ausgangsvolumen ab, wie eine Analyse ergeben hat.

Steaks gegen Alzheimer

  • 02.04.2026
  • Demenz
  • Nachrichten

Da schmeckt das Rinderfilet gleich doppelt so gut: Fleisch beugt einer aktuellen Studie zufolge einer Demenz vor. Allerdings gilt das nur für ApoE4-Träger. Diese haben sich im Laufe der Evolution offenbar an einen hohen Fleischkonsum angepasst – und brauchen ihre Steak-Rationen.

Residuelle Tagesschläfrigkeit trotz guter Compliance

Dass Tagesschläfrigkeit (EDS) nicht zwingend ein Compliance‑Problem ist und selbst umfangreiche Therapieversuche an Grenzen stoßen können, zeigt der auf dem DGP‑Kongress vorgestellte Fall eines 50‑jährigen LKW‑Fahrers. Letztendlich gab der Mann seinen Beruf auf und benötigte weitere Therapien.

Update Neurologie

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.

Bildnachweise
Die Leitlinien für Ärztinnen und Ärzte, Hirn-MRT zeigt bilaterale subdurale Blutung (blaue Pfeile)/© Salvatore Perrone et al doi.org/10.1007/s00277-023-05392-2 unter CC-BY 4.0, Erschöpfter Mann im LKW/© Africa Studio / stock.adobe.com (Symbolbild mit Fotomodell)