Introduction:
Instruments for the assessment of syntax in Brazilian Portuguese remain scarce, despite their crucial role in identifying comprehension and production deficits in clinical populations. This study aimed to develop a computerized version of the Test for the Assessment of Syntactic Processing in Comprehension (TASComp) and to investigate validity evidence related to sentence reading comprehension among Brazilian adults.
Methods:
Procedures included stimulus adaptation for the PsyToolkit platform, content validation by expert judges, and psychometric analyses of a final version administered to 102 healthy adults, stratified by in-person (n = 45) and online (n = 57) administration modalities.
Results:
The TASComp demonstrated satisfactory content validity, with a Scale Content Validity Index (S-CVI) of 0.87 and adequate internal consistency (KR-20 = 0.66). Syntactic complexity effects were confirmed by significant differences in the results of reversible passives vs. irreversible (t = 3.08, p = 0.003, d = 0.31), object clefts vs. subject clefts (t = 5.30, p < 0.001, d = 0.53), and center-embedded vs. right-branching relatives for both subject clauses (t = 6.80, p < 0.001, d = 0.67) and object clauses (t = 3.52, p = 0.001, d = 0.35). Additionally, right-branching subject relatives were more accurate than right-branching object relatives (t = 3,78; p < 0.001, d = 0.37). No significant differences in accuracy were found between in-person and online administration modalities across all sentence types (p > 0.05). The beta regression model revealed that years of education (χ2 (1) 35 = 15.74, p < 0.001) and age (χ2 (1) = 7.98, p = 0.01) were significant predictors of comprehension accuracy, while the administration modality was not (χ2 (1) = 0.27, p = 0.61).
Discussion:
Validity evidence based on response processes was observed, as the instrument successfully discriminated between levels of syntactic complexity, with significant performance differences in structures that impose high working memory demands, such as object clefts and center-embedded and object relative clauses. No significant differences were found between administration modalities (in-person vs. online), indicating that both modes assess sentence reading comprehension without compromising psychometric quality. Furthermore, the findings demonstrate that age and years of education significantly impact syntactic processing and reading comprehension.
Conclusion:
TASComp demonstrates adequate validity of evidence and clinical relevance. The instrument shows promise for detecting subtle language deficits that are often overlooked by other assessments, due to its inclusion of a wide range of 46 complex syntactic structures.
1 IntroductionSyntactic processing encompasses the set of cognitive operations that support the incremental construction of grammatical structures during language comprehension and production (Blake et al., 2025; Caplan et al., 2016; Rodrigues et al., 2024). This process entails the hierarchical organization of constituents, the computation of dependency relations, and the integration of syntactic and semantic information in real time (Breder and Rodrigues, 2026; Caplan, 2016). Disruptions to these mechanisms may result in sentence-level comprehension deficits (Rodrigues et al., 2024).
Language assessments, specifically at the syntactic level, are highly sensitive to neurocognitive disorders and early changes associated with cognitive decline and dementia in adults (Gilardone et al., 2024; Ivanova and Dronkers, 2022; Kaltsa et al., 2024; Sung et al., 2020). In adult neurodevelopmental disorders, such as Attention-Deficit/Hyperactivity Disorder (ADHD) (Hill et al., 2025; Jepsen et al., 2025; Oliveira and Corrêa, 2023; Parks et al., 2022) and Developmental Language Disorder (Bishop et al., 2017; Blake et al., 2025; Corrêa and Augusto, 2013; Tucci and Choi, 2023), syntactic production and comprehension may likewise be impaired, with significant deficits manifesting at both sentential and discursive levels. Consequently, the assessment of syntactic processing should not be restricted to aphasic populations; rather, it warrants inclusion in comprehensive investigations of syntactic and semantic interface across a broad spectrum of adult clinical cohorts (Gilardone et al., 2024; Ivanova and Dronkers, 2022; Kaltsa et al., 2024; Murray, 2018; Thothathiri et al., 2023; Wallace et al., 2023).
The diagnostic value of syntactic assessments lies in the high processing demands of complex structures, whose performance is affected not only by linguistic deficits but also by reduced working memory capacity (Caplan et al., 2016; Murray, 2018; Poulisse et al., 2019; Sung et al., 2017). Although linguistic processing remains stable in aging (McKoon and Ratcliff, 2018; Silagi et al., 2023; Solomons et al., 2025), syntactic difficulties in processing complex sentences may be present due to decreased executive functions (Murray, 2018; Poulisse et al., 2019; Sung et al., 2020). These deficits, even in healthy aging, can serve as preclinical indicators of cognitive difficulties (Feigin et al., 2017; Kaltsa et al., 2024; Van Oers et al., 2018). Furthermore, education and lifelong literacy can serve as a protective factor for linguistic performance against cognitive decline (López-Higes et al., 2023; Riffo et al., 2025). These factors, essential for maintaining linguistic abilities (Pereira and Ortiz, 2022), can also stimulate the attentional and executive resources necessary for processing complex sentences (Akashi and Ortiz, 2018; Malcorra et al., 2022).
Sentence comprehension assessments must be grounded in psycholinguistic parameters that identify syntactic configurations associated with increased processing cost. Such costs are primarily linked to deviations from the canonical Subject–Verb–Object (SVO) order and to structures involving syntactic displacement (Rodrigues et al., 2024). Passive sentences exemplify this deviation by promoting the object of the corresponding active clause to subject position (“The fence was painted by the worker”; A cerca foi pintada pelo funcionário [Brazilian Portuguese, pt-BR]) and by demoting or omitting the agent (“The worker was fired last week”; pt-BR: O funcionário foi demitido na última semana), which disrupts the canonical mapping between thematic roles and grammatical), functions and increases reliance on syntactic parsing (Corrêa et al., 2017; Sung et al., 2020). Processing difficulty is amplified in reversible passives, where animacy does not constrain thematic role assignment (“The man was greeted by his friend”; pt-BR: O homem foi cumprimentado pela sua amiga); and in cases of passives with implausible semantic content (“The police officer was chased by the criminal”; pt-BR: O policial foi perseguido pelo bandido), which limit the use of semantic heuristics and force greater dependence on structural cues (Christianson et al., 2006; Ferreira et al., 2002). These demands are particularly taxing for populations that rely on good-enough processing strategies, such as older adults, who may prioritize semantic plausibility over detailed syntactic analysis, with evidence showing higher processing costs for passives in Brazilian Portuguese within these groups (Ferrari, 2017; Hübner et al., 2022; Souza, 2023).
Additional processing costs arise in constructions that involve filler–gap dependencies, such as cleft sentences and relative clauses. Cleft sentences consist of a matrix clause, typically introduced by the copular construction “it is” (in Brazilian Portuguese, “É”), and a relative clause that encoded the dependency associated with the focused constituent (Alves et al., 2015). The complexity of these sentences lies in the syntactic movement of the focused argument, leaving a trace in its original position (“It is the girl [who __ is helping the woman]”; É a menina que está sendo ajudada pela mulher [BP]) (Alves et al., 2015; Lambrecht, 2001). Relative clauses involve the movement of a constituent to an antecedent position in the sentence, leaving a gap in its original location immediately following the relative pronoun (The boy [who__ honored the teacher] is an excellent student; O menino que elogiou o professor é um excelente estudante). Object clefts and object relative clauses are consistently more demanding than their subject counterparts due to the greater linear and structural distance between the displaced constituent and its gap, increasing working memory load and susceptibility to interference (Forster and Corrêa, 2016; Friedmann et al., 2009; Gibson, 1998). Center-embedded relative clauses further exacerbate this cost by interrupting the main clause and requiring the maintenance of multiple unresolved dependencies, in contrast to right-branching structures, which impose lower processing demands (Gibson et al., 2005). Finally, sentences with perspective verbs introduce noncanonical mappings between syntactic position and thematic role, as the subject denotes a patient rather than an agent (“The patient received the diagnosis from the doctor”; pt-BR: O paciente recebeu o diagnóstico do médico.), increasing computational demands despite active syntax (Breder and Rodrigues, 2026). Together, these structures highlight the importance of controlling syntactic complexity when designing assessment tools sensitive to sentence-level processing deficits (Rodrigues et al., 2024).
Language assessments with different syntactic structures could be essential to identify deficits that compromise reading comprehension and the interpretation of complex messages, thereby affecting communicative competence in daily life. Although individuals often employ compensatory strategies, such as re-reading, slowing down, and minimizing distractions, processing deficits tend to surface specifically under conditions of increased syntactic complexity (Webster et al., 2023). These difficulties can significantly interfere with modern reading habits, particularly in the realm of digital communication, such as instant messaging, which requires the rapid processing of sentence structures (Kinsey et al., 2021; Lee and Cherney, 2022). In clinical settings, detailed syntactic assessment serves as a sensitive screening tool for subtle linguistic impairments that frequently go undetected in spontaneous conversation or simple comprehension tasks (Sung et al., 2020). Consequently, the heightened cognitive effort required for language processing can lead to reduced social participation, affecting both elderly populations and adults with acquired language disorders who rely on technology for interpersonal connection (Kinsey et al., 2021). Ultimately, a comprehensive evaluation of syntax is also fundamental for designing interventions in clinical settings, considering a detailed evaluation of language comprehension.
Despite robust evidence of deficits associated with high-cost syntactic structures, such as passive, relative, and cleft sentences, in clinical and aging populations, standardized instruments for evaluating sentence comprehension remain scarce in Brazil. Current assessment options primarily include the adapted versions of the Montreal-Toulouse Battery (MTL-BR) (Parente et al., 2016) and the Test for Reception of Grammar (TROG-2) (Carthery-Goulart et al., 2022; Pereira et al., 2009). Both instruments provide essential validity evidence for the Brazilian population but present some limitations. The MTL-BR, which evaluates different levels of language processing, includes only eight items assessing written sentence comprehension. By contrast, the TROG-2, designed to evaluate oral syntactic comprehension, exhibits an imbalance in the representation of sentence subtypes (e.g., discrepancies between reversible and irreversible items in blocks) and variability in constituent features, such as animate versus inanimate subjects. This stimulus heterogeneity within blocks may confound qualitative error analysis and hinder the precise assessment of discrete syntactic structures (Ferrari, 2022).
In this context, the Test for the Assessment of Syntactic Processing in Comprehension (TASComp) was developed (Ferrari, 2022). Designed in Brazilian Portuguese, the instrument employs age-appropriate stimuli depicting human agents performing various actions. Importantly, TASComp includes semantically implausible sentences to require the inhibition of good-enough processing strategies (Frances, 2024), thereby limiting reliance on semantic heuristics and ensuring that performance depends on syntactic parsing (Ferrari, 2022).
The TASComp was constructed based on psycholinguistics theory about processing cost of sentences (Ferrari, 2022). Regarding its structure, TASComp comprises five blocks, each representing a specific syntactic construction: active, passive, cleft, relative, and interrogative clauses. These blocks include subcategories reflecting varying levels of linguistic complexity, totaling 56 stimuli and encompassing 14 sentence types: subject and object relative clauses (in both right-branching and center-embedded forms); subject and object cleft sentences; reversible passives (classified as neutral or implausible) and irreversible passives; constructions with perspective verbs; simple active sentences (reversible and irreversible); and subject and object interrogatives. TASComp includes four items per sentence type.
The original TASComp version was designed for stimulus presentation using PowerPoint. Each trial presents a target sentence alongside a quadrant containing four image-based response options (A, B, C, and D). By keeping the sentence visible throughout the visual array, this design reduces working memory demands during syntactic processing (Tan et al., 2017), allowing for a more precise assessment of domain-specific linguistic impairments. Responses were recorded on a printed protocol, completed either by the participant during group administration or by the examiner in individual sessions (Ferrari, 2022).
The assessment score considers one point per correct response, which permits both a quantitative evaluation of cumulative accuracy in each block and in the total score of the task (Ferrari, 2022). Furthermore, the TASComp incorporates an analysis of error patterns to elucidate the nature of underlying linguistic impairments based on the respondent’s selections. Grammatical errors are characterized by thematic role of reversals, such as agent-patient inversion, which indicate specific syntactic parsing failures. Lexical errors involve the substitution of core lexical constituents, such as the selection of an alternative verb, while maintaining an interpretation consistent with the original thematic structure. Finally, mixed errors manifest as the simultaneous alteration of both syntactic configuration and lexical content, reflecting a multidimensional impairment across various domains of linguistic processing (Ferrari, 2022).
The applicability of TASComp was evaluated in 43 adults aged 18 to 57 years (M = 23.41, SD = 8.76) and 10 older adults aged 60 to 74 years (M = 65.50, SD = 5.17), all with a university degree (Ferrari, 2022). In the adult group, the mean number of correct responses was 54.63 out of 56 sentences (approximately 97.50%), with eight participants making errors on relative clauses (the structure with the highest error rate). The older adult group achieved a mean of 51.60 correct responses (approximately 92.40%), with errors predominantly occurring in relative clauses and across the test as a whole (Ferrari, 2022). However, the small sample of older adults (n = 10) limited the number of comparisons that could be made between groups.
Although TASComp is a promising tool for assessing adults’ syntactic comprehension, its psychometric properties have not yet been analyzed. Developing a computerized version would enable administration in both face-to-face and remote/online settings, supporting teleneuropsychology and interdisciplinary applications (Alva et al., 2025; Serrano-Juárez et al., 2023). In this context, it is essential to validate the test for online use and to examine whether its psychometric properties differ from the traditional pencil-and-paper version. Current research indicates strong correlations between remote and in-person assessment results, with differences generally lacking significant clinical impact across most tests analyzed (Brown et al., 2024; Chapman et al., 2021; Krynicki et al., 2023).
This study aimed to adapt TASComp to a computerized online version and to investigate its psychometric properties. The specific objectives were: (a) to design the computerized version of TASComp, (b) to analyze evidence of content validity through expert judgment, (c) to test the instrument in a pilot study, and (d) to examine evidence of validity based on internal consistency, response processes across administration modalities (in-person and online) and syntactic complexity, as well as relationships with external variables (age and years of education) in each modality. Overall, the study seeks to contribute a Brazilian computerized instrument for assessing the syntactic level of written comprehension in adults.
2 Materials and methodsThe computerized version of TASComp was developed in six stages: (1) adaptation of stimuli and data-recording procedures; (2) expert review; (3) revisions; (4) pilot study; (5) refinement of administration instructions; and (6) analysis of the psychometric properties of the final version. Each stage involved distinct participants and procedures, as detailed below. The study was approved by the institutional ethics committee (approval no. 7.719.167).
2.1 Step 1: stimulus adaptation for computerized version of TASCompAt this stage, the TASComp stimuli was designed into the PsyToolkit (Stoet, 2016). This platform was selected for its ability to present stimuli in an assessment design, record participants’ responses, and reaction times. The interrogative sentences block contained in the original version (Ferrari, 2022) was excluded due to the variability in stimulus presentation. The online version proposed contains 48 items organized into five sentence types: active (12 items), passive (12 items), relative (16 items), and cleft sentences (8 items). Two training stimuli were included. Table 1 shows examples of sentences from each of the TASComp blocks.
Type of syntactic clauseBlockExample (pt-BR)Example translated (EN)Irreversible activeActive sentencesA menina está lendo o livro.The girl is reading the book.Reversible activeActive sentencesA menina está puxando o menino.The girl is pulling the boy.Sentences with perspective verbActive sentencesO rapaz está ganhando o presente da moça.The boy is receiving a gift from the girl.Irreversible passivesPassive sentencesO quadro está sendo pintado pelo rapaz.The painting is being painted by the boy.Reversible passivesPassive sentencesA mulher está sendo consolada pela menina.The woman is being comforted by the girl.Semantic implausible reversible passivesPassive sentencesO repórter está sendo entrevistado pela médica.The journalist is being interviewed by the doctor.Object cleftsCleftsÉ o soldado que o astronauta está empurrando.It is the soldier that the astronaut is pushing.Subject cleftsCleftsÉ o menino que está abraçando o homem.It is the boy who is hugging the man.Right-branching subject relativeRelative clausesO menino cutuca o homem que está quicando a bola.The boy pokes the man who is bouncing the ball.Center-embedded subject relativeRelative clausesO rapaz que filma a moça está descendo a escada.The boy that films the girl is going down the stairs.Right-branching object relativeRelative clausesA figura certa mostra o homem que a mulher está consolando.The correct figure shows the man that the woman is consoling.Center-embedded object relativeRelative clausesA menina que a mulher observa está subindo o degrau.The girl that the woman observes is going up the step.Types of syntactic clauses in TASComp and examples of items in Brazilian Portuguese (pt-BR) and their translation into English.
Stimuli were randomized to minimize performance differences attributable to item difficulty, independent of block. An instruction manual was developed to standardize administration across evaluators, and task instructions were embedded in the computerized platform. Finally, a qualitative questionnaire was administered at the end of the task to assess participants’ response strategies and motivation.
2.2 Steps 2 and 3: content validity and reformulations2.2.1 ParticipantsIn Step 2, 22 researchers from a completed item evaluation form administered via Google Forms, including undergraduate and graduate students in Psychology and Linguistics. The research participants consisted entirely of Brazilian graduate students from the state of Rio de Janeiro. To establish content validity, the Content Validity Index (CVI) aggregates expert ratings regarding item relevance and clarity (Yusoff, 2019). The index is derived from the proportion of experts endorsing an item’s essentiality. This quantitative validation is crucial for minimizing measurement error, ensuring that the instrument accurately operationalizes the intended theoretical construct (Lynn, 1986).
For the calculation of the CVI, data from seven judges meeting a minimum qualification of a master’s degree and affiliated with graduate programs in Clinical Psychology or Language Studies were considered. The panel comprised one master’s-level and two doctoral researchers in Literature and Language Studies with expertise in psycholinguistics, one doctoral researcher in Clinical Psychology, the panel was comprised of three master’s-level researchers in Clinical Psychology with experience in neuropsychological assessment and psychometrics.
2.2.2 Materials and proceduresTo assess the content validity of the online TASComp, judges reviewed all 48 items, each comprising the sentence and four image-based response options. For each item, they provided dichotomous (“Yes”/“No”) responses to two criteria: (1) whether the target image adequately represented the sentence content, and (2) whether the distractors were sufficiently differentiated from the target for accurate sentence interpretation. In the second section of the form, judges rated the test instructions, layout, and qualitative feedback questionnaire on a 4-point Likert scale (1 = totally inadequate, 4 = quite adequate), with space for additional comments.
2.2.3 Statistical analysesMicrosoft Excel (Md Lukmanul Hakim et al., 2025; Microsoft Corporation, 2024) was used to calculate the indices according to Yusoff (2019) tutorial. For the calculation for the content validity index (CVI) Item and Scale-CVI (Yusoff, 2019), a score of 1 was assigned for “Yes” and zero for “No” for the dichotomous questions. The CVI was calculated as the proportion of agreement among the judges, with values above 0.83 considered acceptable recommended by Lynn (1986), considering six to eight experts’ evaluation. For Likert-scale responses, values of 1 and 2 were recorded as zero (inadequate), and 3 and 4 were recorded as 1 (adequate) for calculation purposes. The responses of undergraduate researchers were compiled for qualitative evaluation.
2.3 Steps 3 and 4: pilot study and final version design2.3.1 ParticipantsA pilot study was conducted with six brazilian adults, between 18 to 48 years, to evaluate the usability and functionality of the computerized TASComp after the item modifications. Participants were classified as cognitively healthy based on the self-reported absence of any history of psychiatric, neurological, or cognitive disorders in a clinical interview. The pilot study sample comprised Brazilian residents of Rio de Janeiro who were enrolled in undergraduate or Master’s degree programs. Following the pilot study, the computerized version of the TASComp was reformulated for the final version.
2.3.2 Materials and proceduresA convenience sample was recruited via social media. After providing informed consent, participants completed a sociodemographic and health questionnaire, followed by the TASComp. Four participants were assessed online via Zoom, with the evaluator sharing their screen to present instructions, training, and test items. Two participants were assessed in person in the research laboratory, with the task displayed on the researcher’s computer. In both modalities, the test was administered by the researcher on a computer or laptop with a minimum 14-inch screen in a quiet environment, ensuring participant confidentiality; microphones and headphones were used in the online condition. Participants responded verbally by indicating the letter corresponding to their chosen option (A–D), which was recorded by the evaluator. Based on accuracy data and qualitative usability feedback, the final version of the TASComp was developed.
2.3.3 Statistical analysesIn Step 3, analyses included computing each participant’s total number of correct responses and extracting results using RStudio and the Tidyverse package (R Core Team, 2025; Wickham et al., 2019), along with the collection of qualitative data. Qualitative information was obtained from evaluator–participant conversations and focused on feedback regarding the test’s design and interface, clarity of the final instructions, and overall platform usability from the evaluator’s perspective.
2.4 Step 5: psychometrics properties of the online version of TASComp2.4.1 ParticipantsBased on sample calculations for multiple linear regression analysis with two predictors, the minimum sample size required to conduct the study was 44 individuals, considering a probability of α = 0.05 and statistical power of 1−β = 0.95. The software used for the sample calculation was G*Power (Faul et al., 2007). The final testing sample comprised 102 Brazilian adults, all residents of Brazil from different states across the country, native speakers of Brazilian Portuguese with at least 5 years of formal education and no self-reported neurological or psychiatric diagnoses. Exclusion criteria included a history of self-reported neurodevelopmental disorders, psychiatric disorders without medical treatment or with partial symptom remission, illiteracy, and uncorrected hearing or visual impairments. Participants were divided into two groups: 45 completed the in-person assessment and 57 completed the remote assessment. Participants assessed online were younger than those assessed in person. A summary of the sample’s sociodemographic characteristics is presented in Table 2.
VariableIn-person (n = 45)Online (n = 57)Total (n = 102)t/(χ2)pAge (years) M (SD)51.20 (20.00)42.6 (16.30)45.8 (18.06)2.650.009*Years of study M (SD)17.40 (4.75)18.00 (4.05)17.80 (4.36)−0.640.53Gender %Woman24.5033.3057.800.170.68Men19.6022.5042.20Family incomea %A—21,826.7412.7016.5029.105.000.42B1—10,361.407.6021.5029.10B2—5,755.2310.1010.1020.30C1—3,276.766.305.1011.40C2—1,965.875.102.507.60D, E—900.601.301.302.50Sociodemographic characteristics of the participants.
aSocioeconomic classification based on the possession of consumer goods and education level (Brazil Economic Classification Criteria/ABEP).
2.4.2 Materials and proceduresThe sample was non-probabilistic and convenience-based, recruited via university announcements, social media, and referrals. Data were collected individually by the principal researcher or by trained undergraduate students in Psychology and Neuroscience. Procedures followed those described in the pilot study. For online assessments, participants joined a Zoom video call, were expected to be familiar with technology or have assistance and viewed the evaluator’s shared screen. Both participants and evaluators were in a quiet, interruption-free environment, using headphones to minimize noise and ensure confidentiality.
In-person TASComp assessments were conducted on computers in research laboratories or private rooms. Participants sat beside the examiner, facing the screen at an appropriate distance for clear visibility. The examiner recorded the participant’s responses after they indicated their choice verbally or by pointing to the corresponding letter. Following the task, the examiner completed a qualitative questionnaire, documenting response mode and interaction, as described in the pilot study.
2.4.3 Statistical analysesDescriptive analyses of the sample were performed, including normality testing (Shapiro–Wilk) for sociodemographic variables (age and years of schooling) and TASComp performance, both overall and by block (Richardson and Machan, 2021). Group differences by administration modality (in-person vs. online) were assessed using Student’s t-test for age and years of schooling and the Chi-square test (χ2) for gender and family income. TASComp data were extracted using RStudio and the Tidyverse package (R Core Team, 2025; Wickham et al., 2019), recording correct responses and constructing dataframes for analysis in Jamovi (The Jamovi Project, 2025). Extracted data included item-level accuracy, total correct responses per participant, and correct responses by syntactic block. Means and standard deviations were calculated for each block and the total test score.
Evidence of validity based on internal structure was assessed using the Kuder–Richardson 20 coefficient (KR-20) (Kuder and Richardson, 1937; Zhang, 2022) for the full scale. Validity evidence based on response processes was examined by comparing performance differences associated with the structural complexity of each test block using paired-samples t-tests. Manipulated syntactic factors included: reversibility in active and passive sentences (reversible vs. irreversible); clause type in relative and cleft sentences (subject vs. object); relative clause position (center vs. right-branching); and semantic plausibility in reversible passive sentences. Additionally, mean performance per block across syntactic complexity levels was compared between administration modalities (in-person vs. online) using independent-samples t-tests. All t-tests were conducted in SPSS 27 (IBM Corp, 2020) with Bias-Corrected and Accelerated Bootstrap (BCa) resampling (1,000 samples) to ensure robust estimates (Field, 2017; Haukoos and Lewis, 2005).
To examine criterion validity, the influence of age and education on the total proportion of correct TASComp responses was analyzed, with administration modality (in-person vs. online) included as a factor. A Beta Regression Model (Schmid et al., 2013) was implemented in the GAMlj3 module of Jamovi (Gallucci, 2019) using a logit link function to model the overall proportion of accuracy (0–1) within a Generalized Linear Model framework. Beta regression is appropriate for dependent variables expressed as proportions, offering a more suitable alternative to linear regression for non-parametric data (Shapiro–Wilk < 0.001; Kerby, 2014). Robustness of parameter estimates was assessed via 95% confidence intervals using the BCa with 1,000 resamples (Haukoos and Lewis, 2005).
3 Results3.1 Step 1: stimulus adaptation for online computerized version of TASCompThe initial phase of the adaptation process yielded the elaborated instructions specifically designed for the computerized version of TASComp. The computerized TASComp must be administered and recorded solely by the evaluator and is not suitable for self-administration; all test interactions are performed using the evaluator’s mouse. Both remote and in-person assessments require a quiet, well-lit, and interruption-free environment. Equipment requirements include a computer with a minimum 14.6-inch screen and an internet connection of at least 20 Mbps. The computer should be placed on a flat surface at eye level, maintaining an approximate viewing distance of 60 cm from the participant.
The computerized TASComp is administered in four sequential steps: (1) review of instructions with the participant, (2) two training items, (3) administration of the main test comprising 48 sentences, and (4) completion of a qualitative questionnaire by the evaluator. Designed as a silent reading task, participants indicate their responses from the training stage onward either verbally (stating the letter of the chosen answer) or pointing. Each item provides feedback on correctness, with a maximum of five errors allowed per training item; exceeding this limit results in test termination. Participants are informed that answers cannot be changed once submitted and are encouraged to respond promptly, as the test also measures processing speed (time per item).
The estimated duration of the TASComp is approximately 15 min, although there is no strict time limit for completion. An example of the steps preceding the training phase is shown in Figure 1.

Computerized TASComp instructions and training example on the PsyToolkit platform.
Upon completion of the 48 items, a screen indicates the end of the task. The evaluator then stops screen sharing or removes the display from the participant’s view to complete a qualitative questionnaire for evaluator use only. The questionnaire records the participant’s primary response mode (verbal or pointing), including responses mediated by assistants if applicable. It also documents whether sentences were read aloud, instances of re-reading, and self-corrections (errors corrected before proceeding to the next item). Additionally, the evaluator notes the participant’s motivation, collaboration, and any relevant behavioral observations, providing essential clinical context for interpreting results.
The final TASComp version was published via a PsyToolkit link accessible only to study researchers. A sample link containing three test items was also created to train administrators on instructions and platform use.
3.2 Steps 2 and 3: content validity and reformulationsThe Content Validity Index (CVI) was calculated for each item and for the overall scale. Most items showed an I-CVI of 1.00. One item containing a perspective verb, “The boy is winning the girl’s gift”, received an I-CVI of 0.57. All distractor images achieved I-CVI values above 0.83. For the assessment, the Scale-CVI (S-CVI) was 0.98 for correct answers and 0.99 for distractor distinctiveness. The test layout and final qualitative questionnaire were considered adequate (CVI = 0.86), whereas the test instructions required clarification (CVI = 0.71). Overall, considering instructions, layout, and the evaluator’s qualitative questionnaire, the scale achieved an S-CVI of 0.87.
In the qualitative analysis, judges’ comments and results were compiled and evaluated according to technical and theoretical criteria, with suggestions either accepted or discarded. Following an expert review, the image of the item with the lowest I-CVI was modified for clarity. The image was modified by adding a birthday hat to better identify the character receiving the gift. This revised image was approved by the test authors, resulting in a final TASComp version deemed satisfactory by both expert reviews and authors.
3.3 Steps 3 and 4: pilot study and final versionThe pilot study indicated high acceptability and overall comprehension of the TASComp task. Participants demonstrated familiarity with the online platform and a clear understanding of instructions, regardless of age or prior technology experience. Acceptability was further supported by the absence of participant questions during the task and evaluators’ confirmation of the platform’s usability.
Performance data showed that most items were answered correctly, with
Comments (0)