Adolescent depression is a significant global public health concern and a leading contributor to disability among young people worldwide [,]. It is consistently linked to long-term psychosocial difficulties in adulthood, including persistent mental health issues, lower academic achievement, impaired social functioning, and unstable employment [,]. These challenges are particularly acute in low- and middle-income countries (LMICs), where recent studies from South Africa and Uganda estimate that up to 20% of adolescents experience depressive disorders [-]. In these settings, adolescent depression is also associated with substance use, sexual risk behaviors, and entrenched poverty and inequality, further exacerbating social and economic burdens [-]. Although effective treatments exist [,], access remains limited in LMICs due to shortages of trained professionals, under-resourced health systems, and stigma [,]. These barriers are often more severe for adolescents than for adults because youth mental health services, policies, and resources are frequently underdeveloped, contributing to substantial unmet need [-].
Promise and Limits of Digital Mental HealthDigital mental health interventions (DMHIs), such as mobile apps and web-based platforms, may offer scalable, cost-effective solutions to bridge access gaps [-]. By delivering therapeutic content remotely, DMHIs can help overcome workforce shortages and geographic and logistical barriers to care [,]. The feasibility of DMHIs in LMICs is supported by increasing access to mobile phones, including smartphones, and rising digital literacy among young people. National survey data from several Southern African countries indicate that mobile phone ownership among youth aged 15 to 24 years ranges from 32% in Malawi to 83% in Eswatini and is trending upward []. In South Africa, recent data from urban adolescents indicate that more than 85% report owning a smartphone and nearly 70% use the internet to seek health-related information []. These patterns suggest that mobile-based interventions have become a more viable strategy for reaching adolescents with mental health support. Realizing this potential, however, requires attention to equity, contextual relevance, and inclusive design, priorities that are increasingly underscored in the digital health promotion literature [].
Human Support and Task-SharingDespite this potential, digital-only interventions face important limitations, including high attrition and lower-than-expected effectiveness outcomes [,]. Meta-analytic evidence suggests that blended models, which combine technology with real-time communication, social connection, and accountability [,], improve engagement, adherence, and clinical outcomes [-]. Task-sharing approaches have emerged as a promising strategy to address workforce gaps in LMICs [,]. In these models, nonspecialists, defined as individuals without formal professional training in mental health (eg, lay counselors or community health workers), are trained and supervised to provide low-intensity psychosocial support, often alongside digital tools []. A newer variation within task-sharing is the peer support model. Peer supporters are nonspecialists who share similar sociodemographic characteristics (eg, age, ethnicity, or cultural background) and/or lived experience of mental health challenges with the target population []. Among adolescents, such shared identity may enhance trust, relatability, and engagement, potentially overcoming barriers often encountered in traditional clinical settings [-].
Training, Supervision, and FidelityAlthough blended approaches have demonstrated success across diverse settings [,], variations in training and supervision can affect their effectiveness []. Central to this challenge is fidelity, defined as providers’ adherence to intervention protocols and their competence or skill in delivering the intervention [,]. Fidelity influences outcomes across psychological treatments in both routine clinical practice and controlled trials [-]. Maintaining fidelity is especially critical when interventions are delivered by nonspecialists, given their less extensive mental health training, clinical experience, and the potential impact on treatment effectiveness [,]. At the same time, overly rigid protocols can constrain provider competence and authentic engagement with participants, necessitating a balance between adherence and flexibility.
Structured training programs for nonspecialist health care workers and lay counselors can improve the effectiveness of psychological treatments [,]. Ongoing supervision supports sustained fidelity and provider competence over time []. However, evidence on how best to design and deliver training and supervision programs for peer-supported digital interventions remains limited, particularly in low-resource settings. The rapid expansion of digital technologies creates opportunities to provide training and supervision remotely, potentially increasing the scalability and reach of treatments [,]. Rigorous evaluation of the feasibility, acceptability, and outcomes of such models can inform program adaptation, implementation, and scale-up. This is particularly important in relation to fidelity and downstream intervention effectiveness.
Study Context and ObjectiveKuamsha is a gamified mobile app that delivers behavioral activation (BA) through interactive narrative stories []. Users choose between 2 storylines, each comprising 6 sequential modules. Each module takes approximately 15 to 20 minutes and introduces a core BA skill (further details are available in separate publications [,]). The intervention schedule comprised 6 modules over 11 weeks. To enhance engagement and reinforce BA concepts, trained and supervised peer mentors made brief weekly telephone calls to participants. Each adolescent was paired with a peer mentor and received up to 7 calls: 1 introductory call and 1 call per module. We reported outcomes of Kuamsha among users elsewhere [,].
The objective of this study was to evaluate the feasibility, acceptability, and fidelity of peer mentor training and supervision among delivery agents (ie, peer mentors) in South Africa and Uganda.
This mixed methods evaluation of the peer mentor training and supervision program was embedded within 2 related feasibility studies conducted in South Africa and Uganda. In South Africa, the evaluation was nested within the digital delivery of BA therapy to overcome depression and facilitate social and economic transitions of adolescents (DoBAt) study, a 2-arm pilot randomized controlled trial assessing the feasibility, acceptability, and preliminary effectiveness of the Kuamsha app with peer support []. The DoBAt study was prospectively registered with the Pan African Clinical Trials Registry (PACTR202206574814636). In Uganda, the evaluation was conducted within Ebikolwa n’Empisa, a single-arm feasibility study of the Kuamsha app with peer support, adapted for a peri-urban Ugandan context [].
This study focused specifically on evaluating the feasibility, acceptability, and fidelity of the peer mentor training and supervision model across both settings, using a convergent mixed methods design. Quantitative and qualitative data were collected and analyzed simultaneously with equal emphasis. The findings were integrated using a merging approach to provide complementary insights into feasibility, acceptability, and fidelity (see the GRAMMS [Good Reporting of A Mixed Methods Study] checklist; ).
Ethical ConsiderationsEthical approval was obtained from institutional review boards in South Africa (University of the Witwatersrand Human Research Ethics Committee, MED20-05-01, and the Mpumalanga Province Research Committee), Uganda (Makerere University School of Public Health [HDREC750] and the Uganda National Council for Science and Technology [HS724ES]), and the United Kingdom (Oxford Tropical Research Ethics Committee, OxTREC 72‐19 and OxTREC 34‐20).
Peer mentors were provided with written information about the study objectives, procedures, potential risks and benefits, and their rights as participants, including the voluntary nature of participation and their right to withdraw at any time without consequence to their employment. Written informed consent was obtained prior to participation. Adolescents provided informed assent or consent, and parental consent was obtained for their participation and for the recording of peer mentor calls.
Confidentiality and privacy were protected throughout the study. Audio recordings and transcripts were deidentified and stored on secure, password-protected institutional servers accessible only to authorized members of the research team. Quotations included in this manuscript have been anonymized to prevent identification.
Peer mentors received reimbursement for their role in delivering the intervention, as described in Table 2; no additional compensation was provided specifically for participation in this evaluation. Risks arising during the study were managed in accordance with a predefined Risk Management Protocol [], with supervisors conducting assessments and facilitating referrals to local services when necessary.
Study SettingIn South Africa, the DoBAt study [,] took place in the Bushbuckridge subdistrict of the Mpumalanga Province, within the Health and Socio-Demographic Surveillance System study area of the South African Medical Research Council/Wits Rural Public Health and Health Transitions Research Unit (Agincourt) []. Bushbuckridge is a rural yet densely populated region located near the Mozambique border and is home to approximately 1 million people. Most adolescents attend secondary school, and youth literacy rates are high (91%‐95.7%); however, 36.8% of young adults (18‐30 y) are not employed, in tertiary education, or in training [,].
In Uganda, the Ebikolwa n’Empisa study [] was implemented by BRAC Uganda in catchment areas mapped around the Katabi town in Wakiso District, a peri-urban region approximately 30 km from the capital, Kampala, with an estimated population of 170,000 []. Approximately half of adolescents attend secondary school; however, given the predominance of subsistence farming, not employed, in tertiary education, or in training rates are around 15% [].
In both contexts, access to mental health services remains sparse, with specialist mental health care usually concentrated in district or regional hospitals. Accessing these facilities often requires lengthy travel, commonly a 2-hour journey by public transportation, and families face additional barriers due to transport costs and the opportunity costs of lost work or schooling [,]. The scarcity of mental health human resources further exacerbates these access challenges. South Africa has an estimated 1.3 psychiatrists and 8.5 mental health nurses per 100,000 people, while Uganda has fewer than 0.1 psychiatrists and 1.4 mental health nurses per 100,000, with most based in Kampala [,]. Given these constraints, we adopted a peer-mentor model to complement the Kuamsha app and extend the reach of psychological treatments to low-resource environments.
Peer Mentor Training and SupervisionDevelopment of the Peer Mentor ProgramThe peer mentor training and supervision program was codeveloped through a series of workshops. First, we conducted 6 online workshops with 8 mental health professionals (HOM, MGC, CL, TS, EK, AS, BM, and JLG-O) to develop a draft peer mentor manual. Next, we held in-person workshops with youth with lived experience in South Africa (4 workshops with 6‐8 participants each) and Uganda (6 workshops with 8‐10 participants each) to refine the program’s content and structure. Specialists from the United Kingdom, United States, and Uganda (HOM, AS, MGC, EK) provided expert input on BA. The training program drew from established models of peer-to-peer coaching, particularly the Screening and Treatment for Anxiety and Depression program developed by MGC at University of California, Los Angeles [,]. The co-design process shaped the structure of the program, which was iteratively refined to (1) build core competencies required to deliver telephone-based support to Kuamsha users and (2) optimize cultural and contextual suitability. outlines the development process.
The co-design process shaped the role specifications for peer mentors. The specifications focused on recruiting supporters close in age to, but slightly older than, adolescent users (15‐19 years), typically 18 to 30 years, as age proximity was viewed as key to relatability and shared experiences of adolescence. Additional requirements included fluency in the local language and familiarity with the local culture. Other qualities included respect for confidentiality; a basic understanding of mental health problems; and a nonjudgmental, supportive attitude. The lived experience of mental health challenges was also considered valuable for empathy and rapport; however, we did not conduct formal assessments nor require mentors to disclose personal experiences of depression due to concerns about mental health–related stigma in South Africa and Uganda.
Recruitment of Peer MentorsIn South Africa, we recruited applicants through advertisements circulated via university psychology and social work departments. A total of 19 Xitsonga-speaking students or recent graduates with a bachelor’s degree in psychology or social work participated in training. In Uganda, 4 youth workers (3 women and 1 man) were trained as peer mentors. Their academic backgrounds included counseling, guidance and psychology, and library and information science. All were fluent in Luganda.
Training Content and FormatTraining was conducted online between May and November 2021, focusing on both program-specific competencies (ie, delivering BA telephone-based support) and nonspecific skills needed to establish a therapeutic alliance (). In addition to these core competencies, peer mentors received training on the study context, research ethics, and risk management. The training included didactic elements (presentations) and interactive components (skills rehearsal, role-plays) and used multimedia materials, including video content, learner workbooks, and structured call sheets.
Table 1. Core competencies of the peer mentor program.Competency domainObservable behavior or evidenceProgram-specific competenciesBehavioral activation: knowledge and skillsExplains how avoidance and withdrawal reinforce low moodWeekly online group supervision sessions (1 hour in South Africa; 30 minutes in Uganda) encouraged reflection, addressed challenges, and supported ongoing skills development. To optimize fidelity and enhance peer mentor competencies, all calls were recorded. Supervisors reviewed selected recordings and provided structured feedback to peer mentors. In both settings, supervisors were mental health professionals (a clinical psychologist in South Africa and a psychiatric clinical officer in Uganda). Supervisors conducted additional brief reflective check-ins with individual peer mentors to monitor well-being and reduce the risk of burnout.
Delivery of Training and SupervisionThe content and materials used for training and supervision were standardized across sites; however, delivery was adapted to account for local resource availability and logistical considerations related to the study designs. For example, in South Africa, due to the wider geographical distribution of peer mentors, training was conducted entirely remotely via Microsoft Teams, with no in-person contact between peer mentors or with their trainers. In Uganda, peer mentors met at the BRAC office for online training delivered via Microsoft Teams. We have summarized the key differences in .
Table 2. Delivery of the peer mentor program in South Africa and Uganda.Program characteristicSouth AfricaUgandaPeer mentors trained194Peer mentors delivering program134Employment periodPart-time, approximately 12 monthsFull-time, approximately 3 monthsReimbursementApproximately US $123 per month per mentor; mobile phone credit and data vouchers providedUS $100 per month per mentorTraining hours3678Trainer-trainee ratio1:10 (1:5 during intervention period)1:2Supervisor characteristicsClinical psychologistPsychiatric clinical officerA TIDieR (Template for Intervention Description and Replication) checklist describing the training and supervision components (content, dose, providers, mode, tailoring, modifications, and fidelity) is provided in .
OutcomesFeasibility and AcceptabilityFeasibility and acceptability were assessed quantitatively using supervisors’ recruitment and retention records and training and supervision attendance registers. Peer mentors also completed a structured 36-item Peer Mentor Training Evaluation questionnaire at the end of training (). Focus group discussions () with peer mentors in South Africa (n=10) and Uganda (n=4) and follow-up in-depth interviews () with peer mentors in South Africa (n=9) provided further insights into satisfaction, perceived value, implementation challenges, and opportunities for improvement.
FidelityPosttraining CompetenceWe assessed peer mentors’ competence post training using a posttraining assessment developed for this study ().
Several established competency frameworks guide the training and assessment of nonspecialist mental health providers, including the Enhancing Assessment of Common Therapeutic Factors and the Working With Children – Assessment of Competencies Tool, both developed within the World Health Organization’s Ensuring Quality in Psychological Support initiative [-]. These frameworks emphasize the assessment of core, transdiagnostic counseling competencies (eg, communication skills, ethical practice) that are applicable across psychological interventions and delivery contexts. While these tools provide a robust foundation for evaluating nonspecific competencies, we intentionally developed a program-specific assessment to capture competencies central to the peer mentor role in this intervention, including BA-specific skills (eg, activity monitoring, goal setting) and structured support tasks required to accompany a digital treatment.
Intervention Period FidelityIndependent assessors evaluated peer-mentor fidelity during intervention delivery. In South Africa, approximately 10% of calls were randomly selected using proportional sampling, stratified by peer mentor, excluding introductory, incomplete, corrupted, or termination calls. In Uganda, we randomly selected 1 recording per peer mentor per week.
Recordings were evaluated using a supervisor feedback form that assessed competence (score out of 30) and adherence (scored out of 37). Fidelity scores were calculated as the sum of adherence and competence scores ().
To ensure interrater reliability, 2 independent assessors double-rated 5 recordings in South Africa and 7 recordings in Uganda. The assessors resolved discrepancies through discussion and, once they reached consensus on their scoring, proceeded independently.
Data AnalysisQuantitative data analyses were conducted using Microsoft Excel and Stata 19 SE (StataCorp LLC). Descriptive statistics summarized recruitment, retention, attendance, and implementation outcomes. Continuous variables are presented as means and SDs or medians and IQR, as appropriate, and categorical variables as counts and percentages. To compare posttraining competence scores between South Africa and Uganda, we used the Wilcoxon rank sum test.
Intervention period fidelity (0‐37), adherence (0‐7), and competence (0‐30) were scored at the session level and converted to percentages of the maximum possible score to facilitate comparison across outcomes. Because multiple sessions were delivered by the same peer mentor, observations were clustered within providers. To account for this hierarchical structure, we fitted linear mixed-effects regression models with a random provider intercept and included site (Uganda vs South Africa) as a fixed effect. Adjusted marginal means and 95% CI were estimated using postestimation margins commands, and absolute percentage-point differences between sites were derived from pairwise comparisons of marginal predictions. Intraclass correlation coefficients were calculated from model variance components to quantify provider-level clustering.
Given that only 1 implementation site was included per country, comparisons reflect site-level differences rather than country-level effects. Statistical significance was defined as P<.05 (2-sided).
For the qualitative component, experienced bilingual research assistants translated audio recordings from focus group discussions and in-depth interviews into English and transcribed them using a naturalized verbatim approach. Two researchers then analyzed transcripts thematically, following Braun and Clarke’s 6-phase analytical approach [-]: (1) familiarization with the data, (2) generating initial codes, (3) generating themes, (4) reviewing potential themes, (5) defining and naming themes, and (6) producing the report. We used NVivo 10 Software (QRS International) for data management and coding.
The qualitative analysis was informed by a contextualist epistemological position, acknowledging that participants’ accounts reflect both individual experiences and the social and implementation contexts in which the peer mentor program was delivered. Reflexivity was supported through analytic discussions among the research team, during which assumptions, positionalities, and emerging interpretations were examined and refined.
Within this broader mixed methods framework, a convergent mixed methods design was used, with quantitative and qualitative data analyzed separately and then integrated during interpretation. The findings from both data strands were compared to identify convergence, complementarity, and divergence across feasibility, acceptability, and fidelity outcomes.
In South Africa, 19 individuals who met the peer mentor person specifications participated in training. Of these, 13 (68%) peer mentor trainees achieved a competency score ≥50% (median 70.7%, IQR 45.8%‐78.2%). We recruited them to participate as peer mentors (of these trainees, 2 with borderline initial scores passed on their second attempts). During the DoBAt pilot randomized controlled trial (approximately 12 mo), 4 of 13 (31%) peer mentors withdrew from the program, resulting in a retention rate of 69%. Baseline characteristics of peer mentors who withdrew (n=4) and those who completed (n=9) are presented descriptively (see ). Given the small sample size and number of withdrawals, formal statistical comparisons were not conducted, as such analyses would be underpowered and potentially misleading. Those who withdrew were slightly older on average (mean 24.8, SD 3.0 y vs mean 22.7, SD 1.3 y), while distributions of gender and qualification background appeared similar across groups. These descriptive observations should be interpreted cautiously, as the small sample sizes preclude meaningful inference regarding baseline differences. Reported reasons for withdrawal included new employment or study opportunities, dissatisfaction with compensation, and declining to address performance concerns raised by supervisors.
In Uganda, 4 individuals who met the role specifications underwent training. All 4 (100%) trainees achieved a competency score ≥50% (median 75.4%, IQR 73.9%‐77.3%) and were recruited as peer mentors. During the Ebikolwa n’Empisa feasibility study (approximately 3 months), no peer mentors withdrew (100% retention).
Attendance RatesTraining attendance was high across both sites. In South Africa, trainees attended 6.3 (89%) out of 7 of the initial training sessions, and those selected attended 9.2 (92%) out of 10 of the booster training sessions. In Uganda, all 4 trainees attended 100% of the sessions.
Attendance at weekly supervision sessions varied. In South Africa, median attendance among mentors was 23 (IQR 21.0‐36.5) out of 40 sessions (57.5%) over 12 months. In Uganda, supervisors reported near-full attendance over the 3 months, despite no formal attendance logs being kept.
Peer Mentor Training EvaluationThe results indicated high overall satisfaction in both South Africa and Uganda (). Peer mentors reported that training enhanced learning, boosted confidence in mentoring skills, and improved the understanding of BA. Facilitators were commended for clear communication and for fostering a supportive learning environment. Role-plays, instructional videos, and learner workbooks were rated as particularly useful for bridging theory and practice. While most mentors felt well prepared, a few expressed a desire for additional support to better prepare them for their roles.
Figure 1. Peer mentor training evaluation scores in South Africa and Uganda. One area of divergence was the feasibility of online learning, which received lower ratings than other domains. In both settings, trainees reported occasional disruptions due to power outages and network interruptions, as well as intermittent disengagement. Suggestions for improvement included scheduled breaks and adjustments to training hours.
Qualitative findings support the high acceptability of the training.
When we did the training, I felt like, “Yes, I’m on top of this thing.Taken together, these findings suggest strong overall acceptability of the training model, while also identifying opportunities for refinement, particularly in settings relying on remote delivery.
FidelityPosttraining CompetenceWe assessed trainees’ competence and engagement using a structured posttraining assessment; median competence was high at both sites ().
Table 3. Posttraining competence assessment scores.Competency domainSouth Africa (n=19), median (IQR)Uganda (n=4), median (IQR)P valueNonspecific competencies74.3 (50.0‐83.6)75.7 (73.6‐78.6).83Program-specific competencies75.0 (56.3‐81.7)78.3 (72.5‐80.8).63Posttraining total for group70.7 (45.8‐78.2)75.4 (73.8‐77.3).37Overall reliability (Krippendorff α)0.67——aMann-Whitney U test (exact P value).
bNot applicable.
Intervention Period FidelityA total of 89 sessions were rated (South Africa: 40 sessions across 10 providers; Uganda: 49 sessions across 4 providers). provides adjusted marginal means for competence, adherence, and overall fidelity. Competence was 83.8% (95% CI 79.7%‐87.8%) in South Africa and 88.6% (95% CI 84.6‐92.6) in Uganda. Competence was higher in Uganda, with an adjusted mean difference of 4.88 percentage points (95% CI 1.23‐8.53; P=.009). Adherence was 79.7% (95% CI 75.6%‐83.7%) in South Africa and 93.0% (95% CI 89.0%‐96.9%) in Uganda, corresponding to an adjusted mean difference of 13.30 percentage points (95% CI 8.99‐17.61; P<.001). Overall fidelity was 84.7% (95% CI 80.3%‐89.0%) in South Africa and 87.7% (95% CI 83.4%‐92.1%) in Uganda; the adjusted mean difference of 3.06 percentage points was not statistically significant (95% CI −0.98 to 7.10; P=.14). Likelihood ratio tests supported the inclusion of provider-level random effects for adherence (P=.009), competence (P<.001), and fidelity (P<.001), indicating meaningful between-provider variability.
Table 4. Fidelity of peer mentor calls in South Africa and Uganda.OutcomeSouth Africa (n=40 calls)Uganda (n=49 calls)P valueCompetence83.8 (79.7‐87.8)88.6 (84.6‐92.6).009Adherence79.7 (75.6‐83.7)93.0 (89.0‐96.9)<.001Fidelity score84.7 (80.3‐89.0)87.7 (83.4‐92.1).14aAdjusted marginal means with 95% CI.
Across outcomes, scores were generally high, with consistently higher adherence and competence in the Ugandan site. Because only 1 implementation site was included per country, these comparisons reflect site-level differences and should not be interpreted as country-level effects. Fidelity did not vary significantly across the 6 treatment calls (F6,82=1.31; P=.26), indicating that fidelity did not differ systematically across treatment phases. We also examined fidelity over calendar time (week of delivery) and found no significant trend (F17,69=0.76; P=.73), indicating that peer mentor performance remained stable throughout the study.
Qualitative perspectives on supervision highlighted barriers and facilitators to program delivery, underscoring opportunities to enhance peer mentor engagement, competence, and adherence.
Peer mentors reported that feedback on call recordings helped them develop their skills.
That actually was very important because you could receive feedback and actually see how you’re doing and where to improve.Group supervision was considered valuable because it enabled peer mentors to learn from one another.
If I come across a challenge, then I can say, “My mentee was difficult, and this is how I worked around it.” The next person has now taken that skill to say, “I can look out for certain things, and when that time comes, I now have something that I can use to mitigate that problem.”For peer mentors in South Africa employed across the DoBAt pilot trial (approximately 12 months), barriers to engagement included competing work and studies.
It was just a bit difficult for me to juggle around (my work) and be able to give all my mentees my full attention at the same time.While many South African peer mentors found the online nature of the work convenient, some felt that periodic in-person contact with supervisors and fellow mentors would have supported engagement and retention.
I think that can be a barrier in itself, not having the in-person contact.This study evaluated the feasibility, acceptability, and fidelity of an online training and supervision program for peer mentors supporting the delivery of a BA digital intervention. This intervention (Kuamsha app) targeted adolescents with depression across 2 studies conducted in South Africa and Uganda. The findings suggest that the peer mentor program offered a feasible and acceptable approach in both settings and achieved high overall fidelity. Differences in retention, adherence, and fidelity across sites underscore the importance of implementation factors, such as supervision models.
In South Africa, we trained 19 peer mentors and recruited 13 for the study, with a 12-month retention rate of 9 of 13 (69%). In Uganda, all 4 trained mentors were retained in the 3-month program. Training attendance was consistently high in both countries (approximately 90% in South Africa; 100% in Uganda), but supervision attendance was more variable in South Africa (64%); in Uganda, supervisors reported few absences, although formal logs were not kept. In South Africa, weekly online group supervision followed a structured format, whereas in Uganda, supervision was more informal.
Satisfaction ratings and qualitative findings support the high overall acceptability of training and supervision. Peer mentors reported positive learning experiences, greater confidence in their acquired mentorship skills, and appreciation for multimodal and interactive training methods (eg, role-plays, guided workbooks, supervised practice). However, technical challenges, such as network disruptions and power outages, hindered participation in online training. For some peer mentors in South Africa, the part-time nature of the work and competing priorities reduced engagement in the program. Addressing these obstacles may improve retention and adherence. Examples of possible solutions include flexible online training schedules, opportunities for periodic in-person contact, and, where feasible, full-time roles with higher remuneration.
Competence assessments showed that all peer mentors met the minimum competency threshold (50%). Fidelity assessments indicated moderate-to-high adherence in both countries and remained stable across treatment phases and over time. These findings align with prior research demonstrating that digital interventions are feasible and can provide effective mental health support in LMICs []. Competence and adherence were higher in the Ugandan than the South African site, which is unsurprising given the shorter program duration (3 months as opposed to 12 months) and higher supervisor: peer mentor ratio (1:4, as opposed to 1:13). It is also possible that some in-person contact of peer mentors with their supervisors in Uganda contributed to engagement, retention, and performance. These findings highlight the need for adequate supervisory capacity to maintain fidelity.
Comparison With Prior WorkThe findings of this study contribute to the broader literature on training nonspecialist providers to deliver mental health interventions in LMICs and add emerging evidence on training and supervising peer supporters. A recent systematic umbrella review found that peer-delivered interventions are more likely to succeed when peer supporters receive structured training, ongoing supervision, and clear role guidance. However, most of this evidence comes from high-income settings []. The DoBAt pilot trial findings indicate that peer mentors promoted engagement and adherence to the app, as evidenced by a positive association between call frequency and usage metrics, as well as adolescents’ reports of the value of peer guidance [].
Research on DMHIs suggests that human-supported approaches tend to yield better adherence than digital-only interventions. Our study contributes to this evidence base by demonstrating that structured online supervision and booster training can help maintain fidelity in peer-led interventions in low-resource settings []. The differences observed between South Africa and Uganda mirror previous task-sharing research, which suggests that economic and contextual factors influence volunteer retention and performance. Prior research also highlights the necessity of ongoing supervision and refresher training to maintain fidelity in peer-delivered mental health interventions []. Collectively, these findings support context-specific training and supervision to optimize program outcomes.
Implications for Future Research and PracticeThe results have several implications for future training and supervision models in LMICs. Retention remains a key challenge, as competing work or study priorities, or difficulty achieving competencies, may reduce long-term engagement and performance []. Competence assessment should remain central in peer-delivered interventions. Established tools, such as Enhancing Assessment of Common Therapeutic, Working with Children – Assessment of Competencies Tool, and the broader Ensuring Quality in Psychological Support platform, have been widely used to evaluate the clinical competencies of nonspecialist providers and share commonalities with the structured supervisor rating form used in this study [-]. However, our assessment was intentionally designed to capture BA-specific skills (eg, activity monitoring, structured goal setting) that are not fully encompassed in generalist tools. Future research could examine how BA-specific assessments integrate with or benchmark against existing competency frameworks.
Flexible and hybrid training and supervision models, financial incentives, career development opportunities, and credentialing or certification should be explored to improve retention. Additionally, broadening recruitment to peer mentors who meet the role requirements (eg, age, language, cultural familiarity, confidentiality, and supportive attitude) but are not limited to psychology or social work students or graduates could expand the applicant pool. Recruiting peers from local communities would also facilitate periodic in-person contact among peer mentors, mentees, and supervisors. To further reduce barriers to engagement, future programs should budget for data or mobile credit or negotiate zero-rated access, as digital inequities can disrupt call quality and limit app engagement. This recommendation is consistent with digital health scale-up frameworks that emphasize data affordability as a key barrier in LMICs [,].
Structured, ongoing supervision was critical for maintaining fidelity, particularly among mentors with lower adherence scores. Programs can build on these findings by optimizing remote digital and peer-led group supervision to support cost-effective and sustainable implementation []. Addressing logistical challenges, such as network disruptions and power outages, will further improve delivery. Access to reliable technology and alternative communication strategies may enhance engagement and prevent disruptions.
Finally, future research should evaluate clinical outcomes associated with peer mentorship in adolescent mental health because while peer-delivered and peer-supported approaches are increasingly used, the evidence base remains mixed, particularly in LMIC settings. Recent reviews of peer-led mental health interventions for youth (10‐24 y) found promising effects in several programs, but also highlighted substantial variability in intervention models (eg, structured vs informal support), outcomes, and implementation quality, alongside limited reporting on supervision, fidelity, and mechanisms through which peer support may influence symptom change [,]. Additionally, the outcome findings from the DoBAt pilot trial indicate that peer support may plausibly influence downstream outcomes through engagement pathways (eg, associations between mentor call frequency and intervention use). However, the pilot was not powered to isolate peer effects on depressive symptoms []. Together, these observations motivate fully powered studies that (1) link peer mentor fidelity and contact “dose” to user engagement metrics; (2) test whether engagement mediates symptom change; and (3) examine how modifiable implementation factors, such as supervision intensity, mentor employment conditions, and training modality, affect both fidelity and clinical outcomes.
Strengths and LimitationsOverall, the findings highlight both promise and practical challenges of implementing peer mentor models in LMIC contexts. We have noted several strengths of this study. First, the peer mentor program was implemented and evaluated across 2 distinct low-resource settings, a rural district in South Africa and a peri-urban area in Uganda, enhancing the generalizability and relevance of the findings to diverse contexts. Second, the program was co-designed with mental health professionals and youth with lived experience, ensuring that the training content, delivery format, and support structures were grounded in clinical expertise and the realities of providers (peer mentors) and end users (adolescents). Third, the use of a mixed methods approach provided a comprehensive evaluation with quantitative data providing objective assessments of the program’s feasibility, acceptability, and fidelity, while qualitative insights contextualized peer mentors’ experiences, helping to interpret variability in outcomes. Fourth, the study assessed both the training phase and intervention delivery, offering insights into retention, supervision, and quality over time. Fifth, rigorous fidelity evaluation was conducted through structured tools and independent ratings of a proportion of recorded sessions, with a subset double-coded to ensure interrater reliability. Finally, the program incorporated adaptable training and supervision models that drew on interactive methods (eg, role-plays, guided workbooks) and targeted booster sessions, which are critical for supporting peer-delivered care in real-world settings. This study, therefore, contributes to the literature on task-sharing models and underscores the potential scalability of peer-delivered digital mental health programs [].
Several limitations need to be acknowledged. First, the relatively small sample size, particularly in Uganda, restricts the generalizability of site-specific findings and precludes granular analyses (eg, by gender or education level). Second, self-reported measures, such as the peer mentor training evaluation survey, are susceptible to response bias. Third, variations in the recruitment of peer mentors and the delivery of training and supervision, as well as sample size and study duration, hinder direct comparisons of outcomes between South Africa and Uganda. Finally, we did not measure the direct clinical effect of peer mentorship on adolescent mental health outcomes. Future studies should evaluate the clinical effectiveness of peer mentorship within DMHIs. Prior research has shown that structured training and supervision can enhance the effectiveness of nonspecialist health workers delivering mental health care []. Additionally, digitally supported training and supervision may improve scalability while maintaining fidelity [].
ConclusionsThis study demonstrates that peer support for a digital adolescent depression intervention in South Africa and Uganda is feasible and acceptable among providers and achieves moderate overall fidelity. These findings highlight the potential of DMHIs, supported by trained nonspecialists or peer mentors in LMICs. Further research should optimize training and supervision models and assess the effects of peer support on adolescent mental health outcomes.
We thank the peer mentors who participated in this study, cosupervisor Tinyiko Mafumo (registered counselor), and field supervisor Princess Makhubela for their dedication and hard work. We also thank the adolescents; community stakeholders; and the staff at BRAC, the South African Medical Research Council/Wits University-Agincourt Research Unit, and Sea Monster for their contributions and expertise.
The digital delivery of behavioral activation therapy and Ebikolwa Consortium collaborators are as follows: Alan Stein, Kathleen Kahn, Munshi Sulaiman, Julia R Pozuelo, Bianca D Moffett, Michelle G Craske, Stephen Tollman, Crick Lund, Tholene Sodi, Heather A O’Mahen, Eugene Kinyanda, Kate Orkin, Mahreen Mahmud, Sarah-Jayne Blakemore, Eustasius Musenge, Zamakhanya Makhanya, Joy Onoria, Megan Davis, Christine Nabulumba, Doreen Sikoti, Princess Makhubela, Meriam Maritze, Nokhutula Mayindi, Tlangelani Baloyi, Gabriele Chierchia, Emma J Kilford, Sophie L Fielmann, and Sophie Dallison.
During manuscript preparation, ZM used ChatGPT (OpenAI) and Grammarly for language editing. The authors reviewed and edited all output and take full responsibility for the manuscript’s integrity and accuracy. No generative artificial intelligence tools were used to generate data, results, or images.
Research reported in this publication was supported by the South African Medical Research Council with funds received through the Medical Research Council-Newton UK-South Africa Joint Initiative on Mental Health (grant: MR/S008748/1). The development of the peer supporter training model and fidelity scales was supported by a Medical Research Council Global Challenges Research Fund grant supplement to the University of Exeter. Additional support was provided by a UK Research and Innovation Official Development Assistance Institutional grant supplement to the University of Oxford. The study was conducted within the South African Medical Research Council /Wits Rural Public Health and Health Transitions Research Unit and Agincourt Health and Socio-Demographic Surveillance System, a node of the South African Population Research Infrastructure Network, which is supported by the Department of Science and Innovation; the University of the Witwatersrand; and the Medical Research Council, South Africa, and previously the Wellcome Trust, UK (grants: 058893/Z/99/A, 069683/Z/02/Z, 085477/Z/08/Z, and 085477/B/08/Z).
The deidentified quantitative data and analysis code that support the findings of this study are available from the corresponding author upon reasonable request and subject to institutional data-use agreements. Audio recordings and full interview transcripts are not publicly shared to protect participant privacy; deidentified exemplar quotations are included in the text. Training materials and assessment tools are provided in -.
Methodology: BM, AS, AVH, CL, CT, EK, HOM, JLG-O, JRP, KK, MD, MGC, SG, TB, TS, ZM
Writing – review & editing: AS, AVH, BM, CL, CT, EK, HOM, JRP, JLG-O, KK, MD, MGC, SG, TB, TS, ZM
MGC reports editorial honoraria and/or payments from Oxford University Press, the American Psychological Association, Elsevier (serving as Editor-in-Chief of Behaviour Research and Therapy), and UpToDate, Inc., outside the submitted work. MGC also holds editorial leadership positions that are unrelated to this manuscript. All other authors declare no conflicts of interest.
Edited by John Torous; submitted 24.Oct.2025; peer-reviewed by Maria Loades, Richard Munthali; final revised version received 23.Feb.2026; accepted 24.Feb.2026; published 09.Apr.2026.
© Zamakhanya Makhanya, Bianca Moffett, Julia R Pozuelo, Meghan Davis, Joy Louise Gumikiriza-Onoria, Shayni Geffen, Tlangelani Baloyi, Tholene Sodi, Eugene Kinyanda, Michelle G Craske, Christine Tusiime, Crick Lund, Alastair Van Heerden, Kathleen Kahn, Alan Stein, Heather O'Mahen, DoBAt and Ebikolwa Consortium. Originally published in JMIR Mental Health (https://mental.jmir.org), 9.Apr.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.
Comments (0)