pith. machine review for the scientific record. sign in

arxiv: 2603.29159 · v2 · submitted 2026-03-31 · 💻 cs.CL · cs.CY· cs.HC

Recognition: no theorem link

Kwame 2.0: Human-in-the-Loop Generative AI Teaching Assistant for Large Scale Online Coding Education in Africa

Authors on Pith no claims yet

Pith reviewed 2026-05-14 00:33 UTC · model grok-4.3

classification 💻 cs.CL cs.CYcs.HC
keywords generative AIteaching assistanthuman-in-the-looponline coding educationAfricaretrieval-augmented generationeducational technology
0
0 comments X

The pith

Human-in-the-loop generative AI combines scalability with human reliability for large-scale coding support across Africa.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper presents Kwame 2.0, a bilingual generative AI teaching assistant that retrieves course materials to generate context-aware answers for students in an introductory mobile coding course. Over 15 months it handled more than 3,700 enrollments in 15 cohorts across 35 countries, delivering timely responses that were largely accurate on curriculum questions. Human facilitators and peers reviewed outputs and corrected errors, especially on administrative matters, showing how AI speed can pair with human judgment in resource-limited settings.

Core claim

Kwame 2.0 retrieves relevant course materials and generates responses in English or French while running inside a forum that encourages human oversight and community participation. Evaluation via community feedback and expert ratings in the 15-month study found high accuracy on curriculum-related questions, with human intervention effectively mitigating errors on administrative queries and demonstrating that the combined system offers scalable learning assistance for underrepresented populations in constrained environments.

What carries the argument

Kwame 2.0, a retrieval-augmented generation system placed in a human-in-the-loop forum that incorporates oversight and community input to refine AI outputs.

If this is right

  • Large numbers of learners can receive immediate course-specific help without a proportional increase in full-time human staff.
  • Students in multiple countries and languages gain consistent access to accurate guidance on coding topics.
  • AI generation errors are caught through human review, particularly for queries outside the core curriculum.
  • The model supports education delivery in resource-constrained regions by leveraging both automated speed and targeted human correction.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same human-in-the-loop structure could extend to other subjects such as mathematics or basic science in similar low-resource settings.
  • Over repeated deployments the amount of required human review might decrease if the retrieval and generation components improve from accumulated corrections.
  • Educational providers could test cost reductions by substituting some traditional tutoring hours with this AI-plus-oversight approach.

Load-bearing premise

Community feedback and expert ratings provide unbiased and comprehensive evidence of support quality, and human oversight remains scalable without becoming a bottleneck.

What would settle it

A larger deployment in which the volume of AI responses overwhelms human reviewers, resulting in a measurable rise in uncorrected errors and lower expert ratings on support quality.

Figures

Figures reproduced from arXiv: 2603.29159 by George Boateng, Samuel Boateng, Victor Kumbol.

Figure 1
Figure 1. Figure 1: Architecture of Kwame 2.0 computes an embedding of the question using Sentence-BERT, and then computes cosine similarity scores with all saved embeddings in ElasticSearch to retrieve the top 5 relevant passages from our course materials based on students’ questions, any attached images, and code snippets from relevant course material using question tags, and then uses them as context to the GPT-4 API [1] t… view at source ↗
Figure 2
Figure 2. Figure 2: Prompt for Kwame 2.0 along with learners who could respond to queries, upvote, or downvote answers. Furthermore, a user who asked a question could accept only one answer as the correct one. This human-in-the-loop model ensured that Kwame 2.0’s responses were adequately supported and supervised. 5 Evaluation We evaluated the helpfulness of Kwame 2.0’s answers using community ratings (i.e., the number of upv… view at source ↗
Figure 3
Figure 3. Figure 3: Screenshots of Kwame 2.0 in the SuaCode course forum were provided almost instantaneously, they had a good chance to address learners’ questions early and then be accepted as the correct answer and or upvoted. Our expert evaluation showed that there were 490 valid questions, out of which 50.8% were curriculum questions and 49.2% were administrative questions. We computed accuracy as the number of correct a… view at source ↗
read the original abstract

Providing timely and accurate learning support in large-scale online coding courses is challenging, particularly in resource-constrained contexts. We present Kwame 2.0, a bilingual (English-French) generative AI teaching assistant built using retrieval-augmented generation and deployed in a human-in-the-loop forum within SuaCode, an introductory mobile-based coding course for learners across Africa. Kwame 2.0 retrieves relevant course materials and generates context-aware responses while encouraging human oversight and community participation. We deployed the system in a 15-month longitudinal study spanning 15 cohorts with 3,717 enrollments across 35 African countries. Evaluation using community feedback and expert ratings shows that Kwame 2.0 provided high-quality and timely support, achieving high accuracy on curriculum-related questions, while human facilitators and peers effectively mitigated errors, particularly for administrative queries. Our findings demonstrate that human-in-the-loop generative AI systems can combine the scalability and speed of AI with the reliability of human support, offering an effective approach to learning assistance for underrepresented populations in resource-constrained settings at scale.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript presents Kwame 2.0, a bilingual (English-French) generative AI teaching assistant built with retrieval-augmented generation (RAG) and deployed in a human-in-the-loop forum for the SuaCode introductory mobile coding course. It describes a 15-month longitudinal deployment across 15 cohorts with 3,717 enrollments in 35 African countries. Evaluation relies on community feedback and expert ratings, with the central claim that the system achieves high accuracy on curriculum questions while human facilitators and peers effectively mitigate errors, particularly for administrative queries, thereby combining AI scalability with human reliability at scale for resource-constrained settings.

Significance. If the effectiveness claims are substantiated, the work could offer a practical template for hybrid AI-human support systems in large-scale online education targeting underrepresented learners in Africa and similar contexts. The deployment scale (3,717 enrollments) and bilingual focus are notable strengths. However, the absence of objective metrics, baseline comparisons, and quantified human intervention rates substantially weakens the ability to assess whether the approach genuinely scales without human bottlenecks or selection bias in feedback.

major comments (3)
  1. [Abstract] Abstract: The claim that Kwame 2.0 achieved 'high accuracy on curriculum-related questions' provides no supporting details on measurement (e.g., expert-written reference answers, multiple-choice probes, error rates, or inter-rater reliability for expert ratings), which is load-bearing for the central effectiveness assertion.
  2. [Evaluation] Evaluation/Deployment description: No counts or fractions are reported for queries handled fully automatically versus those requiring human intervention across the 3,717 enrollments, preventing assessment of whether human oversight remains scalable or introduces bottlenecks as claimed.
  3. [Longitudinal study] Longitudinal study section: The 15-cohort results lack baseline comparisons to non-AI support methods and details on how post-hoc adjustments or potential biases in community feedback were handled, undermining the cross-cohort reliability claims.
minor comments (2)
  1. [Abstract] Abstract: Consider specifying response latency metrics or exact accuracy percentages if quantified elsewhere in the manuscript to strengthen the timeliness claim.
  2. [Methods] Notation and terminology: The term 'human-in-the-loop' is used without a precise definition of the escalation protocol or decision criteria for when humans intervene, which could be clarified for reproducibility.

Simulated Author's Rebuttal

3 responses · 1 unresolved

We thank the referee for their constructive feedback on our manuscript. We address each major comment below, proposing revisions where feasible while noting limitations inherent to the study design.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claim that Kwame 2.0 achieved 'high accuracy on curriculum-related questions' provides no supporting details on measurement (e.g., expert-written reference answers, multiple-choice probes, error rates, or inter-rater reliability for expert ratings), which is load-bearing for the central effectiveness assertion.

    Authors: We agree that the abstract would benefit from greater specificity on the accuracy measurement. In the revised manuscript, we will update the abstract to briefly describe the expert rating process (a sample of curriculum responses rated by domain experts against reference materials from the course, with details on the rating scale and agreement metrics provided in the Evaluation section). revision: yes

  2. Referee: [Evaluation] Evaluation/Deployment description: No counts or fractions are reported for queries handled fully automatically versus those requiring human intervention across the 3,717 enrollments, preventing assessment of whether human oversight remains scalable or introduces bottlenecks as claimed.

    Authors: We acknowledge this omission and will add the requested statistics to the Evaluation/Deployment section. Our system logs tracked intervention events, allowing us to report the fraction of queries resolved automatically by the RAG model versus those escalated to human facilitators or peers. revision: yes

  3. Referee: [Longitudinal study] Longitudinal study section: The 15-cohort results lack baseline comparisons to non-AI support methods and details on how post-hoc adjustments or potential biases in community feedback were handled, undermining the cross-cohort reliability claims.

    Authors: We will expand the Limitations section to discuss potential selection bias in voluntary community feedback and note that no post-hoc adjustments were applied to the data; instead, findings were triangulated with expert ratings for reliability. However, the observational deployment design precludes direct baseline comparisons. revision: partial

standing simulated objections not resolved
  • Direct baseline comparisons to non-AI support methods, as the study was a single-arm longitudinal deployment without a control condition.

Circularity Check

0 steps flagged

Empirical deployment study with no derivation chain or self-referential claims

full rationale

The paper presents a system description and longitudinal deployment results evaluated via community feedback and expert ratings across 3,717 enrollments. No equations, fitted parameters, predictions derived from inputs, uniqueness theorems, or ansatzes appear in the abstract or described structure. The central claim rests on observed outcomes from external user interactions rather than any reduction to self-defined quantities or self-citations. This is a standard non-circular empirical report.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claim depends on untested assumptions about retrieval accuracy for curriculum content and the sustained effectiveness of human oversight at scale; no free parameters or invented entities are introduced.

axioms (2)
  • domain assumption RAG retrieval produces relevant and accurate context for curriculum-related student questions.
    Invoked to support the reported high accuracy on course content.
  • domain assumption Human facilitators and peers can reliably detect and correct AI errors without introducing new delays or biases.
    Central to the human-in-the-loop reliability claim.

pith-pipeline@v0.9.0 · 5500 in / 1256 out tokens · 45043 ms · 2026-05-14T00:33:19.248666+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

15 extracted references · 15 canonical work pages · 2 internal anchors

  1. [1]

    GPT-4 Technical Report

    Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F.L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al.: Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)

  2. [2]

    In: Proceedings of the 10th Computer Science Education Research Conference

    Annor, P.S., Kayang, E., Boateng, S., Boateng, G.: Autograd: Automated grading software for mobile game assignments in suacode courses. In: Proceedings of the 10th Computer Science Education Research Conference. pp. 79–85 (2021)

  3. [3]

    In: International Conference on Artificial Intelligence in Education

    Boateng, G.: Kwame: A bilingual ai teaching assistant for online suacode courses. In: International Conference on Artificial Intelligence in Education. pp. 93–97. Springer (2021) 8 G. Boateng et al

  4. [4]

    In: Proceedings of the 10th Computer Science Education Research Conference

    Boateng, G., Annor, P.S., Kumbol, V.W.A.: Suacode africa: Teaching coding online to africans using smartphones. In: Proceedings of the 10th Computer Science Education Research Conference. pp. 14–20 (2021)

  5. [5]

    In: 2018 IEEE Integrated STEM Education Conference (ISEC)

    Boateng, G., Kumbol, V.: Project iswest: Promoting a culture of innovation in africa through stem. In: 2018 IEEE Integrated STEM Education Conference (ISEC). pp. 104–111. IEEE (2018)

  6. [6]

    In: Proceedings of the 8th Computer Science Education Research Conference

    Boateng, G., Kumbol, V.W.A., Annor, P.S.: Keep calm and code on your phone: A pilot of suacode, an online smartphone-based coding course. In: Proceedings of the 8th Computer Science Education Research Conference. pp. 9–14 (2019)

  7. [7]

    https://www.dw.com/en/why- mobile-internet-is-so-expensive-in-some-african-nations/a-55483976 (Nov 2020)

    Why mobile internet is so expensive in africa (2020). https://www.dw.com/en/why- mobile-internet-is-so-expensive-in-some-african-nations/a-55483976 (Nov 2020)

  8. [8]

    In: Proceedings of the 2024 on ACM Virtual Global Computing Education Conference V

    Feng, T., Liu, S., Ghosal, D.: Courseassist: Pedagogically appropriate ai tutor for computer science education. In: Proceedings of the 2024 on ACM Virtual Global Computing Education Conference V. 2. pp. 310–311 (2024)

  9. [9]

    Advances in Neural Information Processing Systems 33, 9459–9474 (2020)

    Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., Küttler, H., Lewis, M., Yih, W.t., Rocktäschel, T., et al.: Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33, 9459–9474 (2020)

  10. [10]

    In: Proceedings of the 23rd Koli Calling International Conference on Computing Education Research

    Liffiton, M., Sheese, B.E., Savelka, J., Denny, P.: Codehelp: Using large language models with guardrails for scalable support in programming classes. In: Proceedings of the 23rd Koli Calling International Conference on Computing Education Research. pp. 1–11 (2023)

  11. [11]

    In: Proceedings of the 55th ACM technical symposium on computer science education V

    Liu, R., Zenke, C., Liu, C., Holmes, A., Thornton, P., Malan, D.J.: Teaching cs50 with ai: leveraging generative artificial intelligence in computer science education. In: Proceedings of the 55th ACM technical symposium on computer science education V. 1. pp. 750–756 (2024)

  12. [12]

    In: Proceedings of the eleventh ACM conference on learning@ scale

    Lyu, W., Wang, Y., Chung, T., Sun, Y., Zhang, Y.: Evaluating the effectiveness of llms in introductory computer science education: A semester-long field study. In: Proceedings of the eleventh ACM conference on learning@ scale. pp. 63–74 (2024)

  13. [13]

    In: Proceedings of the 56th ACM Technical Symposium on Computer Science Education V

    Raihan, N., Siddiq, M.L., Santos, J.C., Zampieri, M.: Large language models in computer science education: A systematic literature review. In: Proceedings of the 56th ACM Technical Symposium on Computer Science Education V. 1. pp. 938–944 (2025)

  14. [14]

    Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks

    Reimers, N., Gurevych, I.: Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084 (2019)

  15. [15]

    https://www.c4dhi.org/news/lecture-by-boateng-suacode-africa- 20210122/

    Suacode africa 2.0: Teaching coding online to africans using smartphones during covid-19. "https://www.c4dhi.org/news/lecture-by-boateng-suacode-africa- 20210122/" (Jan 2021)