Generative artificial intelligence (genAI, or here simply AI) has arrived in the educational discourse – often with grand promises, sometimes with apocalyptic fears. A broad, albeit heterogeneous, empirical base now exists. Meta-analyses report significantly positive effects of AI chatbots like ChatGPT on learning performance, motivation, and higher thinking, but under very different conditions, with sometimes questionable study designs, and predominantly through studies with older pupils or university students (Alemdag, 2025; Deng et al., 2025; Wang & Fan, 2025; Wu & Yu, 2024; Zheng et al., 2023). The question, therefore, is not whether, but rather how these positive effects can be achieved. Because, of course, ChatGPT and similar technologies are not a guaranteed success in the field of learning support, but rather a tool that can be used didactically.
Therefore, authors warn against misunderstanding the sometimes observed "learning effect with AI" as a property of the AI technology itself, instead of asking about the underlying learning environment and didactic pre-structuring (Dinsmore & Fryer, 2026; Kirschner, 2025). In short: AI can effectively support learning – but only if it is designed and integrated in a didactically sound manner. However, it can be just as pointless if it replaces and prevents learning instead of supporting it.
The question I want to explore here is: How can AI be used to support learning processes through personalized feedback? In the following, I outline fundamental, empirically and theoretically supported considerations regarding design principles for AI learning companions and AI tutors in mathematics learning, with a particular focus on primary education. These considerations form the basis for the development and testing of app-integrated, AI-supported learning guidance within the project. PRIMA-KI.
Adaptive scaffolding: Help for self-help instead of quick fixes
AI-supported learning environments generate learning gains particularly when they adapt to learning progress, errors, and strategies, rather than simply providing correct answers. In the TALPer study, for example, lower-achieving fifth-grade students benefited especially from adaptive support in mathematics lessons, while higher-achieving students developed more complex interaction patterns with the AI learning companion (Kuo et al., 2025). Similarly, a comparable learning system showed significant performance gains in word problems, with the perceived benefit—that is, the quality of the support—having a strong influence on motivation and learning outcomes (Liu et al., 2025). Son's (2024) systematic review confirms the positive effects of well-designed intelligent tutoring systems on mathematical learning performance, especially when these systems adapt to individual learning needs.
In the field of self-regulated learning, several studies indicate that adaptive, AI-supported scaffolding (instead of static support sequences) can improve the quality of self-regulated learning processes and offers advantages over a concept of "equal support for all students" (Liu et al., 2025; Wu et al., 2025). New multi-agent approaches and LLM-as-Judge procedures show that the quality and reliability of scaffolds can be improved through self-assessment procedures, and that hallucinations during feedback can be significantly reduced (Cohn et al., 2025; Gonnermann-Müller et al., 2025; Qian et al., 2026).
At the same time, limitations of some tools not specifically designed to support learning became apparent: AI systems that provide unregulated access to complete solutions on demand during practice may increase performance in the short term, but impair long-term learning when the AI is withdrawn (Bastani et al., 2025). This is also referred to as "de-skilling," as it hinders rather than promotes actual learning and only superficially creates the illusion of learning. Learning analyses from AI-based tutoring systems also show that some learners try to "click through" tasks without truly engaging with them (Jancafik et al., 2023). This should be prevented as much as possible through appropriate pre-structuring and design of the AI feedback in order to prevent superficial learning and enable deeper, more comprehension-enhancing learning.
Conclusions: An AI learning companion should primarily:
- Diagnose errors and misconceptions based on subject-specific didactic background information on the topic and adapt accordingly. Notes give (Bewersdorff et al., 2023),
- Encourage a change in strategy and provide informative guidance for independent problem-solving, instead of delivering finished results.,
- Provide support in measured doses, stagger it over time (fading), and, if necessary, "withdraw" it if students become too dependent (i.e., request too many hints).
- Implement mechanisms against superficial problem-solving by strengthening children's ability to solve problems independently through questioning and referencing previous work.
- Individualize: Adapt task formats to the competence level to avoid over- or under-challenging students (Son, 2024).
Good AI tutors encourage independent thinking.
Research on Large Language Models (LLMs) in learning suggests that the sequence of thinking and explaining is crucial. AI feedback is particularly effective when learners first attempt to solve problems themselves (Kumar et al., 2023). Even rare, erroneous LLM explanations in such settings still led to learning gains, without participants systematically adopting incorrect strategies (Kumar et al., 2023). Systems that intervene too early ("reactive feedback") risk suppressing metacognitive processes (Chudziak & Kostka, 2025). Furthermore, hybrid intelligence models—that is, the combination of human thinking and AI support—often lead to better results and lower cognitive load than pure AI solutions (Cosentino et al., 2025).
This corresponds to the learning psychology critique that genAI often tempts users to take a "shortcut" in their thinking: While directly consuming the solution saves effort, it hardly builds robust knowledge structures (Gisiger, 2025; Dinsmore & Fryer, 2026). Studies on AI in practice also show that if AI offers complete solutions during practice, long-term learning outcomes are worse than without AI—unless the AI is regulated to provide only step-by-step hints (Bastani et al., 2025). In their theoretical framework for LLM-based educational agents, Cohn et al. (2025) explicitly emphasize the necessity of "guided discovery over direct answers"—that is, guided discovery instead of direct answers.
Analysis of real-life tutoring dialogues also shows that pure instruction without active learner participation generates little constructive learning. Primary school students respond positively to targeted, interactive questions, but remain passive when the tutor remains monologic and solution-focused (Wang et al., 2025). Ruan et al. (2020) demonstrate a similar finding in their study of narrative-based chatbot tutors: learning gains were achieved primarily when the system provided interactive feedback and hints rather than direct solutions.
Conclusions: An AI learning companion in mathematics lessons should consistently work according to the principle: "First you – then me."„
That means:
- First, a own attempt at a solution demanded.
- The AI only comes into play when needed, either on request or automatically in case of errors, and asks for ideas, sub-strategies, and observations.
- Explanations will be "docked" to existing foundations of understanding„.
- Complete model solutions should remain the exception and be used as a reflection tool, not as the primary learning format.
- Learners are encouraged to critically examine AI responses – a skill that children can practice as early as primary school, for example by detecting errors in AI statements (Helal et al., 2024).
Without a sound pedagogical foundation, the AI tutor becomes a "linguistically gifted random number generator".„
While generative AI can produce linguistically and formally impressive results, it often generates superficial, poorly problem-oriented tasks from a subject-didactic perspective if it is not specifically controlled and the output is not pre-framed according to subject-didactic principles (Schneider, 2025). Literature reviews on AI in mathematics education emphasize the potential for visualization, individualization, and problem-solving, but also point to the weak or overly unspecific theoretical foundation of many systems, presumably because the systems are not specifically optimized for subject-didactic issues (Almheiri et al., 2025; Awang et al., 2024; Holmes & Tuomi, 2022; Pesemowo & Adewuyi, 2024). The systematic analysis by Cárdenas et al. (2025) identifies a lack of theoretical framework as one of the key obstacles to effective AI tutoring systems. Successful systems like ChatTutor or specific frameworks are therefore explicitly based on educational theories such as Social Cognitive Theory or Evidence-Centered Design (Cohn et al., 2025; Dwivedi & Rejina, 2025). It is not enough to view AI as a mere technical "add-on".
Theory-driven tutoring systems are therefore significant: Building on Generative Learning Theory, the study by Makransky et al. (2024) shows that a model specifically trained in didactics promotes conceptual knowledge, confidence, and enjoyment significantly more than a generic LLM or traditional instruction—and that these effects remain stable in the follow-up. The same applies to models specifically adapted for learning tutoring, such as Google's LearnLM. Studies on GeoGebra and AI-supported environments clearly show that conceptual understanding and self-efficacy only increase when subject matter and subject-specific didactics are explicitly incorporated into the system design, rather than simply adding "technology on top" (Cononigo, 2024; Kortenkamp, 2024).
Cohn et al. (2025) made an important contribution with their framework for adaptive scaffolds in LLM-based educational agents, which combines evidence-centered design with social cognitive theory. This theoretically grounded approach enables high-quality formative assessment and interaction.
Conclusion: An AI tutor requires extensive content-specific knowledge as context in the prompt or special follow-up training with subject-specific didactic knowledge, for example:
- a mathematical model Competence development for the content area,
- Error and strategy taxonomies, which depict typical student pathways and errors (Nauryzbayev et al., 2023),
- explicit rules and guidelines, Which forms of support are useful in which learning phase? are,
- A clear understanding and recognition of misconceptions is crucial so that the AI can tailor its approach to promote routine skills (e.g., computational fluency) or conceptual understanding. This can also mean, for example, embedding a learning history so that the generative AI can respond appropriately and avoid endless loops.
Metacognition, self-regulated learning and "teachable agents": The tutor as a mirror of one's own thinking
Chatbots and AI systems can be used not only as guides and task explainers. They can also stimulate processes of planning, monitoring, and reflection in problem-solving—if they are designed to do so. The meta-analysis by Wu et al. (2025) shows that chatbots can support self-regulated learning technically, socially, and reflectively when their scaffolds are linked to models of self-regulated learning. Guo et al. (2025) confirm in their systematic review that AI systems can fulfill basic psychological needs for autonomy, competence, and relatedness—key factors for motivation and engagement.
Studies on so-called teachable agents demonstrate that learners can perceive AI-supported systems as learning companions, moderators, and collaborative problem solvers when they address their own explanations to these agents (Song et al., 2024). At the same time, it is evident that many current generative AI systems do not yet reliably fulfill key tutorial roles—such as the targeted stimulation of planning, strategy selection, and reflection—without specific prompts or retraining, and tend toward a prescriptive style (Chudziak & Kostka, 2025; Contel & Cusi, 2025). This underscores the need for proactive metacognitive scaffolds that go beyond reactive feedback.
Conclusions: AI learning companions can be explicitly designed as metacognitive learning partners::
- The tutor asks questions such as "What did you notice?", "What strategy did you try?", "Why do you think that works?".
- It encourages reflection on mistakes ("What idea from before could help here?").
- Teach-back-Elements can be included where appropriate: children explain to the AI what they have understood, and the AI reflects back, and so on.
„Human in the loop“: Hybrid approaches instead of fully automated solutions
A common shortcoming of current AI tools is the lack of teacher involvement (Guerino et al., 2023). AI must not become a "black box" in the classroom. The most effective scenarios are those in which AI relieves the teacher of some tasks (e.g., through automated differentiation or real-time analysis), allowing them more time for pedagogical interaction (Wezendonk & Veldhuis, 2024; Gonnermann-Müller et al., 2025). Reviews of AI in (mathematical) education emphasize that AI systems should be understood as feedback loops in which data collection, pattern recognition, and adaptive support must always be framed by human decisions (Holmes & Tuomi, 2022; Ninaus & Sailer, 2022). AI can support learning processes, but it cannot replace pedagogical professionalism or the responsibility of teachers (Aru & Laak, 2025; Aufenanger, 2023; Buchholtz et al., 2024).
The meta-analysis by Kaliisa et al. (2025) shows that AI feedback is no less effective than human feedback, but also not systematically superior. In the studies examined—mostly from the field of linguistics—hybrid approaches are particularly promising, combining reliable, direct, and accessible AI feedback with framed feedback from human teachers. Cosentino et al. (2025), in their study on generative AI in embodied learning environments, demonstrate that hybrid feedback models have the potential to reduce cognitive load and support differentiated information processing strategies.
Further studies on hybrid systems confirm this: CoPilot, for example, supports human teachers in real time and leads to significantly better mathematical learning outcomes for younger students – especially where teachers would otherwise provide weaker feedback, for example, due to teaching outside their subject area (Wang et al., 2024). Kestin et al. (2025) note that AI tutors, based on the didactic principles of active learning, can enrich very good face-to-face instruction in certain phases without replacing it. The effective integration in primary schools depends heavily on teachers' attitudes, the available technology, and the involvement of parents and the school community (KMK, 2024; Li, 2024).
However, the involvement of teachers in the design and use of AI tutoring systems often remains insufficient. Guerino et al. (2023) and Wezendonk and Veldhuis (2024) emphasize that teacher-centered design approaches and corresponding AI literacy programs are necessary to ensure practical classroom integration and acceptance, and to avoid using AI as a mere add-on.
Conclusions: AI learning companions should be designed from the outset as Tools and support for teachers to be thought of:
- The generation of feedback must be based on subject-specific didactic information and not left to chance.
- Transparency: Teachers need dashboards to see where students stand and what help the AI has provided.
- AI can suggest tasks, provide support, or offer diagnoses – but the final decision remains with the teacher. Teachers evaluate the quality of AI-supported assistance and decide whether and how AI-based learning support is integrated and in which areas support is beneficial. The "micro-didactic" area of providing support at the task level during problem-solving is particularly promising, as teachers are often unable to offer direct support at crucial points, or can only do so inadequately.
- Professional development and training in the integration and use of AI (AI literacy) are prerequisites for responsible use (Holmes et al., 2018; KMK, 2024; Wang & Nie, 2023) and should therefore be integrated into teacher training and professional development. This should not only extend to knowledge about the technology, but also include pedagogical aspects that address both opportunities and risks.
Motivation and emotion: The tutor as an empathetic partner
Recent reviews emphasize the role of the affective dimension. AI systems that utilize social robots, storytelling, or gamification can fulfill the basic psychological needs for autonomy, competence, and relatedness (Guo et al., 2025; Vitale & Iacono, 2024). Narratives and playful elements, in particular, help maintain motivation even during tedious practice sessions („Enhancing student engagement…“, 2022). Help-seeking research has shown that learners avoid seeking help when it is perceived as public, status-threatening, or embarrassing. Shame, fear of evaluation, and „loss of face“ thus inhibit children from asking questions in class. AI systems can effectively complement teacher support in this regard, as they patiently and personally provide help and feedback without risking the perceived social costs and stigma of seeking help (Aleven et al., 2016).
ConclusionsAI learning support should be a metacognitive and motivational partner:
- Building relationships: Use of friendly, personalized avatars or narratives (Ruan et al., 2020).
- Growth-oriented feedback: Providing feedback that praises effort and framing mistakes as learning opportunities (growth mindset).
- Stimulate reflectionQuestions like "How did you come up with that?" encourage reflection on one's own learning (metacognition).
Ethical aspects: What AI must not do!
Ethical questions are fundamental to learning with genAI – especially when it involves children. Holmes et al. (2021) call for a collaboratively developed ethical framework for AI in education that, beyond data protection, encompasses aspects such as fairness, transparency, agency, and pedagogical responsibility. The Standing Conference of the Ministers of Education and Cultural Affairs (KMK, 2024) explicitly recommends a cautious, research-based use of AI in primary and special education schools, focusing on basic skills, inclusion, equal opportunities, and data protection-compliant, age-appropriate solutions. Scoping reviews on AI and human flourishing show that previous research is strongly performance-oriented and focused on learning outcomes, while ethical, metacognitive, and teacher-related perspectives remain under-researched (Fock & Siller, 2025).
Almheiri et al. (2025) and Cárdenas et al. (2025) identify ethical challenges and scaling problems as key obstacles to the widespread use of AI tutoring systems in their systematic reviews. Furthermore, studies indicate that frequent AI use can be associated with a decline in critical thinking skills, especially when AI is primarily used as an answer-providing machine (Gerlich, 2025). Research on psychological profiling with LLMs demonstrates how accurately models can derive personality and value patterns from texts—thus highlighting the potential for misuse inherent in opaque systems (Rosenfelder et al., 2025). Therefore, the pedagogical pre-structuring and regulation of AI tools is of paramount importance, particularly for younger children.
Gulz et al. (2021) also emphasize the need to combine adaptivity with inclusive pedagogy without stigmatizing learners with special needs – an important aspect that is often neglected.
Conclusions: A responsible AI learning facilitator in mathematics lessons must:
- data-saving to avoid psychological profiling, for example by consistently processing data preferably on the device or in data-secured infrastructures, separating AI evaluations from personal data, which should be limited to the bare minimum.
- be barrier-free: Utilize multimodal interaction (language, text, image) for diverse learning needs (Hocine et al., 2023) and promote equal opportunities by specifically considering disadvantaged learners in the design process.
- to explain its functionality, at least in its basic outlines,
- Students to critical review Stimulate AI responses (Kortenkamp, 2024),
- not control learners, but Strengthening self-reliance.
- be pedagogically pre-structured and controlled.
For the PRIMA-AI project, the SKILL framework model for the use of generative AI in younger children was developed, which links the openness and pre-structured nature of the systems used to the competencies in dealing with AI and can therefore provide important guidance for the design of the use of AI.
AI as a catalyst for our own mathematical discoveries – not as a replacement
AI should not and must not lead to children merely staring at screens. Research on tangible interfaces and social robots shows that AI can also stimulate interaction in the physical world (Ligthart et al., 2023). AI should serve as a catalyst for mathematical activities—for example, to discover patterns that can then be created using tactile materials. There are currently significantly fewer studies for primary education than for secondary and higher education, but the existing results offer cautious optimism: children can benefit from genAI in their learning if learning environments are designed appropriately (Hwang, 2022; Listyaningrum et al., 2024; Mott et al., 2023; Rumbelow & Coles, 2024; Yim & Su, 2025). Some examples:
- Primary school children can understand key AI concepts such as classification and bias when these are addressed in project-based, "unplugged," or playful settings (Yim & Su, 2025; Mott et al., 2023). Such an understanding is important so that children can critically assess how to interact with the scaffolds of AI learning companions—and also that not all assistance is necessarily correct.
- AI-supported object recognition (e.g., for Cuisenaire rods) can help children to increasingly link their actions with abstract mathematical representations (Rumbelow & Coles, 2024).
- AI training can specifically improve arithmetic fluency, but for children with difficulties in arithmetic, it must be carefully combined with other forms of practice (Samuelsson, 2023). Samuelsson shows that AI-supported practice achieves greater fluency gains than memorization-based approaches alone.
- ChatGPT and similar systems show potential for supporting problem-solving, geometric tasks and algebra in primary school, but require high prompt quality and subject-specific filtering by teachers (Harahap, 2024; Listyaningrum et al., 2024; Schneider, 2025).
- Adaptive systems for children with dyscalculia show promising results in maintaining motivation and engagement (Hocine et al., 2023; Holmes, 2024).
Narrative and gamified approaches deserve special attention: Ruan et al. (2020) show that storytelling-based chatbot tutors can promote engagement and learning gains. Sayed et al. (2022) confirm significant improvements, particularly among lower-achieving students, through adaptive, gamified content.
Conclusions: In primary school, AI learning companions should primarily serve as Stimulus for rich mathematical activities They should not serve as a tool to "quickly arrive at a solution." The central task remains to enable children to critically examine, justify, and negotiate mathematical statements—including those from AI (Kortenkamp, 2024; Aufenanger, 2023).
Conclusion: Designing AI-based learning support for mathematics lessons from a subject-specific didactic perspective
Based on the current state of research and previous considerations, the following guiding principles for the design of AI tutors can currently be formulated:
- Adaptive rather than responsive: Good AI tutors recognize patterns and errors, provide measured assistance, and refrain from giving solutions (Bastani et al., 2025; Bewersdorff et al., 2023; Kuo et al., 2025; Son, 2024).
- Think before explaining: They are designed so that learners must first become active themselves before explanations and feedback follow (Gisiger, 2025; Kumar et al., 2023).
- Subject-specific didactic insights as a basis: They are based on explicit models of mathematical learning – not just on linguistic plausibility, which they inherently possess (Cohn et al., 2025; Cononigo, 2024; Holmes & Tuomi, 2022; Makransky et al., 2024).
- Metacognitive partners: They promote planning, monitoring and reflection instead of mere task completion (Wu et al., 2025; Song et al., 2024).
- Together instead of isolated: They support teachers rather than replace them and work best in human-framed settings during those learning phases where teachers cannot support all children in the learning process (Holmes et al., 2018; Ninaus & Sailer, 2022; Wang et al., 2024; Wezendonk & Veldhuis, 2024).
- Ethically justifiable: They take into account data protection, fairness, accessibility and the risk of cognitive „deterioration“ due to excessive relief (Holmes et al., 2021; Fock & Siller, 2025; Gerlich, 2025; KMK, 2024).
- Pre-structured instead of open: Especially with younger children, an (app- or learning environment-)integrated, pedagogically pre-structured and controlled integration of AI as learning support is necessary, as children are often not yet able to use AI independently as learning support (Gulz et al., 2021).
If we design and further research generative AI learning support based on guiding principles, then it is not a replacement for teaching – but a potential expansion of the didactic repertoire and a tool for more personalized feedback and scaffolding, particularly in primary education. The task for the coming years will be to translate these principles into concrete, empirically tested designs for primary school mathematics instruction. Within the PRIMA-AI project, app-integrated AI learning support tools are currently being developed based on the formulated fundamental principles for AI tutors. These tools are being researched, further developed, and optimized within the framework of design-based research.
literature
Alemdag, E. (2025). The effect of chatbots on learning: A meta-analysis of empirical research. Journal of Research on Technology in Education, 57(2), 459–481. https://doi.org/10.1080/15391523.2023.2255698
Aleven, V., Roll, I., McLaren, BM, & Koedinger, KR (2016). Help Helps, But Only So Much: Research on Help Seeking with Intelligent Tutoring Systems. International Journal of Artificial Intelligence in Education, 26(1), 205-223. https://doi.org/10.1007/s40593-015-0089-1
Almheiri, ASB, Albastaki, H., & Alrashdan, H. (2025). AI-based tutoring systems in education. Advances in Computational Intelligence and Robotics Book Series, 185–210. https://doi.org/10.4018/979-8-3373-0847-0.ch007
Aru, J., & Laak, K.-J. (2025). Developing an AI-based General Personal Tutor for education. Trends in Cognitive Sciences, 29(11), 957–960. https://doi.org/10.1016/j.tics.2025.09.010
Aufenanger, S., Herzig, B., & Schiefner-Rohs, M. (2023). Artificial intelligence and schools. Tasks for teaching and the organization of schools. In C. de Witt, C. Gloerfeld, & S. E. Wrede (Eds.), Artificial intelligence in education (pp. 199–218). Springer Fachmedien. https://doi.org/10.1007/978-3-658-40079-8_10
Awang, L.A., Yusop, FD, & Danaee, M. (2025). Current practices and future directions of artificial intelligence in mathematics education: A systematic review. International Electronic Journal of Mathematics Education, 20(2), em0823. https://doi.org/10.29333/iejme/16006
Bach, KM, Reinhold, F., & Hofer, S. (2025). Unlocking math potential in students from lower SES backgrounds – using instructional scaffolds to improve performance. npj Science of Learning, 10(1).
Bastani, H., Bastani, O., Sungu, A., Ge, H., Kabakcı, Ö., & Mariman, R. (2025). Generative AI without guardrails can harm learning: Evidence from high school mathematics. Proceedings of the National Academy of Sciences, 122(26), 2422633122.
Bewersdorff, A., Seßler, K., Baur, A., Kasneci, E., & Nerdel, C. (2023). Assessing student errors in experimentation using artificial intelligence and large language models: A comparative study with human raters. Computers and Education: Artificial Intelligence, 5, 100177. https://doi.org/10.1016/j.caeai.2023.100177
Buchholtz, N., Schorcht, S., Baumanns, L., Huget, J., Noster, N., Rott, B., Siller, H.-S., & Sommerhoff, D. (2024). Nobody expects this! Six guiding principles on the implications and research needs of AI technologies in mathematics education. Communications of the Society for Didactics of Mathematics, 117.
Canonigo, A.M. (2024). Leveraging AI to enhance students' conceptual understanding and confidence in mathematics. Journal of Computer Assisted Learning, 40(6), 3215–3229. https://doi.org/10.1111/jcal.13065
Cárdenas, R., Vásquez, HGE, Gamboa, DAP, Arteaga-Arcentales, E., & Carrera, JEM (2025). Exploring AI-powered adaptive learning systems and their implementation in educational settings: A systematic literature review. International Journal of Innovative Research and Scientific Studies, 8(4), 832–842. https://doi.org/10.53894/ijirss.v8i4.7961
Chudziak, JA, & Kostka, A. (2025). AI-powered math tutoring: Platform for personalized and adaptive education. Lecture Notes in Computer Science, 462–469. https://doi.org/10.1007/978-3-031-98465-5_58
Cohn, C., Rayala, S., Srivastava, N., Fonteles, J., Jain, S., Luo, X., Mereddy, D., Mohammed, N., & Biswas, G. (2025). A theory of adaptive scaffolding for LLM-based pedagogical agents. arXiv. https://doi.org/10.48550/arxiv.2508.01503
Contel, F., & Cusi, A. (2025). Investigating the Role of ChatGPT in Supporting Metacognitive Processes During Problem-Solving Activities. Digital Experiences in Mathematics Education, 11(1), 167–191. https://doi.org/10.1007/s40751-024-00164-7
Cosentino, G., Anton, J., Sharma, K., Gelsomini, M., Giannakos, M. N., & Abrahamson, D. (2025). Generative AI and multimodal data for educational feedback: Insights from embodied math learning. British Journal of Educational Technology. https://doi.org/10.1111/bjet.13587
Deng, R., Jiang, M., Yu, X., Lu, Y., & Liu, S. (2025). Does ChatGPT enhance student learning? A systematic review and meta-analysis of experimental studies. Computers & Education, 227, 105224. https://doi.org/10.1016/j.compedu.2024.105224
Dinsmore, D.L., & Fryer, L.K. (2026). What does current genAI actually mean for student learning? Learning and Individual Differences, 125, 102834. https://doi.org/10.1016/j.lindif.2025.102834
Fock, A., & Siller, H.-S. (2025). Generative Artificial Intelligence in Secondary STEM Education in the Light of Human Flourishing: A Scoping Literature Review. Research Square. https://doi.org/10.21203/rs.3.rs-6923010/v1
Gerlich, M. (2025). AI Tools in Society: Impacts on Cognitive Offloading and the Future of Critical Thinking. Societies, 15(1), 6. https://doi.org/10.3390/soc15010006
Gisiger, M. (2025, April 17). The role of artificial intelligence in learning – opportunities and risks. Michael Gisiger. https://text.tchncs.de/gisiger/die-rolle-von-kunstlicher-intelligenz-im-lernen-chancen-und-risiken
Gonnermann-Müller, J., Haase, J., Fackeldey, K., & Pokutta, S. (2025). FACET: Teacher-centered LLM-based multi-agent systems – Towards personalized educational worksheets. arXiv. https://doi.org/10.48550/arxiv.2508.11401
Guerino, G., Challco, GC, Veloso, TE, Oliveira, L., Penha, RSD, Melo, RF, Vieira, T., Marinho, MLM, Macario, V., Bittencourt, II, Isotani, S., & Dermeval, D. (2023). Teacher-centered intelligent tutoring systems: Design considerations from Brazilian public school teachers. Anais do XXXIV Simpósio Brasileiro de Informática na Educação. https://doi.org/10.5753/sbie.2023.235159
Gulz, A., & Haake, M. (2021). No child left behind, nor singled out: Is it possible to combine adaptive instruction and inclusive pedagogy in early math software? SN Social Sciences, 1, 205. https://doi.org/10.1007/s43545-021-00205-7
Guo, J., Ma, Y., Jang, H., Li, T., Wu, J., Huang, D., Han, F., Noetel, M., Liao, K., Tang, X., & Kui, X. (2025). The impact of artificial intelligence on primary school students' motivation and engagement: A systematic review. PsyArXiv. https://doi.org/10.31234/osf.io/ecspn_v1
Harahap, R. (2024). The role of ChatGPT in enhancing mathematics education: A systematic review. Annals of the Vietnam Academy of Science and Technology, 28(2s), 511–524. https://doi.org/10.52783/anvi.v28.2753
Hocine, N., Moussa, MBO, & Ali, S.A. (2023). Posicalculia: An adaptive virtual environment for children with learning difficulties. IEEE INSTA 2023. https://doi.org/10.1109/inista59065.2023.10310592
Holmes, V.M. (2024). Designing an AI math tutor for children with dyslexia, dysgraphia, and dyscalculia. https://doi.org/10.58445/rars.2035
Holmes, W., Bialik, M., & Fadel, C. (2019). Artificial Intelligence in Education: Promise and Implications for Teaching and Learning. Center for Curriculum Redesign.
Holmes, W., Porayska-Pomsta, K., Holstein, K., Sutherland, E., Baker, T., Shum, S., Santos, OC, Rodrigo, M., Cukurova, M., Bittencourt, I., & Koedinger, K. (2021). Ethics of AI in Education: Towards a Community-Wide Framework. International Journal of Artificial Intelligence in Education, 32, 504–526. https://doi.org/10.1007/s40593-021-00239-1
Holmes, W., & Tuomi, I. (2022). State of the art and practice in AI in education. European Journal of Education, 57(4), 542–570. https://doi.org/10.1111/ejed.12533
Hwang, S. (2022). Examining the Effects of Artificial Intelligence on Elementary Students' Mathematics Achievement: A Meta-Analysis. Sustainability, 14(20), 13185. https://doi.org/10.3390/su142013185
Jančařík, A., Michal, J., & Novotná, J. (2023). Using AI Chatbot for Math Tutoring. Journal of Education Culture and Society, 14(2), 285–296. https://doi.org/10.15503/jecs2023.2.285.296
Kaliisa, R., Misiejuk, K., López-Pernas, S., & Saqr, M. (2025). How does artificial intelligence compare to human feedback? A meta-analysis of performance, feedback perception, and learning dispositions. Educational Psychology, 1–32. https://doi.org/10.1080/01443410.2025.2553639
Kestin, G., Miller, K., Klales, A., Milbourne, T., & Ponti, G. (2025). AI tutoring outperforms in-class active learning: An RCT introducing a novel research-based design in an authentic educational setting. Scientific Reports, 15(1), 17458. https://doi.org/10.1038/s41598-025-97652-6
KirschnerED (2025, August 15). ChatGPT in Education: An Effect in Search of a Cause? https://www.kirschnered.nl/2025/08/15/chatgpt-in-education-an-effect-in-search-of-a-cause/
KMK (2024). Recommendations for educational authorities on dealing with artificial intelligence in school education processes. https://www.kmk.org/fileadmin/veroeffentlichungen_beschluesse/2024/2024_10_10-Handlungsempfehlung-KI.pdf
Kortenkamp, U. (2024). How much math does humanity need? Core mathematical competencies in the face of AI. https://doi.org/10.20378/irb-104036
Kumar, H., Rothschild, DM, Goldstein, DG, & Hofman, JM (2023). Math Education with Large Language Models: Peril or Promise? (SSRN Scholarly Paper No. 4641653). Social Science Research Network. https://doi.org/10.2139/ssrn.4641653
Kuo, B.-C., Bai, Z.-E., & Lin, C.-H. (2026). Developing an AI learning companion for mathematics problem solving in elementary schools. Computers & Education, 240, 105463. https://doi.org/10.1016/j.compedu.2025.105463
Li, M. (2024). Integrating Artificial Intelligence in Primary Mathematics Education: Investigating Internal and External Influences on Teacher Adoption. International Journal of Science and Mathematics Education. https://doi.org/10.1007/s10763-024-10515-w
Ligthart, MEU, de Droog, SM, Bossema, M., Elloumi, L., Hoogland, K., Smakman, MHJ, Hindriks, KV, & Ben Allouch, S. (2023). Design specifications for a social robot math tutor. In G. Castellano, L. Riek, M. Cakmak, & J. Leite (Eds.), Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction (pp. 321–330). ACM/IEEE. https://doi.org/10.1145/3568162.3576957
Listyaningrum, P., Retnawati, H., Harun, H., & Ibda, H. (2024). Digital learning using ChatGPT in elementary school mathematics learning: A systematic literature review. Indonesian Journal of Electrical Engineering and Computer Science, 36(3), 1701–1710. https://doi.org/10.11591/ijeecs.v36.i3.pp1701-1710
Liu, J., Sun, D., Sun, J., Wang, J., & Yu, PLH (2025). Designing a generative AI enabled learning environment for mathematics word problem solving in primary schools: Learning performance, attitudes and interaction. Computers and Education: Artificial Intelligence, 9, 100438. https://doi.org/10.1016/j.caeai.2025.100438
Makransky, G., Shiwalia, BM, Herlau, T., & Blurton, S. (2024). Beyond the „Wow“ factor: Using Generative AI for Increasing Generative Sense-Making. Review. https://doi.org/10.21203/rs.3.rs-5622133/v1
Mott, B., Gupta, A., Glazewski, K., Ottenbreit-Leftwich, A., Hmelo-Silver, C., Scribner, A., Lee, S., & Lester, J. (2023). Fostering Upper Elementary AI Education: Iteratively Refining a Use-Modify-Create Scaffolding Progression for AI Planning. Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 2, 647. https://doi.org/10.1145/3587103.3594170
Ninaus, M., & Sailer, M. (2022). Closing the loop – The human role in artificial intelligence for education. Frontiers in Psychology, 13. https://doi.org/10.3389/fpsyg.2022.956798
Opesemowo, OAG, & Adewuyi, HO (2024). A systematic review of artificial intelligence in mathematics education: The emergence of 4IR. Eurasia Journal of Mathematics, Science and Technology Education, 20(7), em2478. https://doi.org/10.29333/ejmste/14762
Qian, K., Liu, S., Li, T., Raković, M., Li, X., Guan, R., Molenaar, I., Nawaz, S., Swiecki, Z., Yan, L., & Gašević, D. (2026). Towards reliable generative AI-driven scaffolding: Reducing hallucinations and enhancing quality in self-regulated learning support. Computers & Education, 240, 105448. https://doi.org/10.1016/j.compedu.2025.105448
Rosenfelder, A., Levitin, MD, & Gilead, M. (2025). Towards social superintelligence? AI infers various psychological traits from text without specific training, outperforming human judges. Computers in Human Behavior: Artificial Humans, 6, 100228. https://doi.org/10.1016/j.chbah.2025.100228
Ruan, S., He, J., Ying, R., Burkle, J., Hakim, D., Wang, A., Yin, Y., Zhou, L., Xu, Q., AbuHashem, AA, Dietz, G., Murnane, EL, Brunskill, E., & Landay, JA (2020). Supporting children's math learning with feedback-augmented narrative technology. Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3392063.3394400
Rumbelow, M., & Coles, A. (2024). The Promise of AI Object-Recognition in Learning Mathematics: An Explorative Study of 6-Year-Old Children's Interactions with Cuisenaire Rods and the Blockplay.ai App. Education Sciences, 14(6), 591. https://doi.org/10.3390/educsci14060591
Samuelsson, J. (2023). Arithmetic fact fluency supported by artificial intelligence. Frontiers in Education Technology, 6(1), 13. https://doi.org/10.22158/fet.v6n1p13
Sayed, W.S., Noeman, A., Abdellatif, A., Abdelrazek, M., Badawy, MG, Hamed, AEA, & El-Tantawy, S. (2022). AI-based adaptive personalized content presentation and exercises navigation for an effective and engaging e-learning platform. Multimedia Tools and Applications, 82(3), 3303–3333. https://doi.org/10.1007/s11042-022-13076-8
Schneider, RJ (n.d.). The use of AI to support lesson preparation: How can AI-generated practice exercises for primary school mathematics be evaluated from a subject-didactic perspective? [Unpublished manuscript].
Son, T. (2024). Artificial intelligence in mathematics education: A systematic literature review on intelligent tutoring systems. Journal of Educational Research in Mathematics, 34(2), 187. https://doi.org/10.29275/jerm.2024.34.2.187
Song, Y., Kim, J., Liu, Z., Li, C., & Xing, W. (2024). Students' perceived roles, opportunities, and challenges of a generative AI-powered teachable agent: A case of middle school math class. Journal of Research on Technology in Education, 1–19. https://doi.org/10.1080/15391523.2024.2447727
Topkaya, Y., Doğan, Y., Batdı, V., & Aydın, S. (2025). Artificial intelligence applications in primary education: A quantitatively-supported mixed-meta method study [Preprint]. Preprints. https://doi.org/10.20944/preprints202501.2263.v1
Vitale, A., & Dello Iacono, U. (2024). Using social robots as inclusive educational technology for mathematics learning through storytelling. European Public & Social Innovation Review, 9, 1–17. https://doi.org/10.31637/epsir-2024-672
Wang, D., Shan, D., Ju, R., Kao, B., Zhang, C., & Chen, G. (2025). Investigating dialogic interaction in K12 online one-on-one mathematics tutoring using AI and sequence mining techniques. Education and Information Technologies, 30(7), 9215–9240. https://doi.org/10.1007/s10639-024-13195-9
Wang, J., & Fan, W. (2025). The effect of ChatGPT on students' learning performance, learning perception, and higher-order thinking: Insights from a meta-analysis. Humanities and Social Sciences Communications, 12(1), 1–21. https://doi.org/10.1057/s41599-025-04787-y
Wang, L., & Nie, Z. (2023). Research on adaptive learning in K-12 education in the perspective of teachers' artificial intelligence literacy: Development, technology, improvement strategies. IEEE CSTE 2023. https://doi.org/10.1109/cste59648.2023.00059
Wang, RE, Ribeiro, AT, Robinson, CD, Loeb, S., & Demszky, D. (2024). Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise. arXiv preprint arXiv:2410.03017. https://arxiv.org/abs/2410.03017
Wezendonk, A., & Veldhuis, M. (2024). Adaptive empty systems and the didactic solution for basic school learning in the field of knowledge. Tijdschrift voor Onderwijs en Praktijk in Statistiek. https://doi.org/10.54657/tops.13844
Wu, R., & Yu, Z. (2024). Do AI chatbots improve students learning outcomes? Evidence from a meta-analysis. British Journal of Educational Technology, 55(1), 10–33. https://doi.org/10.1111/bjet.13334
Wu, X.-Y., Radloff, J., Yeter, I., Wang, L., & Chiu, TKF (2025). Designing artificial intelligence chatbots for self-regulated learning from a systematic review based on Habermas's three interests. Interactive Learning Environments. https://doi.org/10.1080/10494820.2025.2563086
Yim, IHY, & Su, J. (2025). Artificial intelligence literacy education in primary schools: A review. International Journal of Technology and Design Education. https://doi.org/10.1007/s10798-025-09979-w
Zheng, L., Niu, J., Zhong, L., & Gyasi, J.F. (2023). The effectiveness of artificial intelligence on learning achievement and learning perception: A meta-analysis. Interactive Learning Environments, 31(9), 5650–5664. https://doi.org/10.1080/10494820.2021.2015693

Leave a Reply