Agents League was designed to showcase what agentic AI can look like when developers move beyond singleâprompt interactions and start building systems that plan, reason, verify, and collaborate.
Across three competitive tracksâCreative Apps, Reasoning Agents, and Enterprise Agentsâparticipants had two weeks to design and ship real AI agents using productionâready Microsoft and GitHub tools, supported by live coding battles, community AMAs, and async builds on GitHub.
Today, weâre excited to spotlight the winning project for the Reasoning Agents track, built on Microsoft Foundry: CertPrep MultiâAgent System â Personalised Microsoft Exam Preparation by Athiq Ahmed.
The Reasoning Agents Challenge Scenario
The goal of the Reasoning Agents track challenge was to design a multiâagent system capable of effectively assisting students in preparing for Microsoft certification exams. Participants were asked to build an agentic workflow that could understand certification syllabi, generate personalized study plans, assess learner readiness, and continuously adapt based on performance and feedback. The suggested reference architecture modeled a realistic learning journey: starting from freeâform student input, a sequence of specialized reasoning agents collaboratively curated Microsoft Learn resources, produced structured study plans with timelines and milestones, and maintained learner engagement through reminders. Once preparation was complete, the system shifted into an assessment phase to evaluate readiness and either recommend the appropriate Microsoft certification exam or loop back into targeted remediationâemphasizing reasoning, decisionâmaking, and humanâinâtheâloop validation at every step.
All details are available here: agentsleague/starter-kits/2-reasoning-agents at main · microsoft/agentsleague.
The Winning Project: CertPrep MultiâAgent System
The CertPrep MultiâAgent System is an AI solution for personalized Microsoft certification exam preparation, supporting nine certification exam families.
At a high level, the system turns freeâform learner input into a structured certification plan, measurable progress signals, and actionable recommendationsâdemonstrating exactly the kind of reasoned orchestration this track was designed to surface.
Inside the MultiâAgent Architecture
At its core, the system is designed as a multiâagent pipeline that combines sequential reasoning, parallel execution, and humanâinâtheâloop gates, with traceability and responsible AI guardrails.
The solution is composed of eight specialized reasoning agents, each focused on a specific stage of the learning journey:
- LearnerProfilingAgent â Converts freeâtext background information into a structured learner profile using Microsoft Foundry SDK (with deterministic fallbacks).
- StudyPlanAgent â Generates a weekâbyâweek study plan using a constrained allocation algorithm to respect the learnerâs available time.
- LearningPathCuratorAgent â Maps exam domains to curated Microsoft Learn resources with trusted URLs and estimated effort.
- ProgressAgent â Computes a weighted readiness score based on domain coverage, time utilization, and practice performance.
- AssessmentAgent â Generates and evaluates domainâproportional mock exams.
- CertificationRecommendationAgent â Issues a clear âGO / CONDITIONAL GO / NOT YETâ decision with remediation steps and nextâcert suggestions.
Throughout the pipeline, a 17ârule Guardrails Pipeline enforces validation checks at every agent boundary, and two explicit humanâinâtheâloop gates ensure that decisions are made only when sufficient learner confirmation or data is present.
CertPrep leverages Microsoft Foundry Agent Service and related tooling to run this reasoning pipeline reliably and observably:
- Managed agents via Foundry SDK
- Structured JSON outputs using GPTâ4o (JSON mode) with conservative temperature settings
- Guardrails enforced through Azure Content Safety
- Parallel agent fanâout using concurrent execution
- Typed contracts with Pydantic for every agent boundary
- AI-assisted development with GitHub Copilot, used throughout for code generation, refactoring, and test scaffolding
Notably, the full pipeline is designed to run in under one second in mock mode, enabling reliable demos without live credentials.
User Experience: From Onboarding to Exam Readiness
Beyond its backend architecture, CertPrep places strong emphasis on clarity, transparency, and user trust through a wellâstructured frontâend experience. The application is built with Streamlit and organized as a 7âtab interactive interface, guiding learners stepâbyâstep through their preparation journey.
From a userâs perspective, the flow looks like this:
- Profile & Goals Input
Learners start by describing their background, experience level, and certification goals in natural language. The system immediately reflects how this input is interpreted by displaying the structured learner profile produced by the profiling agent. - Learning Path & Study Plan Visualization
Once generated, the study plan is presented using visual aids such as Ganttâstyle timelines and domain breakdowns, making it easy to understand weekly milestones, expected effort, and progress over time. - Progress Tracking & Readiness Scoring
As learners move forward, the UI surfaces an examâweighted readiness score, combining domain coverage, study plan adherence, and assessment performanceâhelping users understand why the system considers them ready (or not yet). - Assessments and Feedback
Practice assessments are generated dynamically, and results are reported alongside actionable feedback rather than just raw scores. - Transparent Recommendations
Final recommendations are presented clearly, supported by reasoning traces and visual summaries, reinforcing trust and explainability in the agentâs decisionâmaking.
The UI also includes an Admin Dashboard and demoâfriendly modes, enabling judges, reviewers, or instructors to inspect reasoning traces, switch between live and mock execution, and demonstrate the system reliably without external dependencies.
Why This Project Stood Out
This project embodies the spirit of the Reasoning Agents track in several ways:
- â Clear separation of reasoning roles, instead of promptâheavy monoliths
- â Deterministic fallbacks and guardrails, critical for educational and decisionâsupport systems
- â Observable, debuggable workflows, aligned with Foundryâs production goals
- â Explainable outputs, surfaced directly in the UX
It demonstrates how agentic patterns translate cleanly into maintainable architectures when supported by the right platform abstractions.
Try It Yourself
Explore the project, architecture, and demo here:
- đ GitHub Issue (full project details): https://github.com/microsoft/agentsleague/issues/76
- đ„ Demo video: https://www.youtube.com/watch?v=okWcFnQoBsE
- đ Live app (mock data): https://agentsleague.streamlit.app/






