Skip to content

Submission for Fossee Python Task 3: Research plan evaluating open-source coding models for high-level student competence analysis in Python, comparing Qwen2.5-Coder with alternatives and validating through a prototype pipeline.

Notifications You must be signed in to change notification settings

hpriyankaa/fossee-python-task

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

92 Commits
 
 
 
 
 
 

Repository files navigation

Fossee-python-task-3

Task Title: "Python Screening Task 3: Evaluating Open Source Models for Student Competence Analysis"
Name: H Priyanka
Email: priyankah2407@gmail.com

Table of Contents

Research Plan

To assess LLMs and Open Source Models on their ability to support high-level student competence analysis in Python, I researched on various models such as CodeLlama[6][7], DeepSeek Coder V2[8], DeepSeek-V3[9], Qwen3-Coder[10], StarCoder[1], WizardCoder[2], and Codestral[3][4]. While each had it’s own advantages,like CodeLlama’s[6][7] specialization in Python, DeepSeek’s[8][9] scalability, Qwen3-Coder’s[10] long context. These models have constraints like high resource demands or less established ecosystems. Out of all these choices, I find Qwen2.5-Coder(instruction-tuned)[5][13] as the most balanced option because it is education-task-oriented, exhibits high-performance on the benchmarks such as HumanEval[11][14] and MBPP[12][14], and can be tested and scaled across parameters 0.5B to 32B on basis of available compute resources. A model is suitable for competence analysis if it can not only solve problems but analyzes the student's code, surfaces misconceptions, and proposes prompts that encourage further reasoning without directly revealing the final solution. Qwen2.5-Coder-Instruct[5][13] fulfills all these criterion better than other models.

To validate this, I developed a Planner–Executor–Critic (PEC)[15] prototype where the Planner detects concepts and gaps, the Executor generates Socratic-style questions, and the Critic checks their depth and clarity. This framework offers a systematic means to quantify whether the model produces meaningful prompts that surface reasoning errors and leads to further additional learning. The results showed that Qwen2.5-Coder-Instruct[5][13] gave interpretable, pedagogically useful feedback with reasonable computational cost.Despite newer models like Qwen3-Coder[10] pushing raw accuracy higher and DeepSeek-V3[9] excelling at general-purpose reasoning, their higher resource demands and less established ecosystems makes them less practical for this use case and Qwen3-Coder[10] is still in earlier stages of development. In contrast, Qwen2.5-Coder[5][13] offers proven reliability, better integration resources, and extensive real-world testing. Collectively, it can be said that these factors assures Qwen2.5-Coder-Instruct[5][13] is not only the most practical choice but also the model best positioned to bring together accuracy, interpretability, and scalability to actual educational applications.

Reasoning

What makes a model suitable for high-level competence analysis?

  • Pedagogical alignment: The model should analyze student-written code, identify misconceptions, and generate probing prompts that encourage reasoning without revealing full solutions.
  • Robust training data: It must be trained on high-quality, code-rich datasets that capture syntax, logic, common patterns, and typical student errors.
  • Practical scalability: The model should be efficient to deploy on common infrastructure (Colab, mid-range GPUs, or CPUs) and adaptable across difficulty levels, from basic loops to advanced algorithms.
  • Fairness and reliability: It should minimize bias in evaluating coding styles, provide interpretable feedback, and be backed by an actively maintained ecosystem with frequent updates and integration support.

How would you test whether a model generates meaningful prompts?

To test whether a model generates meaningful prompts, I set up a PEC-style configuration (Planner-Executor-Critic). In this setup:

  • The Planner analyzed student-written code, identifying underlying ideas, conceptual gaps, and potential misconceptions.
  • The Executor generated Socratic prompts that directly targeted these gaps, probing the student’s reasoning without revealing the final solution.
  • The Critic evaluated each prompt against a clear rubric, assessing its relevance to the code, the depth of its inquiry, the clarity of its phrasing, and its strict avoidance of solution leakage. This was complemented by evaluation on a curated dataset, which confirmed whether prompts actually surfaced misconceptions and guided deeper reasoning. By combining rubric-based evaluation with observed learning outcomes, I was able to determine when a prompt was not only linguistically well-formed, but also pedagogically effective. Sample outputs from this process can be found in the Prototype Output directory.

What trade-offs might exist between accuracy, interpretability, and cost?

  • Accuracy vs complexity: Larger models (e.g., Qwen3, DeepSeek-V3) achieve higher accuracy and handle complex reasoning, but at significantly higher compute cost.
  • Interpretability: Comparitively smaller instruction-tuned models (e.g., Qwen2.5) often produce feedback that is easier for learners to follow, while larger models may generate outputs that are harder to understand or justify.
  • Cost vs deployability: Comparatively smaller models are cheaper to run and can be deployed in practical environments (Colab, classroom servers), while large ones may be impractical.
  • Ecosystem maturity: Established models with active updates and integrations provide more reliable, long-term value than newer but less proven alternatives. The trade-off is choosing a model that balances credible results, interpretability for learners, cost-effective deployment, and long-term reliability rather than optimizing for a single dimension.

Why did you choose the model you evaluated, and what are its strengths or limitations?

  • Instruction-tuning advantage: Qwen2.5-Coder-Instruct was chosen over its base version because it better follows pedagogical prompts, identifies misconceptions, and generates reasoning-focused feedback.
  • Balanced performance: It combines solid coding accuracy with interpretability, scaling across parameter sizes (0.5B–32B) without requiring heavy compute resources.
  • Practical alternative to SOTA: Newer models like Qwen3 or DeepSeek-V3 achieve higher raw accuracy but are costlier to run and less stable for classroom or research use.
  • Ecosystem reliability: Qwen2.5 benefits from active updates, integrations, and a mature community, making it a dependable option for long-term educational deployment. Limitation: Qwen2.5 does not yet match the very top-end accuracy or longest context handling of newer models, but it delivers the best balance of interpretability, efficiency, and reliability for competence analysis.

Benchmarks

Sample Image

Figure 1: Qwen 2.5-Coder Benchmark vs other coding models [13]


Sample Image

Figure 2: Qwen2.5-Coder model size comparison [13]


Sample Image

Figure 3: Qwen2.5-Coder vs peers across tasks [13]


Sample Image

Figure 4: EvalPlus Benchmark Leaderboad [14]

Prototype

As part of this work, I built a prototype PEC pipeline that analyzes student-written Python code, identifies key concepts and misconceptions, generates Socratic-style questions, and evaluates them using a structured rubric.

This prototype is lightweight, runs on open-source models (e.g., Qwen2.5-Coder via vLLM), and produces clean reports in both console and artifact formats (JSONL, CSV, Markdown).

Explore the full prototype, installation steps, and detailed usage here:
Prototype Directory.


Future Research Plan

Cross-Model Comparisons

Conduct side-by-side evaluations of multiple models including Qwen2.5-Coder, Qwen3-Coder, DeepSeek Coder v2 and much more using the same PEC prototype. Track differences in accuracy, interpretability, efficiency, and pedagogical usefulness.

Instruction-Tuning Customization

Fine-tune Qwen2.5-Coder-Instruct on education-specific datasets (e.g., student submissions with annotated misconceptions) to better align the model with pedagogical needs.

  • Use CodeT5/GraphCodeBERT sidecar tagging to automatically generate misconception and concept labels for fine-tuning data.
  • Reduce manual annotation effort and ensure higher-quality supervision for education-specific tuning.

Prototype Enhancement (PEC 2.0)

Expand the Planner–Executor–Critic pipeline into a fully adaptive feedback loop, where identified misconceptions automatically trigger new rounds of targeted Socratic questioning. Add multi-round evaluation to measure whether students improve after model-driven feedback.

  • Incorporate Knowledge Tracing (pyBKT/pyKT) so each feedback round updates student concept mastery scores.
  • Drive the next set of Socratic prompts based on low-mastery concepts to make the loop adaptive and personalized.

References

[1]R. Li et al., “StarCoder: may the source be with you!,” arXiv.org, May 09, 2023. https://arxiv.org/abs/2305.06161.

[2] Z. Luo et al., “WizardCoder: Empowering Code Large Language Models with Evol-Instruct,” arXiv.org, Jun. 14, 2023. https://arxiv.org/abs/2306.08568.

[3] Mistral AI, “Codestral | Mistral AI,” Mistral.ai, May 29, 2024. [Online]. Available: https://mistral.ai/news/codestral

[4] Hugging Face, “mistralai/Codestral-22B-v0.1,” Hugging Face, May 29, 2024. [Online]. Available: https://huggingface.co/mistralai/Codestral-22B-v0.1

[5] B. Hui et al., “QWen2.5-Coder Technical Report,” arXiv.org, Sep. 18, 2024. https://arxiv.org/abs/2409.12186.

[6] B. Rozière et al., “Code llama: Open Foundation Models for code,” arXiv.org, Aug. 24, 2023. https://arxiv.org/abs/2308.12950

[7] Meta, “Meta Code Llama,” Llama.com, Aug. 24, 2023. [Online]. Available: https://www.llama.com/code-llama/

[8] DeepSeek-Ai et al., “DeepSeek-Coder-V2: Breaking the barrier of Closed-Source Models in Code Intelligence,” arXiv (Cornell University), Jun. 2024, doi: 10.48550/arxiv.2406.11931. Available: https://www.researchgate.net/publication/381517674_DeepSeek-Coder-V2_Breaking_the_Barrier_of_Closed-Source_Models_in_Code_Intelligence

[9] DeepSeek-Ai et al., “DeepSeek-V3 Technical Report,” arXiv.org, Dec. 27, 2024. https://arxiv.org/abs/2412.19437v1

[10] A. Yang et al., “QWEN3 Technical Report,” arXiv.org, May 14, 2025. Available: https://arxiv.org/abs/2505.09388

[11] M. Chen et al., “Evaluating large language models trained on code,” arXiv.org, Jul. 07, 2021. https://arxiv.org/abs/2107.03374

[12] J. Austin et al., “Program Synthesis with Large Language Models,” arXiv.org, Aug. 16, 2021. https://arxiv.org/abs/2108.07732

[13] Qwen Team, “QWeN2.5-Coder: Code more, learn more!,” Qwen, Sep. 18, 2024. https://qwenlm.github.io/blog/qwen2.5-coder/

[14] EvalPlus, “EvalPlus leaderboard,” EvalPlus. [Online]. Available: https://evalplus.github.io/leaderboard.html

[15] Anthropic, “Building effective AI agents,” Anthropic. [Online]. Available: https://www.anthropic.com/engineering/building-effective-agents

About

Submission for Fossee Python Task 3: Research plan evaluating open-source coding models for high-level student competence analysis in Python, comparing Qwen2.5-Coder with alternatives and validating through a prototype pipeline.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages