Skip to content

Commit 3bb7617

Browse files
committed
feat: add conclusion and future work sections with detailed insights
1 parent 0aa89f3 commit 3bb7617

File tree

2 files changed

+136
-2
lines changed

2 files changed

+136
-2
lines changed

paper/content/conclusion.typ

Lines changed: 41 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,43 @@
11
= Conclusion
22

3-
#lorem(200)
3+
This paper presented Harmonia, an instructor-facing tool for automated, data-driven assessment of student collaboration in team-based software engineering projects, alongside a complementary student-facing feedback system that provides formative code quality guidance. Together, these systems address critical gaps in the current evaluation process for team projects in introductory programming courses.
4+
5+
== Key Contributions
6+
7+
Harmonia introduces several significant innovations to the field of software engineering education. First, it provides a *comprehensive, multi-dimensional assessment* of collaboration through the Collaboration Quality Index (CQI), which combines effort-based metrics (weighted at 55%) with lines of code balance, temporal spread, and file ownership spread. This approach moves beyond simple commit counting to capture the semantic quality of contributions, distinguishing between a high-effort feature implementation and a trivial formatting change through LLM-driven analysis.
8+
9+
Second, Harmonia achieves *transparency and scalability* by automating the assessment of up to 200 teams in a single course run. The real-time analysis pipeline with Server-Sent Events feedback allows instructors to monitor progress and begin reviewing teams before the full analysis completes. The separation of Git analysis from AI analysis enables partial results to be available immediately, reducing instructor wait time.
10+
11+
Third, the system demonstrates *practical integration* with existing course infrastructure. By connecting directly to the Artemis learning management platform and automating data retrieval, Harmonia eliminates the manual export burden that has historically hindered instructor adoption of assessment tools. Tutors can upload pair programming attendance records and automatically verify compliance across all teams.
12+
13+
Finally, the student-facing feedback system provides a *learning-centric complement* to instructor assessment. By delivering non-graded, priority-categorized feedback in real time, the system encourages early and frequent feedback requests without evaluation anxiety, supporting a growth mindset and deeper engagement with code quality principles throughout the project lifecycle.
14+
15+
== Evaluation Results
16+
17+
Our manual evaluation of five representative teams provided validation of Harmonia's approach. The inter-rater reliability study showed strong agreement between independent reviewers across most teams and dimensions, with an average Pearson correlation above 70% for effort, complexity, and novelty assessments. Team A1 demonstrated exceptional reviewer agreement (up to 99% on some dimensions), suggesting that well-structured commits are consistently recognizable, while Team H showed the greatest disagreement (70% average), highlighting the inherent subjectivity in assessing complex work. These results suggest that Harmonia's LLM-based assessments operate within a reasonable range of human judgment variability.
18+
19+
== Impact and Adoption
20+
21+
Harmonia has been deployed and used in the ITP course for the winter semester 2025/26. The tool has provided instructors with unprecedented visibility into team collaboration patterns across the course, enabling more transparent and consistent grading decisions. By surfacing collaboration imbalances and commit quality issues through the CQI and visual breakdowns, Harmonia has helped instructors identify teams requiring intervention and justify grading decisions with evidence.
22+
23+
The student feedback system has similarly shown positive adoption patterns, with students utilizing the "Request AI Feedback" button frequently throughout the project lifecycle. The priority taxonomy has proven effective in helping students allocate effort, and the integration with the IDE has minimized friction in seeking guidance.
24+
25+
== Limitations and Considerations
26+
27+
Despite its contributions, Harmonia has inherent limitations that should be acknowledged. The reliance on LLM-based effort estimation introduces a dependency on model quality and training data. While the paper evaluated Harmonia using GPT-5-Mini, the approach's effectiveness with other models remains unexplored. Additionally, the sequential AI analysis phase causes the overall analysis time to scale linearly with team and commit counts, exceeding the initial 15-minute target (currently taking up to 24 hours for a full course).
28+
29+
The CQI, while comprehensive, remains a composite metric that can mask important qualitative differences in collaboration. A team with high CQI might nevertheless have divided responsibilities without knowledge sharing, or conversely, might share code ownership through effective pair programming despite imbalanced commit counts. Instructors must therefore treat the CQI as an *indicator* warranting investigation rather than a definitive judgment.
30+
31+
Furthermore, Git history reflects only authorship and timestamps, not the actual effort or time invested by each student. Commits pushed by one student might represent collaborative work, and vice versa. While pair programming verification and the AI effort estimation mitigate this limitation, no purely automated system can fully capture collaboration quality.
32+
33+
== Broader Implications
34+
35+
Harmonia contributes to a growing body of work on using large language models to support software engineering education. The success of LLM-driven effort estimation in this context suggests broader applicability to other educational domains requiring nuanced assessment of student work. The open question of how to calibrate LLM-based scoring against human judgment—and how much variation is acceptable—remains important for future tool development.
36+
37+
The integration of both instructor-facing and student-facing tools demonstrates the complementary roles of summative and formative feedback. While Harmonia provides instructors with evidence for grading, the student feedback system supports learning throughout the process. This two-sided approach aligns with pedagogical best practices and could serve as a model for other courses.
38+
39+
## Conclusion
40+
41+
Harmonia represents a significant step forward in automating and democratizing the assessment of student collaboration in team-based software projects. By combining automated Git analysis with LLM-driven effort estimation, and by providing both instructor and student perspectives, the tool addresses longstanding challenges in software engineering education. The deployment in ITP demonstrates practical feasibility, while the manual evaluation provides evidence of reasonable alignment with human judgment.
42+
43+
As software engineering education continues to scale and emphasizes collaboration more prominently, tools like Harmonia will become increasingly important for maintaining assessment quality and fairness at scale. The approach pioneered in this work provides a foundation for future research and practice in this area.

paper/content/future_work.typ

Lines changed: 95 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,97 @@
11
= Future Work
22

3-
#lorem(200)
3+
While Harmonia provides a comprehensive solution to collaborative assessment in team-based software engineering projects, several avenues remain for further research and development. This section outlines key directions for enhancement of both the system itself and the underlying methodologies.
4+
5+
== Optimization and Performance
6+
7+
The most pressing technical limitation is the analysis completion time. The current sequential AI analysis phase causes end-to-end runtime to scale linearly with team and commit counts, with full course analysis taking up to 24 hours. Three approaches could address this bottleneck:
8+
9+
1. *Parallel AI Processing*: Rather than processing commits sequentially, the system could batch multiple commits and submit them to the LLM API in parallel. This would require careful orchestration to stay within rate limits and to manage costs, but could reduce runtime by an order of magnitude.
10+
11+
2. *Distributed Analysis*: Deploying multiple analysis workers on separate machines could process different teams in true parallel, further reducing total runtime. This would require a message queue (e.g., RabbitMQ, Apache Kafka) and careful state management.
12+
13+
3. *Model Caching and Heuristics*: Pre-computing effort estimates for common commit patterns (e.g., boilerplate changes, common library updates) could avoid redundant LLM calls. A hybrid approach using fast heuristics for obvious cases and LLM analysis only for ambiguous commits could dramatically improve performance.
14+
15+
== Model Calibration and Validation
16+
17+
This work used GPT-5-Mini for effort estimation, but the generalizability to other models remains unknown. Future work should:
18+
19+
1. *Comparative Model Study*: Evaluate Harmonia with multiple LLM backends (Claude, Gemini, open-source models like Llama) to understand how model choice affects CQI reliability and whether conclusions generalize.
20+
21+
2. *Extended Ground Truth Evaluation*: Expand the manual evaluation beyond the current five teams to a larger sample, ideally across multiple course offerings and programming languages. This would provide more robust calibration data.
22+
23+
3. *Effort Estimation Validation*: Conduct controlled studies where actual student effort (via time tracking or surveys) is compared against LLM effort estimates, providing direct validation of a key CQI component.
24+
25+
4. *Temporal Analysis*: Investigate whether AI effort ratings remain consistent across semesters and course contexts, or whether periodic recalibration is necessary.
26+
27+
== Enhanced Collaboration Metrics
28+
29+
The current CQI captures four aspects of collaboration, but additional dimensions could enrich the assessment:
30+
31+
1. *Code Review Quality*: Integrating pull request comments and code review patterns (if teams use these) could surface collaborative practices beyond commit history.
32+
33+
2. *Communication Analysis*: Analyzing team chat logs or commit messages for evidence of coordination and knowledge sharing could capture soft collaboration aspects missed by current metrics.
34+
35+
3. *Shared Ownership Dynamics*: Beyond file-level ownership, analyzing whether one student consistently refactors or improves the work of the other could reveal peer learning and mentorship.
36+
37+
4. *Knowledge Breadth*: Tracking whether both team members have contributed to all functional areas of the codebase could identify specialization imbalances.
38+
39+
== Expanded Integration and Interoperability
40+
41+
Currently, Harmonia integrates only with Artemis due to ITP's exclusive use of the platform. However, broader applicability requires:
42+
43+
1. *Multi-Platform Support*: Implement GitLab and GitHub integrations to serve courses using those platforms, as mentioned in the descoped FR-03.
44+
45+
2. *LMS Integration*: Extend integration to other learning management systems (Canvas, Blackboard, Moodle) to reach a broader population of courses.
46+
47+
3. *Data Export and Standards*: Define standardized formats for exporting CQI scores and collaboration evidence, enabling integration with other grading systems and research tools.
48+
49+
== Tutor and Student-Facing Enhancements
50+
51+
Harmonia currently focuses on instructor assessment, but expanding its scope to support other stakeholders could amplify impact:
52+
53+
1. *Tutor Dashboard*: Create a tutor-specific view that highlights problematic teams, shows pair programming compliance status, and surfaces commits requiring investigation. This would reduce the tutor's workload in monitoring team progress.
54+
55+
2. *Student Aggregate Feedback*: Provide teams with a privacy-respecting view of their own CQI over time, showing how their collaboration metrics evolve throughout the project. This could support reflection and improvement.
56+
57+
3. *Peer Feedback Integration*: Incorporate peer evaluation tools (e.g., CATME) and correlate peer feedback with CQI scores, validating whether collaboration metrics align with teammate perceptions.
58+
59+
4. *Visualization Improvements*: Expand the AI Analysis Feed with more detailed visualizations of effort distribution, temporal patterns, and ownership spread, making collaboration patterns more intuitive to grasp.
60+
61+
== Methodological Directions
62+
63+
From a research perspective, several questions remain open:
64+
65+
1. *Fairness and Bias in AI Assessment*: Investigate whether LLM effort ratings reflect inherent biases based on code style, programming language constructs, or commit message length. Conduct fairness audits to ensure ratings are equitable across different student populations and coding styles.
66+
67+
2. *Learning Outcomes Correlation*: Conduct longitudinal studies to correlate Harmonia's collaboration metrics with long-term learning outcomes, code quality in subsequent courses, and career success in team settings.
68+
69+
3. *Manipulation Resistance*: Explore edge cases where students might game the CQI (e.g., by making many small commits with inflated effort claims). Develop detection methods for manipulation attempts.
70+
71+
4. *Cross-Disciplinary Application*: Adapt Harmonia's approach to assess collaboration in non-programming domains such as hardware design, game development, or data science projects where version control usage varies.
72+
73+
== Expansion to Other Courses and Contexts
74+
75+
While validated in ITP, Harmonia's design principles could extend to other educational contexts:
76+
77+
1. *Advanced Software Engineering Courses*: Apply Harmonia to upper-level capstone projects where team sizes are larger and project complexity is higher.
78+
79+
2. *Industry and Research Settings*: Adapt Harmonia for use in research labs or software companies to support code review processes and collaboration assessment in professional contexts.
80+
81+
3. *Remote and Asynchronous Teams*: Investigate how Harmonia's metrics perform for fully remote teams with asynchronous collaboration patterns, where pair programming takes different forms.
82+
83+
== Long-Term Vision
84+
85+
The ultimate goal of tools like Harmonia is to support more equitable, transparent, and learning-focused assessment in team-based education. Future work should strive toward:
86+
87+
1. *Fully Automated Grading Pipelines*: Seamlessly integrate Harmonia with grading systems to automatically propose grades with evidence, reducing instructor burden while maintaining human oversight and review.
88+
89+
2. *Adaptive Feedback Systems*: Use CQI and collaboration patterns to dynamically adjust student feedback recommendations in real time (e.g., suggesting students request more feedback if collaboration imbalance is detected).
90+
91+
3. *Multi-Modal Assessment*: Combine Harmonia with other assessment tools (code quality checkers, test coverage analysis, design review systems) into a unified assessment platform that provides a holistic view of student projects.
92+
93+
4. *Ethical and Transparent AI in Education*: Develop best practices for integrating LLMs into educational assessment in ways that are transparent to students, educationally beneficial, and aligned with institutional values around fairness and integrity.
94+
95+
== Conclusion on Future Work
96+
97+
The challenges and opportunities outlined above demonstrate that while Harmonia represents a significant advance, the field of automated collaboration assessment in software engineering education is still in its early stages. Addressing these directions will make Harmonia and similar tools more practical, generalizable, and impactful. As large language models continue to evolve and as educational practice increasingly embraces data-driven assessment, the methods pioneered here provide a foundation for future innovation.

0 commit comments

Comments
 (0)