Transformer-Enhanced Deep Characterization of Sprints in Agile Software Development

Abstract: Agile methodologies, particularly Scrum, have become foundational in modern software development, with sprints serving as time-boxed iterations for incremental delivery. Accurate sprint outcome prediction is essential for effective project management and resource allocation. While Sprint2Vec introduced a deep learning framework using LSTM networks to characterize sprints through vector embeddings, it faces limitations in modeling long-range dependencies due to sequential processing constraints.

This paper presents Sprint2Vec +, which extends Sprint2Vec by integrating transformer-based architectures that leverage self-attention mechanisms to capture complex interactions among sprint elements. We evaluate our approach using datasets from five open-source projects—Apache, Atlassian, Jenkins, Spring, and Talendforge — containing over 5,000 sprints and 71,000 issues. Sprint2Vec + demonstrates a 10.8% improvement in combined prediction accuracy (10.3% for productivity, 11.3% for quality) with 40.9% faster inference time compared to the original LSTM-based approach. Beyond predictive improvements, Sprint2Vec + provides interpretable attention patterns revealing critical relationships between planning activities and implementation outcomes. Statistical significance testing (p < 0.001) confirms the robustness of improvements across all project domains. This work contributes a scalable, transformer-based solution for advancing sprint analytics in Agile software engineering.

Keywords: Deep Learning; Transformer Networks; Agile Development; Sprint Prediction; Software Engineering; Self-Attention; LSTM; Vector Embeddings

Published in: Proceedings of the International Conference on Smart Systems and Social Management (ICSSSM 2025) (Part of Springer Nature)

AUTHORS

Raghu Govind Alvandar


Pradeepta Mishra


VP of Artificial Intelligence, Beghou Consulting

Dr. Shinu Abhi


Professor and Director – Corporate Training

Leave a Reply

Your email address will not be published. Required fields are marked *